“NLP Status Report 2017-7-10”版本间的差异
来自cslt Wiki
第23行: | 第23行: | ||
* run two versions of the code on small data sets (Chinese-English) and tested these checkpoint | * run two versions of the code on small data sets (Chinese-English) and tested these checkpoint | ||
found version 1.0 save time about 0.03s per step, | found version 1.0 save time about 0.03s per step, | ||
− | + | and these two version has similar complexity and bleu values | |
found that the bleu is still good when the model is over fitting . | found that the bleu is still good when the model is over fitting . | ||
− | + | (reason: the test set and the train set of small data set are similar in content and style) | |
− | * run two versions of the code on big data sets (Chinese-English) . OOM(Out Of Memory) | + | * run two versions of the code on big data sets (Chinese-English) . |
− | + | OOM(Out Of Memory) error occurred when version 0.1 was trained using large data set,but version 1.0 worked | |
− | + | reason: improper distribution of resources by the tensorflow0.1 frame leads to exhaustion of memory resources | |
I had tried 4 times (just enter the same command), and version 0.1 worked | I had tried 4 times (just enter the same command), and version 0.1 worked | ||
− | + | found version 1.0 save time about 0.06s per step, and these two version has similar complexity and bleu values | |
* downloaded the wmt2014 data set ,used the English-French data set to run the code and | * downloaded the wmt2014 data set ,used the English-French data set to run the code and | ||
found the translation is not good (reason:improper word segmentation) | found the translation is not good (reason:improper word segmentation) |
2017年7月10日 (一) 06:18的版本
Date | People | Last Week | This Week |
---|---|---|---|
2017/7/3 | Jiyuan Zhang |
|
|
Aodong LI |
|
| |
Shiyue Zhang | |||
Shipan Ren |
found version 1.0 save time about 0.03s per step, and these two version has similar complexity and bleu values found that the bleu is still good when the model is over fitting . (reason: the test set and the train set of small data set are similar in content and style)
OOM(Out Of Memory) error occurred when version 0.1 was trained using large data set,but version 1.0 worked reason: improper distribution of resources by the tensorflow0.1 frame leads to exhaustion of memory resources I had tried 4 times (just enter the same command), and version 0.1 worked found version 1.0 save time about 0.06s per step, and these two version has similar complexity and bleu values
found the translation is not good (reason:improper word segmentation) |
|