“NLP Status Report 2017-6-5”版本间的差异
来自cslt Wiki
第37行: | 第37行: | ||
| 36.58, 58.4/40.4/32.1/27.0 BP=0.968 | | 36.58, 58.4/40.4/32.1/27.0 BP=0.968 | ||
|} | |} | ||
+ | * tried to tackle UNK using 36.58 M-NMT, increased vocab to 50000, got bleu=35.63, 58.6/40.0/31.6/26.4 BP=0.953 (not good, ?) | ||
+ | * training uy-zh, 50% zh-uy, 25% zh-uy | ||
+ | * training mem without EOS | ||
+ | * reviewing related papers | ||
|| | || | ||
2017年6月5日 (一) 06:01的版本
Date | People | Last Week | This Week | |||||||
---|---|---|---|---|---|---|---|---|---|---|
2017/6/5 | Jiyuan Zhang | |||||||||
Aodong LI |
Only make the English encoder's embedding constant -- 45.98 Only initialize the English encoder's embedding and then finetune it -- 46.06 Share the attention mechanism and then directly add them -- 46.20
Shrink output vocab from 30000 to 20000 and best result is 31.53 Train the model with 40 batch size and best result until now is 30.63 |
| ||||||||
Shiyue Zhang |
|
|||||||||
Shipan Ren |