“14-10-19 Dongxu Zhang”版本间的差异

来自cslt Wiki
跳转至: 导航搜索
(以“Last week ---------------------- 1. Train LSTM-Rnn LM with 200MB corpus(vocabulary 10k, classes 100). when using 2 kernels, it takes aroung 200min per epoch. 2. Trai...”为内容创建页面)
 
第1行: 第1行:
Last week
+
=== Accomplished this week ===
----------------------
+
* 1. Train LSTM-Rnn LM with 200MB corpus(vocabulary 10k, classes 100). when using 2 kernels, it takes aroung 200min per epoch.
1. Train LSTM-Rnn LM with 200MB corpus(vocabulary 10k, classes 100). when using 2 kernels, it takes aroung 200min per epoch.
+
* 2. Train 5-gram LM using Baiduzhidao_corpus(~30GB after preprocess) with new lexicon. There is a mistake when counted possiblity after merge.
2. Train 5-gram LM using Baiduzhidao_corpus(~30GB after preprocess) with new lexicon. There is a mistake when counted possiblity after merge.
+
* 3. An idea occured to me which may improve word2vec with much more semantic information. But there is huge computation complexity problem that bothers me, which I wish we can discuss.
3. An idea occured to me which may improve word2vec with much more semantic information. But there is huge computation complexity problem that bothers me, which I wish we can discuss.
+
* 4. Read paper "Learning Long-Term Dependencies with Gradient Descent is Difficult". Still in progress.
4. Read paper "Learning Long-Term Dependencies with Gradient Descent is Difficult". Still in progress.
+
  
This week
+
=== Next week ===
----------------------
+
* 1. Test LSTM-Rnn LM.
1. Test LSTM-Rnn LM.
+
* 2. Finished building lexion.
2. Finished building lexion.
+
* 3. Understand the paper.
3. Understand the paper.
+
* 4. May have time to achieve my baseline idea on text8.
4. May have time to achieve my baseline idea on text8.
+

2014年10月19日 (日) 12:42的版本

Accomplished this week

  • 1. Train LSTM-Rnn LM with 200MB corpus(vocabulary 10k, classes 100). when using 2 kernels, it takes aroung 200min per epoch.
  • 2. Train 5-gram LM using Baiduzhidao_corpus(~30GB after preprocess) with new lexicon. There is a mistake when counted possiblity after merge.
  • 3. An idea occured to me which may improve word2vec with much more semantic information. But there is huge computation complexity problem that bothers me, which I wish we can discuss.
  • 4. Read paper "Learning Long-Term Dependencies with Gradient Descent is Difficult". Still in progress.

Next week

  • 1. Test LSTM-Rnn LM.
  • 2. Finished building lexion.
  • 3. Understand the paper.
  • 4. May have time to achieve my baseline idea on text8.