<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-04-11</id>
		<title>2014-04-11 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-04-11"/>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-04-11&amp;action=history"/>
		<updated>2026-04-14T11:24:10Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://cslt.org/mediawiki/index.php?title=2014-04-11&amp;diff=9658&amp;oldid=prev</id>
		<title>Cslt：以内容“==Resoruce Building== * Current text resource has been re-arranged and listed  == Leftover questions== * Asymmetric window: Great improvement on training set(WER 34% to...”创建新页面</title>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-04-11&amp;diff=9658&amp;oldid=prev"/>
				<updated>2014-04-11T02:13:59Z</updated>
		
		<summary type="html">&lt;p&gt;以内容“==Resoruce Building== * Current text resource has been re-arranged and listed  == Leftover questions== * Asymmetric window: Great improvement on training set(WER 34% to...”创建新页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;==Resoruce Building==&lt;br /&gt;
* Current text resource has been re-arranged and listed&lt;br /&gt;
&lt;br /&gt;
== Leftover questions==&lt;br /&gt;
* Asymmetric window: Great improvement on training set(WER 34% to 24%), however the improvement is lost on test. Overfitting? &lt;br /&gt;
* Multi GPU training: Error encountered&lt;br /&gt;
* Multilanguage training&lt;br /&gt;
* Investigating LOUDS FST. &lt;br /&gt;
* CLG embedded decoder plus online compiler.&lt;br /&gt;
* DNN-GMM co-training&lt;br /&gt;
&lt;br /&gt;
== AM development ==&lt;br /&gt;
&lt;br /&gt;
=== Sparse DNN ===&lt;br /&gt;
* GA-based block sparsity&lt;br /&gt;
:* Found a paper in 2000 with similar ideas. &lt;br /&gt;
:* Try to get a student working on high performance computing to do the optimization&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Noise training===&lt;br /&gt;
:* More experiments with no-noise&lt;br /&gt;
:* More experiments with additional noise types&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===AMR compression re-training===&lt;br /&gt;
&lt;br /&gt;
* 1700h MPE adaptation done&lt;br /&gt;
* 1700h stream mode adaptation runs into MPE1&lt;br /&gt;
&lt;br /&gt;
===GFbank===&lt;br /&gt;
* Significant improvement found with GFBank &lt;br /&gt;
* Significant improvement found with FBank + GFBank&lt;br /&gt;
&lt;br /&gt;
===Denoising &amp;amp; Farfield ASR===&lt;br /&gt;
*  Recording done&lt;br /&gt;
*  Prepare to construct the baseline&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===VAD===&lt;br /&gt;
&lt;br /&gt;
* Code ready, need to figure out speech/no-speech smooth&lt;br /&gt;
&lt;br /&gt;
===Farfield recognition===&lt;br /&gt;
&lt;br /&gt;
===Scoring===&lt;br /&gt;
&lt;br /&gt;
* g-score based on MLP is done&lt;br /&gt;
* t-score based on linear regression improves the performance&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Word to Vector==&lt;br /&gt;
&lt;br /&gt;
* LDA baseline (sogou 1700*9 training set) done&lt;br /&gt;
* Wordvector classification is much better than the LDA system&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
word vector: &lt;br /&gt;
           general: dict - 15w;   train_data - ren_ming_ri_bao(5g);  windows-5&lt;br /&gt;
           1. size - 50  time=30m 12thread&lt;br /&gt;
           2. size - 10  time=10m 12thread&lt;br /&gt;
&lt;br /&gt;
data: class_num=9  document_num=9*2000&lt;br /&gt;
      train_num =9*1600&lt;br /&gt;
      test_num  =9*200&lt;br /&gt;
       dev_num  =9*200&lt;br /&gt;
&lt;br /&gt;
train_set:&lt;br /&gt;
                   C000008  C000010 C000013 C000014 C000016 C000020 C000022 C000023 C000024     total &lt;br /&gt;
                       财经    IT      健康     体育      旅游   教育      招聘     文化    军事   &lt;br /&gt;
       lda_inf      0.845   0.2756  0.698   0.9502   0.63499  0.32   0.8080   0.3505 0.864    0.6385&lt;br /&gt;
      lda_inf_10    0.8149  0.0887  0.628   0.9641   0.5739   0.105  0.707363 0.2334 0.8628   0.553167&lt;br /&gt;
 w2v_filter_filer   0.7463  0.713   0.657   0.9106   0.68659  0.54   0.74638  0.692  0.84518  0.72638&lt;br /&gt;
w2v_filter_filer_10 0.7608  0.4323  0.57394 0.865    0.549    0.335  0.577    0.6129 0.78099  0.609769&lt;br /&gt;
&lt;br /&gt;
test_set:&lt;br /&gt;
                     C000008  C000010 C000013 C000014 C000016 C000020 C000022 C000023 C000024   total &lt;br /&gt;
                       财经    IT      健康     体育      旅游   教育      招聘     文化    军事  &lt;br /&gt;
w2v_filter_filter    0.6865   0.7263   0.6716  0.84577 0.7462  0.46268 0.6567  0.7114  0.8905    0.71088&lt;br /&gt;
w2v_filter_filter_10 0.791    0.4079   0.56218 0.74129 0.62189 0.22885 0.562   0.6766  0.84079   0.603648&lt;br /&gt;
    lda_inf          0.8706   0.26368  0.6965  0.8009  0.582   0.2537  0.72139 0.3184  0.82587   0.59259&lt;br /&gt;
   lda_inf_10        0.776    0.1044   0.6467  0.9054  0.62189 0.1144  0.56218 0.24378 0.796     0.530127&lt;br /&gt;
&lt;br /&gt;
note:w2v_filter--remove the stop word in traing word vector&lt;br /&gt;
note:w2v_filter_filter  -- remove the stop word in traing word vector and remove the documnet stop words&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
==LM development==&lt;br /&gt;
&lt;br /&gt;
===NN LM===&lt;br /&gt;
&lt;br /&gt;
* Character-based NNLM (6700 chars, 7gram), 500M data training done.&lt;br /&gt;
:* Non-boundary char LM is better than boundary char LM&lt;br /&gt;
&lt;br /&gt;
* Investigate MS RNN LM training&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==QA==&lt;br /&gt;
&lt;br /&gt;
===FST-based matching===&lt;br /&gt;
:* Word-based FST 1-2 seconds with 1600 patterns. Huilan's implementation &amp;lt;1 second. ?????&lt;br /&gt;
:* Char-FST Implementation is done. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Speech QA===&lt;br /&gt;
* Investigate determinization of G embedding&lt;/div&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	</feed>