<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-08-01</id>
		<title>2014-08-01 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-08-01"/>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-08-01&amp;action=history"/>
		<updated>2026-04-16T07:17:41Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://cslt.org/mediawiki/index.php?title=2014-08-01&amp;diff=10574&amp;oldid=prev</id>
		<title>2014年8月1日 (五) 01:53 Cslt</title>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-08-01&amp;diff=10574&amp;oldid=prev"/>
				<updated>2014-08-01T01:53:02Z</updated>
		
		<summary type="html">&lt;p&gt;&lt;/p&gt;
&lt;table class='diff diff-contentalign-left'&gt;
				&lt;col class='diff-marker' /&gt;
				&lt;col class='diff-content' /&gt;
				&lt;col class='diff-marker' /&gt;
				&lt;col class='diff-content' /&gt;
				&lt;tr style='vertical-align: top;'&gt;
				&lt;td colspan='2' style=&quot;background-color: white; color:black; text-align: center;&quot;&gt;←上一版本&lt;/td&gt;
				&lt;td colspan='2' style=&quot;background-color: white; color:black; text-align: center;&quot;&gt;2014年8月1日 (五) 01:53的版本&lt;/td&gt;
				&lt;/tr&gt;&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;第33行：&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;第33行：&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* By tuning parameters of late-response lag &amp;amp; response time, obtained performance improvement with Lasso.&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* By tuning parameters of late-response lag &amp;amp; response time, obtained performance improvement with Lasso.&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt;&amp;#160;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;pre&amp;gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Simulation results: Baseline:&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;Simulation results: Baseline:&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;--------------------------------------------------------------------------&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;--------------------------------------------------------------------------&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;第68行：&lt;/td&gt;
&lt;td colspan=&quot;2&quot; class=&quot;diff-lineno&quot;&gt;第69行：&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;--------------------------------------------------------------------------&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;--------------------------------------------------------------------------&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td colspan=&quot;2&quot;&gt;&amp;#160;&lt;/td&gt;&lt;td class='diff-marker'&gt;+&lt;/td&gt;&lt;td style=&quot;color:black; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #a3d3ff; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;&lt;ins style=&quot;font-weight: bold; text-decoration: none;&quot;&gt;&amp;lt;/pre&amp;gt;&lt;/ins&gt;&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;tr&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* Adaptation under running&lt;/div&gt;&lt;/td&gt;&lt;td class='diff-marker'&gt;&amp;#160;&lt;/td&gt;&lt;td style=&quot;background-color: #f9f9f9; color: #333333; font-size: 88%; border-style: solid; border-width: 1px 1px 1px 4px; border-radius: 0.33em; border-color: #e6e6e6; vertical-align: top; white-space: pre-wrap;&quot;&gt;&lt;div&gt;* Adaptation under running&lt;/div&gt;&lt;/td&gt;&lt;/tr&gt;
&lt;/table&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	<entry>
		<id>http://cslt.org/mediawiki/index.php?title=2014-08-01&amp;diff=10573&amp;oldid=prev</id>
		<title>Cslt：以内容“==Resoruce Building==  == Leftover questions==  * Investigating LOUDS FST.  * CLG embedded decoder plus online compiler. * DNN-GMM co-training * NN LM  == AM developmen...”创建新页面</title>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-08-01&amp;diff=10573&amp;oldid=prev"/>
				<updated>2014-08-01T01:52:34Z</updated>
		
		<summary type="html">&lt;p&gt;以内容“==Resoruce Building==  == Leftover questions==  * Investigating LOUDS FST.  * CLG embedded decoder plus online compiler. * DNN-GMM co-training * NN LM  == AM developmen...”创建新页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;==Resoruce Building==&lt;br /&gt;
&lt;br /&gt;
== Leftover questions==&lt;br /&gt;
&lt;br /&gt;
* Investigating LOUDS FST. &lt;br /&gt;
* CLG embedded decoder plus online compiler.&lt;br /&gt;
* DNN-GMM co-training&lt;br /&gt;
* NN LM&lt;br /&gt;
&lt;br /&gt;
== AM development ==&lt;br /&gt;
&lt;br /&gt;
=== Sparse DNN ===&lt;br /&gt;
* WJS sparse DNN shows a slightly better than non-sparse cases when the network is in a large scale&lt;br /&gt;
* Pre-training does work for DNN training (for both 4/5/6 layers)&lt;br /&gt;
&lt;br /&gt;
===Noise training===&lt;br /&gt;
:* Journal paper writing on going&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Multilingual ASR===&lt;br /&gt;
&lt;br /&gt;
* Native English speaker + Chinglish speaker obtained better performance.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Drop out &amp;amp; convolutional network==&lt;br /&gt;
&lt;br /&gt;
* Change learning to 0.001, the training process can be started.&lt;br /&gt;
* Frame Accuracy goes to :  (with/without drop probability normalization)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Denoising &amp;amp; Farfield ASR===&lt;br /&gt;
&lt;br /&gt;
* By tuning parameters of late-response lag &amp;amp; response time, obtained performance improvement with Lasso.&lt;br /&gt;
&lt;br /&gt;
Simulation results: Baseline:&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                model/test              |  far_evl92   |   near_evl92&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                   clean_ce             |     59.38    |    19.25&lt;br /&gt;
                 mpe_clean_ce           |     40.46    |    12.94&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
&lt;br /&gt;
Lasso with optimal parameters(lambda=0.05, delta=5, N=10)&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                model/test              |  far_evl92   |   near_evl92&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                clean_ce                |     54.63    |    15.75&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
              mpe_clean_ce              |     36.58    |    11.64&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
&lt;br /&gt;
Real data results:&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                   model/test              |  far_evl92   |   near_evl92&lt;br /&gt;
 --------------------------------------------------------------------------&lt;br /&gt;
                   clean_ce                |     94.86    |    63.48&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                 mpe_clean_ce              |     92.29    |    58.37&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
dereverberated recording :&lt;br /&gt;
&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                   model/test              |  far_evl92   |   near_evl92&lt;br /&gt;
 --------------------------------------------------------------------------&lt;br /&gt;
                   clean_ce                |     94.91    |    61.03&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
                 mpe_clean_ce              |     91.28    |    54.16&lt;br /&gt;
--------------------------------------------------------------------------&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Adaptation under running&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===VAD===&lt;br /&gt;
&lt;br /&gt;
* Waiting for testing results&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Scoring===&lt;br /&gt;
&lt;br /&gt;
* Refine the acoustic model with AMIDA database. problem solved by involving both wsj and AMIDA.&lt;br /&gt;
&lt;br /&gt;
===Confidence===&lt;br /&gt;
&lt;br /&gt;
* Be familiar with Kaldi&lt;br /&gt;
* Need to extract lattice and DNN features&lt;br /&gt;
&lt;br /&gt;
===Embedded decoder===&lt;br /&gt;
&lt;br /&gt;
* Chatting LM released (80k)&lt;br /&gt;
* Train two smaller network: 500x4+600, 400x4+500: on going&lt;br /&gt;
* Need to upload the new client code onto git (+)&lt;br /&gt;
* Build a new graph with MPE3 am and chatting LM.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==LM development==&lt;br /&gt;
&lt;br /&gt;
===Domain specific LM===&lt;br /&gt;
&lt;br /&gt;
h2. Domain specific LM construction&lt;br /&gt;
&lt;br /&gt;
h3. TAG LM&lt;br /&gt;
&lt;br /&gt;
* TAG obtained better performance&lt;br /&gt;
&lt;br /&gt;
h3. Chatting LM&lt;br /&gt;
&lt;br /&gt;
* First version released (80k lexicon)&lt;br /&gt;
* Prepare 2nd released (120k lexicon)&lt;br /&gt;
* Test on Xiaotang long &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Word2Vector==&lt;br /&gt;
&lt;br /&gt;
===W2V based doc classification===&lt;br /&gt;
&lt;br /&gt;
* Initial results variable Bayesian GMM obtained. Performance is not as good as the conventional GMM.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Speaker ID==&lt;br /&gt;
&lt;br /&gt;
* Full-data SRE trial goes into the final stage&lt;br /&gt;
* results will be ready soon&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Translation==&lt;br /&gt;
* collecting more data (Xinhua parallel text, bible, name entity)  for the second version&lt;br /&gt;
* check possible parameters to control phrase pair lexicon&lt;/div&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	</feed>