<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-04-18</id>
		<title>2014-04-18 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-04-18"/>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-04-18&amp;action=history"/>
		<updated>2026-04-14T11:24:10Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://cslt.org/mediawiki/index.php?title=2014-04-18&amp;diff=9727&amp;oldid=prev</id>
		<title>Cslt：以内容“==Resoruce Building== * quota on /nfs/disk this Saturday * release management should be started: Zhiyong * Blaster 0.1 &amp; vivian 0.0 system release  == Leftover question...”创建新页面</title>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-04-18&amp;diff=9727&amp;oldid=prev"/>
				<updated>2014-04-18T02:35:00Z</updated>
		
		<summary type="html">&lt;p&gt;以内容“==Resoruce Building== * quota on /nfs/disk this Saturday * release management should be started: Zhiyong * Blaster 0.1 &amp;amp; vivian 0.0 system release  == Leftover question...”创建新页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;==Resoruce Building==&lt;br /&gt;
* quota on /nfs/disk this Saturday&lt;br /&gt;
* release management should be started: Zhiyong&lt;br /&gt;
* Blaster 0.1 &amp;amp; vivian 0.0 system release&lt;br /&gt;
&lt;br /&gt;
== Leftover questions==&lt;br /&gt;
* Asymmetric window: Great improvement on training set(WER 34% to 24%), however the improvement is lost on test. Overfitting? &lt;br /&gt;
* Multi GPU training: Error encountered&lt;br /&gt;
* Multilanguage training&lt;br /&gt;
* Investigating LOUDS FST. &lt;br /&gt;
* CLG embedded decoder plus online compiler.&lt;br /&gt;
* DNN-GMM co-training&lt;br /&gt;
&lt;br /&gt;
== AM development ==&lt;br /&gt;
&lt;br /&gt;
=== Sparse DNN ===&lt;br /&gt;
* GA-based block sparsity&lt;br /&gt;
:* Found a paper in 2000 with similar ideas. &lt;br /&gt;
:* Try to get a student working on high performance computing to do the optimization&lt;br /&gt;
&lt;br /&gt;
===Noise training===&lt;br /&gt;
:* More experiments with no-noise&lt;br /&gt;
:* More experiments with additional noise types&lt;br /&gt;
&lt;br /&gt;
===AMR compression re-training===&lt;br /&gt;
&lt;br /&gt;
* 1700h MPE adaptation done&lt;br /&gt;
* 1700h stream mode adaptation runs into MPE4 done&lt;br /&gt;
* Stream model is better than non-stream wave&lt;br /&gt;
&lt;br /&gt;
===GFbank===&lt;br /&gt;
* GFBank Sinovoice test on 100h MPE&lt;br /&gt;
* Tencent 100h MPE training done&lt;br /&gt;
&lt;br /&gt;
===Multilingual ASR===&lt;br /&gt;
* all phone strategy baseline done&lt;br /&gt;
* Testing on Mandarin &amp;amp; English individually&lt;br /&gt;
&lt;br /&gt;
===Denoising &amp;amp; Farfield ASR===&lt;br /&gt;
*  re-Recording done&lt;br /&gt;
*  Prepare to construct the baseline&lt;br /&gt;
&lt;br /&gt;
===VAD===&lt;br /&gt;
&lt;br /&gt;
* Code ready, migrate to the VAD code framework&lt;br /&gt;
&lt;br /&gt;
===Scoring===&lt;br /&gt;
* g-score based on MLP is done&lt;br /&gt;
* t-score based on linear regression improves the performance&lt;br /&gt;
&lt;br /&gt;
==Word to Vector==&lt;br /&gt;
&lt;br /&gt;
* Dimension of low space varies from 10-100&lt;br /&gt;
* 8-thread word vector generation is 3 times faster than the LDA.&lt;br /&gt;
&lt;br /&gt;
==LM development==&lt;br /&gt;
&lt;br /&gt;
===NN LM===&lt;br /&gt;
&lt;br /&gt;
* Character-based NNLM (6700 chars, 7gram), 500M data training done.&lt;br /&gt;
:* Non-boundary char LM is better than boundary char LM&lt;br /&gt;
&lt;br /&gt;
* Investigate MS RNN LM training&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==QA==&lt;br /&gt;
&lt;br /&gt;
===FST-based matching===&lt;br /&gt;
:* Word-based FST 1-2 seconds with 1600 patterns. Huilan's implementation &amp;lt;1 second. ????&lt;br /&gt;
:* Char-FST Implementation is done. Not so effective.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Speech QA===&lt;br /&gt;
* Investigate determinization of G embedding&lt;/div&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	</feed>