<?xml version="1.0"?>
<?xml-stylesheet type="text/css" href="http://cslt.org/mediawiki/skins/common/feed.css?303"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="zh-cn">
		<id>http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-01-03</id>
		<title>2014-01-03 - 版本历史</title>
		<link rel="self" type="application/atom+xml" href="http://cslt.org/mediawiki/index.php?action=history&amp;feed=atom&amp;title=2014-01-03"/>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-01-03&amp;action=history"/>
		<updated>2026-04-15T05:13:01Z</updated>
		<subtitle>本wiki的该页面的版本历史</subtitle>
		<generator>MediaWiki 1.23.3</generator>

	<entry>
		<id>http://cslt.org/mediawiki/index.php?title=2014-01-03&amp;diff=8934&amp;oldid=prev</id>
		<title>Cslt：以内容“== AM development ==  === Sparse DNN ===  * Optimal Brain Damage(OBD).   # Online OBD held.  # OBD + L1 norm start to investigation.   * Efficient computing  # Conducti...”创建新页面</title>
		<link rel="alternate" type="text/html" href="http://cslt.org/mediawiki/index.php?title=2014-01-03&amp;diff=8934&amp;oldid=prev"/>
				<updated>2014-01-03T02:20:27Z</updated>
		
		<summary type="html">&lt;p&gt;以内容“== AM development ==  === Sparse DNN ===  * Optimal Brain Damage(OBD).   # Online OBD held.  # OBD + L1 norm start to investigation.   * Efficient computing  # Conducti...”创建新页面&lt;/p&gt;
&lt;p&gt;&lt;b&gt;新页面&lt;/b&gt;&lt;/p&gt;&lt;div&gt;== AM development ==&lt;br /&gt;
&lt;br /&gt;
=== Sparse DNN ===&lt;br /&gt;
&lt;br /&gt;
* Optimal Brain Damage(OBD). &lt;br /&gt;
&lt;br /&gt;
# Online OBD held. &lt;br /&gt;
# OBD + L1 norm start to investigation. &lt;br /&gt;
&lt;br /&gt;
* Efficient computing&lt;br /&gt;
&lt;br /&gt;
# Conducting rearrangement the matrix structure and compose zero blocks by some smart approaches, leading to better computing speed. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== Efficient DNN training ===&lt;br /&gt;
&lt;br /&gt;
# L1-L2 grid checking: L1/L2(&amp;lt; 1e-6) seems good for record1900 but worse for other test sets.&lt;br /&gt;
&lt;br /&gt;
[http://192.168.0.50:3000/series/?action=view&amp;amp;series=199,199.0,199.1,199.2,199.3,199.4,199.5,199.6,199.7,199.8,199.9 link here]&lt;br /&gt;
&lt;br /&gt;
# Asymmetric window: Great improvement on training set(WER 34% to 24%), however the improvement is lost on test. Overfitting? &lt;br /&gt;
# Frame-skipping. Skipping 1 frame speeds up decoding in a consistent way while retaining the accuracy largely. Skipping more frames lead to unacceptable performance degradation.&lt;br /&gt;
# Interpolation does not provide performance gain.&lt;br /&gt;
&lt;br /&gt;
[http://192.168.0.50:3000/series/?action=view&amp;amp;series=199,199.5,199.6,199.7,199.8,199.9,198,198.0,198.1,198.2,198.3,198.4,198.5,198.6 link here]&lt;br /&gt;
&lt;br /&gt;
=== Optimal phoneset===&lt;br /&gt;
&lt;br /&gt;
* Analyze Tencent English phone set. Found some errors in CH/EN phone sharing.&lt;br /&gt;
* Develop a new sharing scheme, start training the new system. &lt;br /&gt;
* Start training for all-separated phones&lt;br /&gt;
* Start training mixed system with Chinglish data.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Engine optimization===&lt;br /&gt;
&lt;br /&gt;
* Investigating LOUDS FST. On progress. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==LM development==&lt;br /&gt;
&lt;br /&gt;
===NN LM===&lt;br /&gt;
&lt;br /&gt;
* Collecting a bigger lexicon: 40k words related to music, 56k words from an official dictionary.&lt;br /&gt;
* Working on NN LM based on word2vector.&lt;br /&gt;
&lt;br /&gt;
==Embedded development==&lt;br /&gt;
&lt;br /&gt;
* Liuchao's cellphone, Qualcomm Snapdragon Krait MSM8960 @ 1.5GHz, using 1 core&lt;br /&gt;
small nnet 100/600/600/600/600/1264 with MFCC input&lt;br /&gt;
&lt;br /&gt;
* 4500 words:&lt;br /&gt;
&lt;br /&gt;
:* construct LG: 0.41s&lt;br /&gt;
:* compose HCLG with det: 13.70s, 5.318 MB&lt;br /&gt;
:* compose HCLG without det: 6.61s, 5.488 MB&lt;br /&gt;
&lt;br /&gt;
* 950 words:&lt;br /&gt;
&lt;br /&gt;
:* construct LG: 0.15s&lt;br /&gt;
:* compose HCLG with det: 2.63s, 0.947 MB, decode RT 0.649&lt;br /&gt;
:* compose HCLG without det: 1.74s, 0.998 MB, decode RT 0.548&lt;br /&gt;
&lt;br /&gt;
* For word list or simple grammars, determinization leads to small RT increase, but can improve HCLG compiling dramatically. This is particularly the case for embedded devices. &lt;br /&gt;
* The accuracy does not change with/without determinization.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Speech QA==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
* Use N-best to expand match in QA. Better performance were obtained.&lt;br /&gt;
:* 1-best matches 96/121 &lt;br /&gt;
:* 10-best matches 102/121&lt;br /&gt;
&lt;br /&gt;
* Use N-best to recover errors in entity check. Working on.&lt;br /&gt;
* Use Pinyin to recover errors in entity check. Future work.&lt;/div&gt;</summary>
		<author><name>Cslt</name></author>	</entry>

	</feed>