“2021-12-06”版本间的差异
来自cslt Wiki
(以“{| class="wikitable" !People !! This Week !! Next Week !! Task Tracking (<font color="red">DeadLine</font>) |- |- |Dong Wang || * || * || * |- |- |Yunqi Cai |...”为内容创建页面) |
|||
(12位用户的20个中间修订版本未显示) | |||
第5行: | 第5行: | ||
|Dong Wang | |Dong Wang | ||
|| | || | ||
− | * | + | * Refine spoof paper |
+ | * Prepare talk for information theory in NN | ||
+ | * Prepare talk for representation investigation. | ||
|| | || | ||
− | * | + | * Finish poof paper |
|| | || | ||
* | * | ||
第15行: | 第17行: | ||
|- | |- | ||
|Yunqi Cai | |Yunqi Cai | ||
− | || | + | || |
− | * | + | *review papers about CQDs |
+ | *Verify the deconvolution of infrared and visible faces | ||
+ | *Verify infrared and visible image fusion based on GLOW model | ||
+ | *Arrange research plans for interns | ||
|| | || | ||
* | * | ||
第27行: | 第32行: | ||
|Lantian Li | |Lantian Li | ||
|| | || | ||
− | * | + | * Finish course on AI. |
+ | * Study speaker separation and think about structural embedding. | ||
|| | || | ||
− | * | + | * Finish ETM response. |
+ | * Exps of hard trials. | ||
|| | || | ||
* | * | ||
第38行: | 第45行: | ||
|Ying Shi | |Ying Shi | ||
|| | || | ||
− | * | + | * Report about e2e kws |
+ | * speech engrave (garbage node, sil training data, text to speech attention) | ||
+ | * analyse fenyinta test data [[http://192.168.0.51:5555/cgi-bin/cvss/cvss_request.pl?account=shiying&step=view_request&cvssid=829 here]] | ||
|| | || | ||
− | * | + | * more analyse about speech engrave(speech to text attention) |
+ | * speech engrave (text to speech attention) | ||
|| | || | ||
* | * | ||
第49行: | 第59行: | ||
|Haoran Sun | |Haoran Sun | ||
|| | || | ||
− | * | + | * some tests on our model |
|| | || | ||
− | * | + | * make some more efficient attempts |
+ | * ——remove rhythm and pitch encoders | ||
+ | * ——increase distance between speakers | ||
+ | * ——improve content encoder | ||
+ | * ——make use of speaker label | ||
|| | || | ||
* | * | ||
第60行: | 第74行: | ||
|Chen Chen | |Chen Chen | ||
|| | || | ||
− | * | + | * pre-process audio data & train GAN with wav2vec2 output data directly |
|| | || | ||
− | * | + | * use kmeans and pca clustering wav2vec2 output to build better segment representation |
|| | || | ||
* | * | ||
第71行: | 第85行: | ||
|Pengqi Li | |Pengqi Li | ||
|| | || | ||
− | * | + | * reproduce a series of CAM method on speaker classification |
|| | || | ||
* | * | ||
第92行: | 第106行: | ||
|Weida Liang | |Weida Liang | ||
|| | || | ||
− | * | + | * Finish the first version on improved exemplar autoencoder with cycle loss |
+ | * Rethink the theory analysis part | ||
|| | || | ||
− | * | + | * Test on never-before-seen speaker conversion |
+ | * Review the code of wav2vec, StarGAN and PPG based GAN | ||
|| | || | ||
* | * | ||
第102行: | 第118行: | ||
|Zixi Yan | |Zixi Yan | ||
|| | || | ||
− | * | + | * Training wav2vec model |
|| | || | ||
* | * | ||
第112行: | 第128行: | ||
|Sirui Li | |Sirui Li | ||
|| | || | ||
− | * | + | * Fine-tune the wav2vec model |
|| | || | ||
− | * | + | * Comparing Tibetan and Chinese fine-tune results |
|| | || | ||
* | * | ||
第123行: | 第139行: | ||
|Haoyu Jiang | |Haoyu Jiang | ||
|| | || | ||
− | * | + | * Face sampling in CNCeleb dataset |
+ | * Filter videos without the target's face | ||
|| | || | ||
* | * | ||
第130行: | 第147行: | ||
|- | |- | ||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
− | |||
|- | |- | ||
|Renmiao Chen | |Renmiao Chen | ||
|| | || | ||
− | * | + | * Sample some audio,listen and analyze |
|| | || | ||
− | * | + | * divide data |
|| | || | ||
* | * |
2021年12月13日 (一) 10:57的最后版本
People | This Week | Next Week | Task Tracking (DeadLine) |
---|---|---|---|
Dong Wang |
|
|
|
Yunqi Cai |
|
|
|
Lantian Li |
|
|
|
Ying Shi |
|
|
|
Haoran Sun |
|
|
|
Chen Chen |
|
|
|
Pengqi Li |
|
|
|
Qingyang Zhu |
|
|
|
Weida Liang |
|
|
|
Zixi Yan |
|
|
|
Sirui Li |
|
|
|
Haoyu Jiang |
|
|
|
Renmiao Chen |
|
|
|