|
|
第80行: |
第80行: |
| |Xiaolou Li | | |Xiaolou Li |
| || | | || |
− | * | + | * Reproduce different structure |
| + | ** Baseline training with less data |
| + | ** Code write and debug |
| + | *** ResNet3D, Branchformer, E-Branchformer, interCTC |
| + | * Paper Reading: Some VSR paper in ICASSP2024 |
| || | | || |
− | * | + | * Experiment in different structure |
| || | | || |
| * | | * |
第91行: |
第95行: |
| |Zehua Liu | | |Zehua Liu |
| || | | || |
− | * | + | * ASR training for model distillation |
| + | * Paper Reading |
| || | | || |
| * | | * |
People |
This Week |
Next Week |
Task Tracking (DeadLine)
|
Dong Wang
|
- Design learn not to listen, to reduce false alarms for KWS [1]
- Design AI courses for primary and middle schools
- Finaly review of NeuralMag paper [2]
- Rebutal for IJCAI paper [3]
|
|
|
Lantian Li
|
|
|
|
Ying Shi
|
- Experimental Result for SPL Paper: Phrase Guided End-to-End Target Sentence Extraction from Overlapping Speech
- Data Preparation
- Chinese Overlap ASR model
- Detect wake-up words from Continuous speech
|
- Finish draft for SPL paper
- Finish model training
|
|
Zhenghai You
|
- A matching encoder experiment in cohort[4]
|
|
|
Junming Yuan
|
- Experimental report on "learn not to listen"[5]
|
|
|
Chen Chen
|
- entropy analysis
- group work [6]
|
|
|
Xiaolou Li
|
- Reproduce different structure
- Baseline training with less data
- Code write and debug
- ResNet3D, Branchformer, E-Branchformer, interCTC
- Paper Reading: Some VSR paper in ICASSP2024
|
- Experiment in different structure
|
|
Zehua Liu
|
- ASR training for model distillation
- Paper Reading
|
|
|
Pengqi Li
|
- Review of speech processing XAI for NSFC
- download 160+ papers(113 about speech processing XAI; Traditional XAI method; review)
- Summarize them using LLM and categorize by speech processing task.
|
|
|
Wan Lin
|
|
|
|
Tianhao Wang
|
- Neural scoring docs and codes reviewing
|
|
|
Zhenyu Zhou
|
|
|
|
Junhui Chen
|
- Neural scoring: mix/overlap/concat test
|
|
|
Jiaying Wang
|
- one cohort distance test(The probability of selecting the source closer to the cohort as the target during testing)
- the rate is still around 0.5
- speakerbeam with no enroll/cohort + minimal loss training
- double-check done
- still training, but seem to converge at val_loss around -3
- confused
|
|
|
Yu Zhang
|
- Financial Backtest indicators check
- Jun Wang R2 SAC codes and papers reading
|
- Reproduce R2 SAC and FinRL policy
|
|
Wenqiang Du
|
- update CN KWS model for AIbabel
- Using real environment and FA data to update the model
|
|
|
Yang Wei
|
- Prepare for children mispronunciation detection and diagnosis base model
|
|
|
Lily
|
- Paper reading and prepare for journal paper
- Data annotation (for perception)
|
|
|
Turi
|
- Data Collection App Backend [7]
- User authentication
- Data storage
|
|
|