Improved Language Identification Through Cross-Lingual Self-Supervised Learning.

IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP)(2022)

引用 35|浏览128
暂无评分
摘要
Language identification greatly impacts the success of downstream tasks such as automatic speech recognition. Recently, self-supervised speech representations learned by wav2vec 2.0 have been shown to be very effective for a range of speech tasks. We extend previous self-supervised work on language identification by experimenting with pre-trained models which were learned on real-world unconstrained speech in multiple languages and not just on English. We show that models pre-trained on many languages perform better and enable language identification systems that require very little labeled data to perform well. Results on a 25 languages setup show that with only 10 minutes of labeled data per language, a cross-lingually pre-trained model can achieve over 93% accuracy.
更多
查看译文
关键词
Language identification,self-supervised learning,pre-training,multilingual,wav2vec
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要