Improving Sign Language Translation with Monolingual Data by Sign Back-Translation

2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021(2021)

引用 94|浏览161
暂无评分
摘要
Despite existing pioneering works on sign language translation (SLT), there is a non-trivial obstacle, i.e., the limited quantity of parallel sign-text data. To tackle this parallel data bottleneck, we propose a sign back-translation (SignBT) approach, which incorporates massive spoken language texts into SLT training. With a text-to-gloss translation model, we first back-translate the monolingual text to its gloss sequence. Then, the paired sign sequence is generated by splicing pieces from an estimated gloss-to-sign bank at the feature level. Finally, the synthetic parallel data serves as a strong supplement for the end-to-end training of the encoder-decoder SLT framework. To promote the SLT research, we further contribute CSL-Daily, a large-scale continuous SLT dataset. It provides both spoken language translations and gloss-level annotations. The topic revolves around people's daily lives (e.g., travel, shopping, medical care), the most likely SLT application scenario. Extensive experimental results and analysis of SLT methods are reported on CSL-Daily. With the proposed sign back-translation method, we obtain a substantial improvement over previous state-of-the-art SLT methods.
更多
查看译文
关键词
parallel data bottleneck,sign back-translation approach,massive spoken language texts,SLT training,text-to-gloss translation model,monolingual text,gloss sequence,paired sign sequence,gloss-to-sign bank,synthetic parallel data,end-to-end training,encoder-decoder SLT framework,SLT research,large-scale continuous SLT dataset,spoken language translations,gloss-level annotations,SLT application scenario,sign back-translation method,previous state-of-the-art SLT methods,sign language translation,monolingual data,parallel sign-text data
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要