谷歌浏览器插件
订阅小程序
在清言上使用

BiVaSE: A Bilingual Variational Sentence Encoder with Randomly Initialized Transformer Layers

Acta linguistica academica(2022)

引用 0|浏览0
暂无评分
摘要
Transformer-based NLP models have achieved state-of-the-art results in many NLP tasks including text classification and text generation. However, the layers of these models do not output any explicit repre-sentations for texts units larger than tokens (e.g. sentences), although such representations are required to perform text classification. Sentence encodings are usually obtained by applying a pooling technique during fine-tuning on a specific task. In this paper, a new sentence encoder is introduced. Relying on an autoencoder architecture, it was trained to learn sentence representations from the very beginning of its training. The model was trained on bilingual data with variational Bayesian inference. Sentence repre-sentations were evaluated in downstream and linguistic probing tasks. Although the newly introduced encoder generally performs worse than well-known Transformer-based encoders, the experiments show that it was able to learn to incorporate linguistic information in the sentence representations.
更多
查看译文
关键词
sentence embedding,variational autoencoder,Transformer,pre-training,pooling
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要