Interpreting Sequence-to-Sequence Models for Russian Inflectional Morphology

Proceedings of the Society for Computation in Linguistics (SCiL)(2020)

引用 1|浏览25
暂无评分
摘要
Morphological inflection, as an engineering task in NLP, has seen a rise in the use of neural sequence-to-sequence models (Kann and Schütze, 2016; Cotterell et al., 2018; Aharoni and Goldberg, 2017). While these outperform traditional systems based on edit rule induction, it is hard to interpret what they are learning in linguistic terms. We propose a new method of analyzing morphological sequence-to-sequence models which groups errors into linguistically meaningful classes, making what the model learns more transparent. As a case study, we analyze a seq2seq model on Russian, finding that semantic and lexically conditioned allomorphy (eg inanimate nouns like ZAVOD ‘factory’and animates like OTEC ‘father’have different, animacyconditioned accusative forms) are responsible for its relatively low accuracy. Augmenting the model with word embeddings as a proxy for lexical semantics leads to significant improvements in predicted wordform accuracy.
更多
查看译文
关键词
morphology,sequence-to-sequence
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要