Solving the Problem of Cascading Errors: Approximate Bayesian Inference for Linguistic Annotation Pipelines.

EMNLP '06: Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing(2006)

引用 79|浏览109
暂无评分
摘要
The end-to-end performance of natural language processing systems for compound tasks, such as question answering and textual entailment, is often hampered by use of a greedy 1-best pipeline architecture, which causes errors to propagate and compound at each stage. We present a novel architecture, which models these pipelines as Bayesian networks, with each low level task corresponding to a variable in the network, and then we perform approximate inference to find the best labeling. Our approach is extremely simple to apply but gains the benefits of sampling the entire distribution over labels at each stage in the pipeline. We apply our method to two tasks -- semantic role labeling and recognizing textual entailment -- and achieve useful performance gains from the superior pipeline architecture.
更多
查看译文
关键词
textual entailment,1-best pipeline architecture,novel architecture,superior pipeline architecture,compound task,end-to-end performance,useful performance gain,Bayesian network,approximate inference,entire distribution,approximate Bayesian inference,cascading error,linguistic annotation pipeline
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要