Bayesian synchronous tree-substitution grammar induction and its application to sentence compression

ACL(2010)

引用 35|浏览19
暂无评分
摘要
We describe our experiments with training algorithms for tree-to-tree synchronous tree-substitution grammar (STSG) for monolingual translation tasks such as sentence compression and paraphrasing. These translation tasks are characterized by the relative ability to commit to parallel parse trees and availability of word alignments, yet the unavailability of large-scale data, calling for a Bayesian tree-to-tree formalism. We formalize nonparametric Bayesian STSG with epsilon alignment in full generality, and provide a Gibbs sampling algorithm for posterior inference tailored to the task of extractive sentence compression. We achieve improvements against a number of baselines, including expectation maximization and variational Bayes training, illustrating the merits of nonparametric inference over the space of grammars as opposed to sparse parametric inference with a fixed grammar.
更多
查看译文
关键词
fixed grammar,bayesian tree-to-tree formalism,nonparametric inference,bayesian synchronous tree-substitution grammar,training algorithm,nonparametric bayesian stsg,parametric inference,extractive sentence compression,sentence compression,monolingual translation task,posterior inference,grammar induction
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要