Mathematical Reasoning via Self-supervised Skip-tree Training

ICLR(2021)

引用 49|浏览41930
暂无评分
摘要
We demonstrate that self-supervised language modeling applied to mathematical formulas enables logical reasoning. To measure the logical reasoning abilities of language models, we formulate several evaluation (downstream) tasks, such as inferring types, suggesting missing assumptions and completing equalities. For training language models for formal mathematics, we propose a novel skip-tree task. We find that models trained on the skip-tree task show surprisingly strong mathematical reasoning abilities, and outperform models trained on standard skip-sequence tasks. We also analyze the models\u0027 ability to formulate new conjectures by measuring how often the predictions are provable and useful in other proofs.
更多
查看译文
关键词
self-supervised,skip-tree
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要