arXiv Analytics

Sign in

arXiv:2109.04912 [cs.CL]AbstractReferencesReviewsResources

ReasonBERT: Pre-trained to Reason with Distant Supervision

Xiang Deng, Yu Su, Alyssa Lees, You Wu, Cong Yu, Huan Sun

Published 2021-09-10Version 1

We present ReasonBert, a pre-training method that augments language models with the ability to reason over long-range relations and multiple, possibly hybrid contexts. Unlike existing pre-training methods that only harvest learning signals from local contexts of naturally occurring texts, we propose a generalized notion of distant supervision to automatically connect multiple pieces of text and tables to create pre-training examples that require long-range reasoning. Different types of reasoning are simulated, including intersecting multiple pieces of evidence, bridging from one piece of evidence to another, and detecting unanswerable cases. We conduct a comprehensive evaluation on a variety of extractive question answering datasets ranging from single-hop to multi-hop and from text-only to table-only to hybrid that require various reasoning capabilities and show that ReasonBert achieves remarkable improvement over an array of strong baselines. Few-shot experiments further demonstrate that our pre-training method substantially improves sample efficiency.

Comments: Accepted to EMNLP'2021. Our code and pre-trained models are available at https://github.com/sunlab-osu/ReasonBERT
Categories: cs.CL, cs.AI, cs.LG
Related articles: Most relevant | Search more
arXiv:2212.07699 [cs.CL] (Published 2022-12-15)
Retrieval-based Disentanglement with Distant Supervision
arXiv:1511.06219 [cs.CL] (Published 2015-11-19)
Knowledge Base Population using Semantic Label Propagation
arXiv:2205.08770 [cs.CL] (Published 2022-05-18)
Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision