arXiv Analytics

Sign in

arXiv:1708.00781 [cs.CL]AbstractReferencesReviewsResources

Dynamic Entity Representations in Neural Language Models

Yangfeng Ji, Chenhao Tan, Sebastian Martschat, Yejin Choi, Noah A. Smith

Published 2017-08-02Version 1

Understanding a long document requires tracking how entities are introduced and evolve over time. We present a new type of language model, EntityNLM, that can explicitly model entities, dynamically update their representations, and contextually generate their mentions. Our model is generative and flexible; it can model an arbitrary number of entities in context while generating each entity mention at an arbitrary length. In addition, it can be used for several different tasks such as language modeling, coreference resolution, and entity prediction. Experimental results with all these tasks demonstrate that our model consistently outperforms strong baselines and prior work.

Related articles: Most relevant | Search more
arXiv:1811.00998 [cs.CL] (Published 2018-11-02)
Analysing Dropout and Compounding Errors in Neural Language Models
arXiv:1901.00398 [cs.CL] (Published 2019-01-02)
Judge the Judges: A Large-Scale Evaluation Study of Neural Language Models for Online Review Generation
arXiv:1908.01817 [cs.CL] (Published 2019-07-22)
Sparsity Emerges Naturally in Neural Language Models