arXiv Analytics

Sign in

arXiv:1811.05544 [cs.CL]AbstractReferencesReviewsResources

An Introductory Survey on Attention Mechanisms in NLP Problems

Dichao Hu

Published 2018-11-12Version 1

First derived from human intuition, later adapted to machine translation for automatic token alignment, attention mechanism, a simple method that can be used for encoding sequence data based on the importance score each element is assigned, has been widely applied to and attained significant improvement in various tasks in natural language processing, including sentiment classification, text summarization, question answering, dependency parsing, etc. In this paper, we survey through recent works and conduct an introductory summary of the attention mechanism in different NLP problems, aiming to provide our readers with basic knowledge on this widely used method, discuss its different variants for different tasks, explore its association with other techniques in machine learning, and examine methods for evaluating its performance.

Related articles: Most relevant | Search more
arXiv:1906.09978 [cs.CL] (Published 2019-06-21)
Multilingual Named Entity Recognition Using Pretrained Embeddings, Attention Mechanism and NCRF
arXiv:2204.13353 [cs.CL] (Published 2022-04-28)
Attention Mechanism with Energy-Friendly Operations
Yu Wan et al.
arXiv:2012.02030 [cs.CL] (Published 2020-11-20)
Data-Informed Global Sparseness in Attention Mechanisms for Deep Neural Networks