A Decomposable Attention Model for Natural Language Inference

@inproceedings{Parikh2016ADA,
  title={A Decomposable Attention Model for Natural Language Inference},
  author={Ankur P. Parikh and Oscar T{\"a}ckstr{\"o}m and Dipanjan Das and Jakob Uszkoreit},
  booktitle={EMNLP},
  year={2016}
}
We propose a simple neural architecture for natural language inference. Our approach uses attention to decompose the problem into subproblems that can be solved separately, thus making it trivially parallelizable. On the Stanford Natural Language Inference (SNLI) dataset, we obtain state-of-the-art results with almost an order of magnitude fewer parameters than previous work and without relying on any word-order information. Adding intra-sentence attention that takes a minimum amount of order… CONTINUE READING
Highly Influential
This paper has highly influenced 53 other papers. REVIEW HIGHLY INFLUENTIAL CITATIONS
Highly Cited
This paper has 322 citations. REVIEW CITATIONS
Related Discussions
This paper has been referenced on Twitter 91 times. VIEW TWEETS

From This Paper

Figures, tables, and topics from this paper.

Citations

Publications citing this paper.
Showing 1-10 of 235 extracted citations

Enhanced LSTM for Natural Language Inference

ACL • 2017
View 7 Excerpts
Highly Influenced

322 Citations

01002002016201720182019
Citations per Year
Semantic Scholar estimates that this publication has 322 citations based on the available data.

See our FAQ for additional information.

References

Publications referenced by this paper.
Showing 1-10 of 30 references

Adaptive Subgradient Methods for Online Learning and Stochastic Optimization

Journal of Machine Learning Research • 2011
View 2 Excerpts
Highly Influenced

Similar Papers

Loading similar papers…