Authors
Zhouhan Lin, Minwei Feng, Cicero Nogueira dos Santos, Mo Yu, Bing Xiang, Bowen Zhou, Yoshua Bengio
Publication date
2017/3/9
Journal
ICLR 2017: 5th International Conference on Learning Representations
Description
This paper proposes a new model for extracting an interpretable sentence embedding by introducing self-attention. Instead of using a vector, we use a 2-D matrix to represent the embedding, with each row of the matrix attending on a different part of the sentence. We also propose a self-attention mechanism and a special regularization term for the model. As a side effect, the embedding comes with an easy way of visualizing what specific parts of the sentence are encoded into the embedding. We evaluate our model on 3 different tasks: author profiling, sentiment classification, and textual entailment. Results show that our model yields a significant performance gain compared to other sentence embedding methods in all of the 3 tasks.
Total citations
2017201820192020202120222023202443218472512496452332171
Scholar articles
Z Lin, M Feng, CN Santos, M Yu, B Xiang, B Zhou… - arXiv preprint arXiv:1703.03130, 2017
Z Lin, M Feng, CN Santos, M Yu, B Xiang, B Zhou… - arXiv preprint arXiv:1703.03130, 2017