BIBAL, Adrien
[UCL]
Cardon, Rémi
[UCL]
Alfter, David
[UCL]
Souza Wilkens, Rodrigo
[UCL]
Wang, Xiaoou
[UCL]
François, Thomas
[UCL]
Watrin, Patrick
[UCL]
The performance of deep learning models in NLP and other fields of machine learning has led to a rise in their popularity, and so the need for explanations of these models becomes paramount. Attention has been seen as a solution to increase performance, while providing some explanations. However, a debate has started to cast doubt on the explanatory power of attention in neural networks. Although the debate has created a vast literature thanks to contributions from various areas, the lack of communication is becoming more and more tangible. In this paper, we provide a clear overview of the insights on the debate by critically confronting works from these different areas. This holistic vision can be of great interest for future works in all the communities concerned by this debate. We sum up the main challenges spotted in these areas, and we conclude by discussing the most promising future avenues on attention as an explanation.
Bibliographic reference |
BIBAL, Adrien ; Cardon, Rémi ; Alfter, David ; Souza Wilkens, Rodrigo ; Wang, Xiaoou ; et. al. Is Attention Explanation? An Introduction to the Debate.Annual Meeting of the Association for Computational Linguistics (Dublin, Ireland, du 22/05/2022 au 27/05/2022). In: Association for Computational Linguistics. Annual Meeting. Conference Proceedings, p. 3889–3900 (2022) |
Permanent URL |
http://hdl.handle.net/2078.1/260014 |