Hidden in the Layers: Interpretation of Neural Networks for Natural Language Processing

Speaker:
David Mareček, Jindřich Libovický, Tomáš Musil, Rudolf Rosa and Tomasz Limisiewicz (ÚFAL MFF UK)
Abstract:
In recent years, deep neural networks dominated the area of Natural Language Processing (NLP). End-to-end-trained models can do tasks as skillfully as never before and develop their own language representations. However, they act as black boxes that are very hard to interpret. This calls for an inspection to what extent the linguistic conceptualizations are consistent with what the models learn. Do neural networks use morphology and syntax the way people do when they talk about language? Or do they develop their own way? In our talk, we will half-open the neural black-box and analyze the internal representations of input sentences with respect to their morphological, syntactic, and semantic properties. We will focus on word embeddings as well as contextual embeddings and self-attentions of the Transformer models (BERT, NMT). We will show both supervised and unsupervised analysis approaches.
Length:
01:06:58
Date:
30/11/2020
views:

Images:
Attachments: (video, slides, etc.)
63.9
82 downloads
136.6
95 downloads
2,3M
35 downloads