The document presents a white box approach called contextual decomposition for interpreting character-aware neural networks (CNNs) in the context of word-level prediction, focusing on morphological tagging. It demonstrates that learned patterns in neural networks often align with known linguistic rules, while also highlighting instances where these models may utilize alternative plausible patterns. The study involves analyzing contributions of specific characters to neural network outputs, using datasets from multiple languages and exploring architectural differences between CNNs and bidirectional LSTMs.
Related topics: