The document presents a neural network model for dialog act classification that incorporates context representations. It uses a CNN to represent each utterance, applies an internal attention mechanism, and models context with RNNs. As baselines, it uses a single-utterance CNN and concatenation of utterances. Results show RNNs better learn context representations and attention mechanisms improve performance, though the optimal attention placement depends on the dataset. The best-performing models outperform the previous state-of-the-art on benchmark datasets.
Related topics: