This document discusses a mixture of experts (MoE) approach for reinforcement learning-based dialogue management. It introduces a MoE language model consisting of: (1) a primitive language model capable of generating diverse utterances, (2) several specialized expert models trained for different intents, and (3) a dialogue manager that selects utterances from the experts. The experts are constructed by training on labeled conversation data. Reinforcement learning is used to train the dialogue manager to optimize long-term dialogue quality by selecting among the expert utterances. Experiments demonstrate the MoE approach can generate more coherent and engaging conversations than single language models.
Related topics: