A Hierarchical Latent Variable Encoder-decoder Model For Generating Dialogues | Awesome LLM Papers Contribute to Awesome LLM Papers

A Hierarchical Latent Variable Encoder-decoder Model For Generating Dialogues

Iulian Vlad Serban, Alessandro Sordoni, Ryan Lowe, Laurent Charlin, Joelle Pineau, Aaron Courville, Yoshua Bengio . Proceedings of the AAAI Conference on Artificial Intelligence 2017 – 734 citations

[Paper]   Search on Google Scholar   Search on Semantic Scholar
AAAI Evaluation Model Architecture

Sequential data often possesses a hierarchical structure with complex dependencies between subsequences, such as found between the utterances in a dialogue. In an effort to model this kind of generative process, we propose a neural network-based generative architecture, with latent stochastic variables that span a variable number of time steps. We apply the proposed model to the task of dialogue response generation and compare it with recent neural network architectures. We evaluate the model performance through automatic evaluation metrics and by carrying out a human evaluation. The experiments demonstrate that our model improves upon recently proposed models and that the latent variables facilitate the generation of long outputs and maintain the context.

Similar Work