Corpus ID: 6126582

Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models

@inproceedings{Serban2016BuildingED,
  title={Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models},
  author={Iulian Serban and Alessandro Sordoni and Yoshua Bengio and Aaron C. Courville and Joelle Pineau},
  booktitle={AAAI},
  year={2016}
}
We investigate the task of building open domain, conversational dialogue systems based on large dialogue corpora using generative models. Generative models produce system responses that are autonomously generated word-by-word, opening up the possibility for realistic, flexible interactions. In support of this goal, we extend the recently proposed hierarchical recurrent encoder-decoder neural network to the dialogue domain, and demonstrate that this model is competitive with state-of-the-art… Expand
Generative Deep Neural Networks for Dialogue: A Short Review
TLDR
Recently proposed models based on generative encoder-decoder neural network architectures are reviewed and it is shown that these models have better ability to incorporate long-term dialogue history, to model uncertainty and ambiguity in dialogue, and to generate responses with high-level compositional structure. Expand
Hierarchical Variational Memory Network for Dialogue Generation
TLDR
A novel hierarchical variational memory network (HVMN) is proposed, by adding the hierarchical structure and the variationalMemory network into a neural encoder-decoder network that can capture both the high-level abstract variations and long-term memories during dialogue tracking, which enables the random access of relevant dialogue histories. Expand
End-to-End latent-variable task-oriented dialogue system with exact log-likelihood optimization
TLDR
An end-to-end dialogue model based on a hierarchical encoder-decoder, which employed a discrete latent variable to learn underlying dialogue intentions, which argues that the latent discrete variable interprets the intentions that guide machine responses generation. Expand
Knowledge Grounded Pre-Trained Model For Dialogue Response Generation
TLDR
A novel pre-training based encoder-decoder model which can enhance the multiturn dialogue response generation by incorporating external textual knowledge and a multi-head attention based decoder is designed to incorporate the semantic information from both knowledge and dialogue hidden representations into decoding process to generate informative and proper dialogue responses. Expand
An Alternative Deep Model for Turn-Based Conversations
TLDR
This paper proposes a new model for building a conversational dialogue system which provides natural, realistic and flexible interaction between human and machine based on large movie subtitles dataset and applies an attention mechanism in order to attend to particular sentences on source side when predicting a turn response. Expand
A Hierarchical Latent Variable Encoder-Decoder Model for Generating Dialogues
TLDR
A neural network-based generative architecture, with latent stochastic variables that span a variable number of time steps, that improves upon recently proposed models and that the latent variables facilitate the generation of long outputs and maintain the context. Expand
Hierarchical Text Generation and Planning for Strategic Dialogue
TLDR
An approach to learning representations of messages in dialogues by maximizing the likelihood of subsequent sentences and actions, which decouples the semantics of the dialogue utterance from its linguistic realization, which outperforms previous work both linguistically and strategically. Expand
Memory-to-Sequence learning with LSTM joint decoding for task-oriented dialogue systems
  • Bing Yu, F. Ren, Yanwei Bao
  • Computer Science
  • 2019 14th IEEE Conference on Industrial Electronics and Applications (ICIEA)
  • 2019
TLDR
A Memory-to-Sequence framework that uses Memory Neural Network (MemNN) and Long Short Term Memory (LSTM) joint decoding, which can better capture the dependence between system response and knowledge base is proposed. Expand
Encoder-Memory-Decoder Model for Long Conversation Dialogue System
TLDR
An Encoder-Memory-Decoder model is proposed to build long conversations dialogue system in neural generative way, viewed as an end-to-end neural network model equipped with memory ability to memorize the dialogue history information for generative dialogue response. Expand
A Dual Encoder Sequence to Sequence Model for Open-Domain Dialogue Modeling
TLDR
This work injects latent context representations into a sequence to sequence neural network in the form of dialog acts using a second encoder to enhance the quality and the coherence of the conversations generated. Expand
...
1
2
3
4
5
...

References

SHOWING 1-10 OF 56 REFERENCES
A Diversity-Promoting Objective Function for Neural Conversation Models
TLDR
This work proposes using Maximum Mutual Information (MMI) as the objective function in neural models, and demonstrates that the proposed MMI models produce more diverse, interesting, and appropriate responses, yielding substantive gains in BLEU scores on two conversational datasets and in human evaluations. Expand
Word-Based Dialog State Tracking with Recurrent Neural Networks
TLDR
A new wordbased tracking method which maps directly from the speech recognition results to the dialog state without using an explicit semantic decoder is presented, based on a recurrent neural network structure which is capable of generalising to unseen dialog state hypotheses, and which requires very little feature engineering. Expand
Joint Language and Translation Modeling with Recurrent Neural Networks
TLDR
This work presents a joint language and translation model based on a recurrent neural network which predicts target words based on an unbounded history of both source and target words which shows competitive accuracy compared to the traditional channel model features. Expand
A Neural Conversational Model
TLDR
A simple approach to conversational modeling which uses the recently proposed sequence to sequence framework, and is able to extract knowledge from both a domain specific dataset, and from a large, noisy, and general domain dataset of movie subtitles. Expand
Neural Responding Machine for Short-Text Conversation
TLDR
Empirical study shows that NRM can generate grammatically correct and content-wise appropriate responses to over 75% of the input text, outperforming state-of-the-arts in the same setting, including retrieval-based and SMT-based models. Expand
Deep Neural Network Approach for the Dialog State Tracking Challenge
TLDR
The paper explores some aspects of the training, and the resulting tracker is found to perform competitively, particularly on a corpus of dialogs from a system not found in the training. Expand
POMDP-Based Statistical Spoken Dialog Systems: A Review
TLDR
This review article provides an overview of the current state of the art in the development of POMDP-based spoken dialog systems. Expand
Sequence to Sequence Learning with Neural Networks
TLDR
This paper presents a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure, and finds that reversing the order of the words in all source sentences improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier. Expand
A Neural Network Approach to Context-Sensitive Generation of Conversational Responses
TLDR
A neural network architecture is used to address sparsity issues that arise when integrating contextual information into classic statistical models, allowing the system to take into account previous dialog utterances. Expand
Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation
TLDR
Qualitatively, the proposed RNN Encoder‐Decoder model learns a semantically and syntactically meaningful representation of linguistic phrases. Expand
...
1
2
3
4
5
...