An Auto-Encoder Matching Model for Learning Utterance-Level Semantic Dependency in Dialogue Generation. EMNLP 2018.

Abstract

Generating semantically coherent responses is still a major challenge in dialogue generation. Different from conventional text generation tasks, the mapping between inputs and responses in conversations is more complicated, which highly demands the understanding of utterance-level semantic dependency, a relation between the whole meanings of inputs and outputs. To address this problem, we propose an Auto-Encoder Matching (AEM) model to learn such dependency. The model contains two auto-encoders and one mapping module. The auto-encoders learn the semantic representations of inputs and responses, and the mapping module learns to connect the utterance-level representations. Experimental results from automatic and human evaluations demonstrate that our model is capable of generating responses of high coherence and fluency compared to baseline models.

Jingjing Xu (许晶晶)
Jingjing Xu (许晶晶)
Postdoc

My research interests include representation learning, multilingual learning, and green (energy efficient) deep learning.