Nevin L. Zhang
2020
Response-Anticipated Memory for On-Demand Knowledge Integration in Response Generation
Zhiliang Tian
|
Wei Bi
|
Dongkyu Lee
|
Lanqing Xue
|
Yiping Song
|
Xiaojiang Liu
|
Nevin L. Zhang
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Neural conversation models are known to generate appropriate but non-informative responses in general. A scenario where informativeness can be significantly enhanced is Conversing by Reading (CbR), where conversations take place with respect to a given external document. In previous work, the external document is utilized by (1) creating a context-aware document memory that integrates information from the document and the conversational context, and then (2) generating responses referring to the memory. In this paper, we propose to create the document memory with some anticipated responses in mind. This is achieved using a teacher-student framework. The teacher is given the external document, the context, and the ground-truth response, and learns how to build a response-aware document memory from three sources of information. The student learns to construct a response-anticipated document memory from the first two sources, and teacher’s insight on memory creation. Empirical results show that our model outperforms the previous state-of-the-art for the CbR task.
2019
Learning to Abstract for Memory-augmented Conversational Response Generation
Zhiliang Tian
|
Wei Bi
|
Xiaopeng Li
|
Nevin L. Zhang
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics
Neural generative models for open-domain chit-chat conversations have become an active area of research in recent years. A critical issue with most existing generative models is that the generated responses lack informativeness and diversity. A few researchers attempt to leverage the results of retrieval models to strengthen the generative models, but these models are limited by the quality of the retrieval results. In this work, we propose a memory-augmented generative model, which learns to abstract from the training corpus and saves the useful information to the memory to assist the response generation. Our model clusters query-response samples, extracts characteristics of each cluster, and learns to utilize these characteristics for response generation. Experimental results show that our model outperforms other competitive baselines.
Search
Co-authors
- Zhiliang Tian 2
- Wei Bi 2
- Dongkyu Lee 1
- Lanqing Xue 1
- Yiping Song 1
- show all...
Venues
- ACL2