Wei Bi
2020
Response-Anticipated Memory for On-Demand Knowledge Integration in Response Generation
Zhiliang Tian
|
Wei Bi
|
Dongkyu Lee
|
Lanqing Xue
|
YIPING SONG
|
Xiaojiang Liu
|
Nevin L. Zhang
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Neural conversation models are known to generate appropriate but non-informative responses in general. A scenario where informativeness can be significantly enhanced is Conversing by Reading (CbR), where conversations take place with respect to a given external document. In previous work, the external document is utilized by (1) creating a context-aware document memory that integrates information from the document and the conversational context, and then (2) generating responses referring to the memory. In this paper, we propose to create the document memory with some anticipated responses in mind. This is achieved using a teacher-student framework. The teacher is given the external document, the context, and the ground-truth response, and learns how to build a response-aware document memory from three sources of information. The student learns to construct a response-anticipated document memory from the first two sources, and teacher’s insight on memory creation. Empirical results show that our model outperforms the previous state-of-the-art for the CbR task.
A Batch Normalized Inference Network Keeps the KL Vanishing Away
Qile Zhu
|
Wei Bi
|
Xiaojiang Liu
|
Xiyao Ma
|
Xiaolin Li
|
Dapeng Wu
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Variational Autoencoder (VAE) is widely used as a generative model to approximate a model’s posterior on latent variables by combining the amortized variational inference and deep neural networks. However, when paired with strong autoregressive decoders, VAE often converges to a degenerated local optimum known as “posterior collapse”. Previous approaches consider the Kullback–Leibler divergence (KL) individual for each datapoint. We propose to let the KL follow a distribution across the whole dataset, and analyze that it is sufficient to prevent posterior collapse by keeping the expectation of the KL’s distribution positive. Then we propose Batch Normalized-VAE (BN-VAE), a simple but effective approach to set a lower bound of the expectation by regularizing the distribution of the approximate posterior’s parameters. Without introducing any new model component or modifying the objective, our approach can avoid the posterior collapse effectively and efficiently. We further show that the proposed BN-VAE can be extended to conditional VAE (CVAE). Empirically, our approach surpasses strong autoregressive baselines on language modeling, text classification and dialogue generation, and rivals more complex approaches while keeping almost the same training time as VAE.
Learning to Customize Model Structures for Few-shot Dialogue Generation Tasks
YIPING SONG
|
Zequn Liu
|
Wei Bi
|
Rui Yan
|
Ming Zhang
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Training the generative models with minimal corpus is one of the critical challenges for building open-domain dialogue systems. Existing methods tend to use the meta-learning framework which pre-trains the parameters on all non-target tasks then fine-tunes on the target task. However, fine-tuning distinguishes tasks from the parameter perspective but ignores the model-structure perspective, resulting in similar dialogue models for different tasks. In this paper, we propose an algorithm that can customize a unique dialogue model for each task in the few-shot setting. In our approach, each dialogue model consists of a shared module, a gating module, and a private module. The first two modules are shared among all the tasks, while the third one will differentiate into different network structures to better capture the characteristics of the corresponding task. The extensive experiments on two datasets show that our method outperforms all the baselines in terms of task consistency, response quality, and diversity.
Search
Co-authors
- YIPING SONG 2
- Xiaojiang Liu 2
- Zhiliang Tian 1
- Dongkyu Lee 1
- Lanqing Xue 1
- show all...
Venues
- ACL3