Hui Liu
2020
Does Multi-Encoder Help? A Case Study on Context-Aware Neural Machine Translation
Bei Li
|
Hui Liu
|
Ziyang Wang
|
Yufan Jiang
|
Tong Xiao
|
Jingbo Zhu
|
Tongran Liu
|
changliang li
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
In encoder-decoder neural models, multiple encoders are in general used to represent the contextual information in addition to the individual sentence. In this paper, we investigate multi-encoder approaches in document-level neural machine translation (NMT). Surprisingly, we find that the context encoder does not only encode the surrounding sentences but also behaves as a noise generator. This makes us rethink the real benefits of multi-encoder in context-aware translation - some of the improvements come from robust training. We compare several methods that introduce noise and/or well-tuned dropout setup into the training of these encoders. Experimental results show that noisy training plays an important role in multi-encoder-based NMT, especially when the training data is small. Also, we establish a new state-of-the-art on IWSLT Fr-En task by careful use of noise generation and dropout methods.
Jointly Learning to Align and Summarize for Neural Cross-Lingual Summarization
Yue Cao
|
Hui Liu
|
Xiaojun Wan
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Cross-lingual summarization is the task of generating a summary in one language given a text in a different language. Previous works on cross-lingual summarization mainly focus on using pipeline methods or training an end-to-end model using the translated parallel data. However, it is a big challenge for the model to directly learn cross-lingual summarization as it requires learning to understand different languages and learning how to summarize at the same time. In this paper, we propose to ease the cross-lingual summarization training by jointly learning to align and summarize. We design relevant loss functions to train this framework and propose several methods to enhance the isomorphism and cross-lingual transfer between languages. Experimental results show that our model can outperform competitive models in most cases. In addition, we show that our model even has the ability to generate cross-lingual summaries without access to any cross-lingual corpus.
Search
Co-authors
- Bei Li 1
- Ziyang Wang 1
- Yufan Jiang 1
- Tong Xiao 1
- Jingbo Zhu 1
- show all...
Venues
- ACL2