Guoping Huang
2020
Evaluating Explanation Methods for Neural Machine Translation
Jierui Li
|
Lemao Liu
|
Huayang Li
|
Guanlin Li
|
Guoping Huang
|
Shuming Shi
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Recently many efforts have been devoted to interpreting the black-box NMT models, but little progress has been made on metrics to evaluate explanation methods. Word Alignment Error Rate can be used as such a metric that matches human understanding, however, it can not measure explanation methods on those target words that are not aligned to any source word. This paper thereby makes an initial attempt to evaluate explanation methods from an alternative viewpoint. To this end, it proposes a principled metric based on fidelity in regard to the predictive behavior of the NMT model. As the exact computation for this metric is intractable, we employ an efficient approach as its approximation. On six standard translation tasks, we quantitatively evaluate several explanation methods in terms of the proposed metric and we reveal some valuable findings for these explanation methods in our experiments.
Regularized Context Gates on Transformer for Machine Translation
Xintong Li
|
Lemao Liu
|
Rui Wang
|
Guoping Huang
|
Max Meng
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Context gates are effective to control the contributions from the source and target contexts in the recurrent neural network (RNN) based neural machine translation (NMT). However, it is challenging to extend them into the advanced Transformer architecture, which is more complicated than RNN. This paper first provides a method to identify source and target contexts and then introduce a gate mechanism to control the source and target contributions in Transformer. In addition, to further reduce the bias problem in the gate mechanism, this paper proposes a regularization method to guide the learning of the gates with supervision automatically generated using pointwise mutual information. Extensive experiments on 4 translation datasets demonstrate that the proposed model obtains an averaged gain of 1.0 BLEU score over a strong Transformer baseline.
Search
Co-authors
- Lemao Liu 2
- Jierui Li 1
- Huayang Li 1
- Guanlin Li 1
- Shuming Shi 1
- show all...
Venues
- ACL2