Nuo Xu
2020
Distinguish Confusing Law Articles for Legal Judgment Prediction
Nuo Xu
|
Pinghui Wang
|
Long Chen
|
Li Pan
|
Xiaoyan Wang
|
Junzhou Zhao
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Legal Judgement Prediction (LJP) is the task of automatically predicting a law case’s judgment results given a text describing the case’s facts, which has great prospects in judicial assistance systems and handy services for the public. In practice, confusing charges are often presented, because law cases applicable to similar law articles are easily misjudged. To address this issue, existing work relies heavily on domain experts, which hinders its application in different law systems. In this paper, we present an end-to-end model, LADAN, to solve the task of LJP. To distinguish confusing charges, we propose a novel graph neural network, GDL, to automatically learn subtle differences between confusing law articles, and also design a novel attention mechanism that fully exploits the learned differences to attentively extract effective discriminative features from fact descriptions. Experiments conducted on real-world datasets demonstrate the superiority of our LADAN.
Learning Architectures from an Extended Search Space for Language Modeling
Yinqiao Li
|
Chi Hu
|
Yuhao Zhang
|
Nuo Xu
|
Yufan Jiang
|
Tong Xiao
|
Jingbo Zhu
|
Tongran Liu
|
changliang li
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Neural architecture search (NAS) has advanced significantly in recent years but most NAS systems restrict search to learning architectures of a recurrent or convolutional cell. In this paper, we extend the search space of NAS. In particular, we present a general approach to learn both intra-cell and inter-cell architectures (call it ESS). For a better search result, we design a joint learning method to perform intra-cell and inter-cell NAS simultaneously. We implement our model in a differentiable architecture search system. For recurrent neural language modeling, it outperforms a strong baseline significantly on the PTB and WikiText data, with a new state-of-the-art on PTB. Moreover, the learned architectures show good transferability to other systems. E.g., they improve state-of-the-art systems on the CoNLL and WNUT named entity recognition (NER) tasks and CoNLL chunking task, indicating a promising line of research on large-scale pre-learned architectures.
Search
Co-authors
- Pinghui Wang 1
- Long Chen 1
- Li Pan 1
- Xiaoyan Wang 1
- Junzhou Zhao 1
- show all...
Venues
- ACL2