Dan Jurafsky
2020
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Dan Jurafsky
|
Joyce Chai
|
Natalie Schluter
|
Joel Tetreault
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Pretraining with Contrastive Sentence Objectives Improves Discourse Performance of Language Models
Dan Iter
|
Kelvin Guu
|
Larry Lansing
|
Dan Jurafsky
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Recent models for unsupervised representation learning of text have employed a number of techniques to improve contextual word representations but have put little focus on discourse-level representations. We propose Conpono, an inter-sentence objective for pretraining language models that models discourse coherence and the distance between sentences. Given an anchor sentence, our model is trained to predict the text k sentences away using a sampled-softmax objective where the candidates consist of neighboring sentences and sentences randomly sampled from the corpus. On the discourse representation benchmark DiscoEval, our model improves over the previous state-of-the-art by up to 13% and on average 4% absolute across 7 tasks. Our model is the same size as BERT-Base, but outperforms the much larger BERT-Large model and other more recent approaches that incorporate discourse. We also show that Conpono yields gains of 2%-6% absolute even for tasks that do not explicitly evaluate discourse: textual entailment (RTE), common sense reasoning (COPA) and reading comprehension (ReCoRD).
Social Bias Frames: Reasoning about Social and Power Implications of Language
Maarten Sap
|
Saadia Gabriel
|
Lianhui Qin
|
Dan Jurafsky
|
Noah A. Smith
|
Yejin Choi
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Warning: this paper contains content that may be offensive or upsetting. Language has the power to reinforce stereotypes and project social biases onto others. At the core of the challenge is that it is rarely what is stated explicitly, but rather the implied meanings, that frame people’s judgments about others. For example, given a statement that “we shouldn’t lower our standards to hire more women,” most listeners will infer the implicature intended by the speaker - that “women (candidates) are less qualified.” Most semantic formalisms, to date, do not capture such pragmatic implications in which people express social biases and power differentials in language. We introduce Social Bias Frames, a new conceptual formalism that aims to model the pragmatic frames in which people project social biases and stereotypes onto others. In addition, we introduce the Social Bias Inference Corpus to support large-scale modelling and evaluation with 150k structured annotations of social media posts, covering over 34k implications about a thousand demographic groups. We then establish baseline approaches that learn to recover Social Bias Frames from unstructured text. We find that while state-of-the-art neural models are effective at high-level categorization of whether a given statement projects unwanted social bias (80% F1), they are not effective at spelling out more detailed explanations in terms of Social Bias Frames. Our study motivates future work that combines structured pragmatic inference with commonsense reasoning on social implications.
Search
Co-authors
- Joyce Chai 1
- Natalie Schluter 1
- Joel Tetreault 1
- Dan Iter 1
- Kelvin Guu 1
- show all...
Venues
- ACL3