Evaluating Compositionality of Sentence Representation Models

Hanoz Bhathena, Angelica Willis, Nathan Dass


Abstract
We evaluate the compositionality of general-purpose sentence encoders by proposing two different metrics to quantify compositional understanding capability of sentence encoders. We introduce a novel metric, Polarity Sensitivity Scoring (PSS), which utilizes sentiment perturbations as a proxy for measuring compositionality. We then compare results from PSS with those obtained via our proposed extension of a metric called Tree Reconstruction Error (TRE) (CITATION) where compositionality is evaluated by measuring how well a true representation producing model can be approximated by a model that explicitly combines representations of its primitives.
Anthology ID:
2020.repl4nlp-1.22
Volume:
Proceedings of the 5th Workshop on Representation Learning for NLP
Month:
July
Year:
2020
Address:
Online
Venues:
ACL | RepL4NLP | WS
SIG:
SIGREP
Publisher:
Association for Computational Linguistics
Note:
Pages:
185–193
URL:
https://www.aclweb.org/anthology/2020.repl4nlp-1.22
DOI:
Bib Export formats:
BibTeX MODS XML EndNote
PDF:
https://www.aclweb.org/anthology/2020.repl4nlp-1.22.pdf
Software:
 2020.repl4nlp-1.22.Software.zip

You can write comments here (and agree to place them under CC-by). They are not guaranteed to stay and there is no e-mail functionality.