Joint learning of constraint weights and gradient inputs in Gradient Symbolic Computation with constrained optimization
Abstract
This paper proposes a method for the joint optimization of constraint weights and symbol activations within the Gradient Symbolic Computation (GSC) framework. The set of grammars representable in GSC is proven to be a subset of those representable with lexically-scaled faithfulness constraints. This fact is then used to recast the problem of learning constraint weights and symbol activations in GSC as a quadratically-constrained version of learning lexically-scaled faithfulness grammars. This results in an optimization problem that can be solved using Sequential Quadratic Programming.- Anthology ID:
- 2020.sigmorphon-1.27
- Volume:
- Proceedings of the 17th SIGMORPHON Workshop on Computational Research in Phonetics, Phonology, and Morphology
- Month:
- July
- Year:
- 2020
- Address:
- Online
- Venues:
- ACL | SIGMORPHON | WS
- SIG:
- SIGMORPHON
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 224–232
- URL:
- https://www.aclweb.org/anthology/2020.sigmorphon-1.27
- DOI:
- PDF:
- https://www.aclweb.org/anthology/2020.sigmorphon-1.27.pdf
You can write comments here (and agree to place them under CC-by). They are not guaranteed to stay and there is no e-mail functionality.