A note on the representation and learning of quantificational determiners
DOI:
https://doi.org/10.18148/sub/2020.v24i1.874Abstract
There is a tight, bidirectional connection between the formalism that defines how linguistic knowledge is stored and how this knowledge can be learned. In one direction, the formalism can be mapped onto an evaluation metric that allows the child to compare competing hypotheses given the input data. In the other direction, an evaluation metric can help the linguist to compare competing hypotheses about the formalism in which linguistic knowledge is written. In this preliminary note we explore this bidirectional connection in the domain of quantificational determiners (e.g., ‘every’ and ‘some’). We show how fixing an explicit format for representing the semantics of such elements – specifically, a variant of semantic automata – yields an evaluation metric, based on the principle of Minimum Description Length (MDL), that can serve as the basis for an unsupervised learner for such denotations. We then show how the MDL metric may provide a handle on the comparison of semantic automata with a competing representational format.Downloads
Published
2020-09-17
How to Cite
Katzir, R., Lan, N., & Peled, N. (2020). A note on the representation and learning of quantificational determiners. Proceedings of Sinn Und Bedeutung, 24(1), 392–410. https://doi.org/10.18148/sub/2020.v24i1.874
Issue
Section
Articles