We propose a computational modeling framework for inducing combinatory categorial grammars from arbitrary behavioral data. This framework provides the analyst fine-grained control over the assumptions that the induced grammar should conform to: (i) what the primitive types are; (ii) how complex types are constructed; (iii) what set of combinators can be used to combine types; and (iv) whether (and to what) the types of some lexical items should be fixed. In a proof-of-concept experiment, we deploy our framework for use in distributional analysis. We focus on the relationship between s(emantic)-selection and c(ategory)-selection, using as input a lexicon-scale acceptability judgment dataset focused on English verbs’ syntactic distribution (the MegaAcceptability dataset) and enforcing standard assumptions from the semantics literature on the induced grammar. more »« less
Kim, Gene Louis; White, Aaron Steven
(, Semantics and Linguistic Theory)
null
(Ed.)
We propose a computational model for inducing full-fledged combinatory categorial grammars from behavioral data. This model contrasts with prior computational models of selection in representing syntactic and semantic types as structured (rather than atomic) objects, enabling direct interpretation of the modeling results relative to standard formal frameworks. We investigate the grammar our model induces when fit to a lexicon-scale acceptability judgment dataset – Mega Acceptability – focusing in particular on the types our model assigns to clausal complements and the predicates that select them.
Masten, Matthew A; Poirier, Alexandre
(, The Econometrics Journal)
There are many kinds of exogeneity assumptions. How should researchers choose among them? When exogeneity is imposed on an unobservable like a potential outcome, we argue that the form of exogeneity should be chosen based on the kind of selection on unobservables it allows. Consequently, researchers can assess the plausibility of any exogeneity assumption by studying the distributions of treatment given the unobservables that are consistent with that assumption. We use this approach to study two common exogeneity assumptions: quantile and mean independence. We show that both assumptions require a kind of nonmonotonic relationship between treatment and the potential outcomes. We discuss how to assess the plausibility of this kind of treatment selection. We also show how to define a new and weaker version of quantile independence that allows for monotonic selection on unobservables. We then show the implications of the choice of exogeneity assumption for identification. We apply these results in an empirical illustration of the effect of child soldiering on wages.
Jin, Lifeng; Schuler, William
(, Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics)
In unsupervised grammar induction, data likelihood is known to be only weakly correlated with parsing accuracy, especially at convergence after multiple runs. In order to find a better indicator for quality of induced grammars, this paper correlates several linguistically- and psycholinguistically-motivated predictors to parsing accuracy on a large multilingual grammar induction evaluation data set. Results show that variance of average surprisal (VAS) better correlates with parsing accuracy than data likelihood and that using VAS instead of data likelihood for model selection provides a significant accuracy boost. Further evidence shows VAS to be a better candidate than data likelihood for predicting word order typology classification. Analyses show that VAS seems to separate content words from function words in natural language grammars, and to better arrange words with different frequencies into separate classes that are more consistent with linguistic theory.
Sexual selection is a powerful force shaping not only the details but also the breadth of what we see in nature. Yet so much unexplained variation remains. Organisms often solve the “problem” of how to pass on their genes in ways that do not fit our current expectations. I argue here that integrating empirical surprises will push our understanding of sexual selection forward. Such “nonmodel” organisms (i.e., species that do not do what we think they should do) challenge us to think deeply, integrate puzzling results, question our assumptions, and consider the new (and arguably better) questions these unexpected patterns pose. In this article, I share how puzzling observations from my long-term research on the ocellated wrasse (Symphodus ocellatus) have shaped my understanding of sexual selection and suggested new questions about the interplay among sexual selection, plasticity, and social interactions. My general premise, however, is not that others should study these questions. Instead, I argue for a change in the culture of our field—to consider unexpected results a welcome opportunity to generate new questions and learn new things about sexual selection. Those of us in positions of power (e.g., as editors, reviewers, and authors) need to lead the way.
The ability to provide comprehensive explanations of chosen actions is a hallmark of intelligence. Lack of this ability impedes the general acceptance of AI and robot systems in critical tasks. This paper examines what forms of explanations best foster human trust in machines and proposes a framework in which explanations are generated from both functional and mechanistic perspectives. The robot system learns from human demonstrations to open medicine bottles using (i) an embodied haptic prediction model to extract knowledge from sensory feedback, (ii) a stochastic grammar model induced to capture the compositional structure of a multistep task, and (iii) an improved Earley parsing algorithm to jointly leverage both the haptic and grammar models. The robot system not only shows the ability to learn from human demonstrators but also succeeds in opening new, unseen bottles. Using different forms of explanations generated by the robot system, we conducted a psychological experiment to examine what forms of explanations best foster human trust in the robot. We found that comprehensive and real-time visualizations of the robot’s internal decisions were more effective in promoting human trust than explanations based on summary text descriptions. In addition, forms of explanation that are best suited to foster trust do not necessarily correspond to the model components contributing to the best task performance. This divergence shows a need for the robotics community to integrate model components to enhance both task execution and human trust in machines.
Kim, Gene Louis, and White, Aaron Steven. Montague Grammar Induction. Retrieved from https://par.nsf.gov/biblio/10299985. Proceedings from Semantics and Linguistic Theory 30. Web. doi:DOI: 10.3765/salt.v30i0.4816.
Kim, Gene Louis, & White, Aaron Steven. Montague Grammar Induction. Proceedings from Semantics and Linguistic Theory, 30 (). Retrieved from https://par.nsf.gov/biblio/10299985. https://doi.org/DOI: 10.3765/salt.v30i0.4816
@article{osti_10299985,
place = {Country unknown/Code not available},
title = {Montague Grammar Induction},
url = {https://par.nsf.gov/biblio/10299985},
DOI = {DOI: 10.3765/salt.v30i0.4816},
abstractNote = {We propose a computational modeling framework for inducing combinatory categorial grammars from arbitrary behavioral data. This framework provides the analyst fine-grained control over the assumptions that the induced grammar should conform to: (i) what the primitive types are; (ii) how complex types are constructed; (iii) what set of combinators can be used to combine types; and (iv) whether (and to what) the types of some lexical items should be fixed. In a proof-of-concept experiment, we deploy our framework for use in distributional analysis. We focus on the relationship between s(emantic)-selection and c(ategory)-selection, using as input a lexicon-scale acceptability judgment dataset focused on English verbs’ syntactic distribution (the MegaAcceptability dataset) and enforcing standard assumptions from the semantics literature on the induced grammar.},
journal = {Proceedings from Semantics and Linguistic Theory},
volume = {30},
author = {Kim, Gene Louis and White, Aaron Steven},
editor = {null}
}
Warning: Leaving National Science Foundation Website
You are now leaving the National Science Foundation website to go to a non-government website.
Website:
NSF takes no responsibility for and exercises no control over the views expressed or the accuracy of
the information contained on this site. Also be aware that NSF's privacy policy does not apply to this site.