How syntax gets you to "think"
A perennial problem in literature on word-learning addresses how children learn the meanings of words like “think” and “want”. Since the mental states they refer to are unobservable—-one can’t point to thoughts or desires even if they can see their effects—-it is unclear how a learner could know to link “think” to the concept thinking and “want” to the concept wanting. One proposal for how children learn these meanings is that they can categorize words by the sorts of sentences they show up in (Gleitman, 1990). For instance, “think” can show up in sentences like (1a) but is odd in sentences like (1b) and (1c) (denoted by the “*”) while “want” can show up in sentences like (2b) and (2c) but is odd in sentences like (2a). 1a. Carla thinks that Janet went to the store. 1b. *Carla thinks Janet to go to the store. 1c. *Carla thinks a piece of cake. 2a. *Carla wants that Janet went to the store. 2b. Carla wants Janet to go to the store. 2c. Carla wants a piece of cake. Here, I present evidence for this proposal—-that features of a sentence carry information about the meaning of words in that sentence—-using both experimental methodologies and computational modeling. I show that there exists a nontrivial correlation between the sentences a word can show up in and that word’s meaning.
Ayelet Gneezy
Faculty
fun and interesting!
first a clarification: what do you mean when you say you looked at “how good different words sound with different sentence types”. could you please explain what yo mean by “good”?
now fr the question: do you think that MTurk participants are a good proxy for children (in the context of learning word meaning)? I see the value in exposing them to new, unfamiliar, words, but I am not sure that is sufficient, primarily because adults’ world, associations, context and learning experiences are significantly more loaded than those of children.
Aaron White
To answer the clarification: by good, I mean judged acceptable by participants. An “acceptable” sentence is something a native speaker of English would say. For example, “Frank wondered what he should eat with” is 100% acceptable even if a prescriptive text would warrant against the ‘dangling preposition’. In contrast, “Frank wondered what he should eat bananas and” is clearly not acceptable. Acceptability judgments are a common form of data used in theoretical linguistics. These judgments can be collected in various ways. The way we used was a Likert scale task, but thermometer and magnitude estimation tasks are also used.
To answer your question (Dr. Drayton had a similar question in the Discussion section): first, a clarification. Adults were seeing words they knew and were either making judgments about their similarity with other words they knew or judgments about their compatibility (acceptability) with a syntactic construction. Experiments (in the human simulation paradigm; Gillette et al. 1999) are conducted that ask adults to make inferences about the meaning of a new word given constrained information—e.g. linguistic context and/or visual context—and we plan to carry some out using data from these experiments (see my answer to Dr. Drayton’s question for some of the reasoning behind this). But these experiments were with known words.
With regard to adults being quite different from children in terms of their biases, this is quite likely. The idea we’re pursuing here is to establish what the end state—the adult’s knowledge of their language—-in fact looks like. That is, what do the connections between different words’ meanings and their syntactic distributions look like once you’ve learned these words’ meanings? The idea is that this is as noiseless a representation as we could hope to get of the end state. The next step is to use this knowledge about the informativity of different syntactic constructions to guide our research with children. What do children infer about a word’s meaning given the syntactic features it cooccurs with and is it what we’d expect given the adult data?
Adults are interesting in their own right also. Given that we now know something about the strength of connections between different syntactic cues and meanings, suppose we feed these distributions back into the adult by giving them instances of a novel word. Do they infer things about the novel word’s meaning that you would expect given our model, or do they differ in systematic ways. This is interesting from the modeler’s perspective because it can give insight into how knowledge about the connections between meaning and syntax is actually used in learning. And even if this just tells us about how adults learn, the hope is that it would expand our knowledge about human cognitive capacities.
Aaron White
The experiments can be viewed at the links below.
Similarity judgment experiment: http://spellout.net/ibexexps/aaronsteven/mortar...
Acceptability judgment experiment: http://spellout.net/ibexexps/aaronsteven/pestle...
Eileen Kowler
Faculty: Project PI
Given the unlimited number of possible utterances, can you say more about how you can develop, and test, a Bayesian model of language learning?
Aaron White
In the context of this research, we’re interested in addressing a possibly unbounded number of possible meaning categories. The way we address this is with nonparametric versions of Bayesian methods for text analysis. One broad class of methods we use are topic models—models that attempt to automatically learn topics from text. When we make these models nonparametric (by placing a prior like the Chinese Restaurant Process or the Indian Buffet Process on topics/categories/features), we allow the model to assume that the number of topics—or in our case, meaning categories—is infinite. What is nice about these models is that we can balance parsimony—less categories—and fit within the workings of the model. Another nice thing about them is that, since many of these models are built for analysis of text, the observed data are contingency tables—i.e. cooccurrence counts. This is exactly the sort of data we can glean from parsed corpora, allowing us to easily fit these models to cooccurrences of words with syntactic features.
Mary Gauvain
Faculty: Project Co-PI
What are the social experiences of language learning in childhood, especially social support, that are integral to this specific learning process and how do you plan to include these in your models?
Aaron White
Since the language children hear will be correlated with their social experiences, we can look at how our models perform (and what meaning categories they find) given transcripts of different social situations. For example, some of my recent work has focused on differences in syntactic complexity between play contexts (where a parent and child are playing together with toys) and meal contexts (where the parents and children are sitting at the dinner table). What I found was that there is higher syntactic complexity in the meal-time contexts than the play-time contexts (we already knew there was higher lexical complexity). The higher the syntactic complexity (and the more varied the syntax across situations), the better we’d expect our models to do. This suggests that different social experiences are an integral part of language learning even when we’re focused just on the linguistic input itself.
Our group is also collaborating with a group in Maryland’s Human Development and Quantitative Methods department to look at Socioeconomic Status (SES) differences both in the time course of word-learning and in the input. We’re interesting in looking at syntactic complexity differences—akin to the play-time v. meal-time contexts work I described—in SES. The idea is to see whether our models can predict differences in the word-learning time courses we may find between different SES groups.
Kristopher Irizarry
Faculty: Project Co-PI
A very cool project! Presumably your bayesian model of language learning might be useful as a diagnostic. Can you discuss the model and how it might be applied to identify / classify children?
Aaron White
This is a very interesting question. I suspect that topic models could be used to good effect here—where in this case a document corresponds to features associated with a single child. I think the bigger obstacle might be in the feature engineering. What features to extract would depend on the sort of classification you want to make. So for instance, if one were interested in syntactic differences among different sets of children, it would be interesting to fit a topic model to features extracted from parsed transcripts of children’s speech (maybe at the same age). (A slightly modified version of an adaptor grammar might also be interestingly applied for this sort of problem.) One might then look at the topics (classes of children) learned. An interesting application for this might be trying to discover common syntactic or lexical differences among disordered population that could aid in early discovery and treatment—something we’re currently looking into.
Timothy Waring
Faculty
Interesting project. You mention that you use MTurk as the most “noiseless” sample of the adult state, are you suggesting the MTurkers are representative English speaking adults? Did you collect demographic data to verify (or at least qualify) that ‘end state’ measurement?
Aaron White
We do collect demographic data to ensure the Turkers are native American English speakers. More importantly, though, we have designed and piloted a short screen to ensure that the participant is a native American English speaker (an MTurk qualification that is viewable here: http://goo.gl/BPJ3s). Having passed this screen is required before the Turker can participate.
Timothy Waring
Faculty
Thanks Aaron. That is great to know!