5,340
Views
0
CrossRef citations to date
0
Altmetric
Introduction

Evaluation measures and the study of language acquisition

&

Language acquisition has traditionally provided one of the primary explanatory goals of theory construction in linguistics (Chomsky Citation1965). A theory of grammar aims not merely to provide analyses for particular phenomena in particular languages, but rather to characterize the nature of possible human languages. In turn, constraints on the range of possible languages should make it possible for learners to generalize from their finite experience to a grammatical system capable of generating an infinite number of sentences, and to do so in a roughly uniform way that is robust to small variations in experience.

In light of this explanatory agenda, research in developmental linguistics has two goals. First, it aims to test a theory of linguistic universals and cross-linguistic variation by showing how that space constrains the hypotheses that learners consider during development (Pinker 1984, Gleitman Citation1990, deVillers, Roeper & Vainikka Citation1990, Thornton Citation1990, Crain Citation1991, Snyder Citation2001, Lidz, Waxman & Freedman Citation2003, Syrett & Lidz Citation2009, Bergelson & Idsardi Citation2009, Berent Citation2013). Do learners only consider hypotheses within the space of possible grammars defined by a linguistic theory? To the extent that they consider hypotheses outside this space, this provides an argument against some piece of the linguistic theory. Second, developmental linguistics aims to understand how learners use the input to traverse the space of possible languages (Wexler & Culicover Citation1980, Dresher & Kaye Citation1990, Lightfoot Citation1991, Gibson & Wexler Citation1994, Tesar & Smolensky Citation1998, Pater Citation2008, Yang 2004, Viau & Lidz Citation2011, Sakas & Fodor Citation2012, Pearl & Sprouse Citation2013). Do learners use all of the data that they are exposed to in determining the grammar? How are grammars updated throughout development? Is grammar acquisition deterministic?

While generative research in first and second language acquisition uses the descriptive vocabulary of theoretical linguistics to understand language development, fewer and fewer papers address these fundamental questions and link a theory of possible grammars to observations about linguistic development. In large measure, this is for good reason. Often, psycholinguistic factors conspire to give the illusion that learners have acquired a grammar unlike that of their environment or outside the space of possible grammars. Thus, we find a focus on diagnosing learners’ patterns of behavior as being due to grammatical knowledge or to factors of language use (Hamburger & Crain Citation1984, Crain & Thornton Citation1998, Musolino & Lidz Citation2006, Conroy et al Citation2009, Clahsen & Felser Citation2006). Indeed, the study of the extralinguistic contributions to both first and second language acquisition has become a growing subfield of its own (Omaki & Lidz, 2015. Phillips & Ehrenhofer 2015). Relatedly, identifying the grammatical knowledge of children and ruling out alternative grammatical analyses is, in itself, a daunting task for any given piece of knowledge (Valian Citation1991, Hoekstra & Hyams Citation1998, Thornton & Wexler Citation1999, Becker Citation2005, Valian, Solt & Stewart Citation2009, Orfitelli & Hyams Citation2012 inter alia). As a consequence, the field is largely dominated by investigations of children’s knowledge at specific ages, the trajectory of linguistic development and the interaction between knowledge and performance factors in shaping children’s linguistic behavior and development. However, the subsequent step of using those results to inform questions of linguistic theory or to inform the theory of how learners use their experience to move through the space of possible grammars is rarely undertaken (but see Yang 2004, Becker & Mitchener Citation2011, Pearl Citation2011 as prominent counterexamples).

Given that 2015 was the 50th anniversary of Aspects of the Theory of Syntax (Chomsky Citation1965), a book that articulated the explanatory goals of generative linguistics, a specially themed GALANA was held in 2015 at the University of Maryland in order to highlight research that foregrounds these traditional questions. The conference theme addressed the question of how linguistic theory informs our understanding of both what children know at any given point in development, and also how the input can allow for the construction of a particular grammar.

Chomsky (Citation1965) proposed that a language acquisition mechanism must minimally consist of three components: a way of representing the input signal, a space of possible grammars and an evaluation measure—a method for selecting a particular grammar on the basis of the first two components. However, even as grammatical theories have diversified from sets of rules (Chomsky Citation1965) to principles and parameters (Chomsky Citation1981, Dresher & Kaye Citation1990, Baker Citation1995) or constraint rankings (Prince & Smolensky Citation1993), the problem of grammatical opacity has remained. The utterances that a learner is exposed to bear an indirect relation to the grammar that generated them, even with highly constrained grammatical theories. The vast majority of surface evidence is compatible with multiple grammatical descriptions (Berwick 1985, Clark Citation1989, Citation1992, Gibson & Wexler Citation1994, Dresher & Kaye Citation1990, Tesar & Smolensky Citation1998, Yang 2004, Pearl Citation2009, Fodor & Sakas 2013). Consequently, even defining the evidence that unambiguously allows for the selection of one particular grammar remains a challenge, let alone identifying the psychological mechanisms through which learners recognize that evidence and put it to use in grammar selection (Lidz & Gagliardi, 2015).

The current special issue presents four articles by leading researchers in the field, all of whom use computational tools to address the opacity of the relation between a grammar and the set of utterances it produces. In this light, the evaluation measure is the mechanism through which this opacity can be overcome.

Janet Dean Fodor’s article approaches the problem of parametric ambiguity. In a theory of grammatical variation defined by syntactic parameters, the evidence a learner is exposed to may be compatible with multiple grammars. Fodor proposes an evaluation measure that inherits its properties from sentence processing mechanisms. She argues that an incremental parser with limited capacity for parallelism circumvents certain problems of parametric ambiguity. In this sense, the evaluation measure is a byproduct of the parser, rather than a component of the learning theory per se. She argues that parameters should be thought of as tree fragments that can be activated with a strength proportional to their frequency. With this conception of the parameters, she reports data that support a mechanism of learning by parsing over one that learns through domain search.

Charles Yang addresses the puzzle of generalization and whether indirect negative evidence is a viable information source for learning morphosyntactic generalizations. He focuses in particular on subset-superset problems (Gold 1967, Berwick 1985) in which a given set of data is compatible with two grammars whose extensions stand in a subset-superset relation. In examining the subset of adjectives in English that can be used predicatively (I am asleep) but not attributively (*the asleep cat), Yang shows that learning by indirect negative evidence fails to identify this subclass accurately and that instead, a grammatical analysis that treats them as belonging to the same syntactic class as prepositional phrases can lead to successful learning on the basis of positive evidence alone. Yang also identifies a common threshold against which any set of forms can be compared in order to determine whether there is a productive generalization that explains those forms. In essence, Yang argues that a learner armed with the right tools of grammatical analysis and a general theory of productivity allows for richer explanations of how learners identify grammatical generalizations than one which simply calculates probabilities. More generally, Yang’s article emphasizes that learners require a rich characterization of possible grammatical representations in order for simple evaluation measures to be effective.

Lisa Pearl raises the question of whether one can use evaluation measures as a probe of grammatical representations. She compares several models of stress placement, which differ in their representational vocabulary, but which all provide accurate descriptions of English stress placement. She asks within each model whether a domain-general bayesian learner converges on the grammar of English stress given child directed speech. Then, she compares differences in convergence in order to see whether one class of models gets closer to the correct grammar. If such a way of evaluating grammars works, it opens up new ways of teasing apart alternative grammatical analyses. That is, to the extent that two grammars with equal descriptive coverage differ in how well they can be learned, we will have used the evaluation measure as a way of separating descriptive from explanatory adequacy.

Finally, Bruce Tesar examines learnability in the context of Optimality Theory (Prince & Smolensky Citation1993). The puzzle in this context comes from the vast space of possible grammars and lexicons defined by an OT framework. Searching this vast space is computationally infeasible in both humans and machines. Tesar’s approach to delimiting this search space takes advantage of output driven maps. Such maps define candidates in terms of the similarities between the output form and the potential input forms. Because entailment relations can be defined between large swaths of candidate input-output maps, the search space for a learner can be defined in highly structured ways that allow for more efficient search over possible lexicons. Tesar illustrates how such an approach can be taken to learning phonotactics, morphophonological generalizations, underlying features, constraint rankings, and contrastiveness. These output driven maps define structure over possible grammars beyond what comes from the OT framework itself, making more efficient learning possible.

While the four articles all tackle different problems in syntax, morphology and phonology, they share an aim of exploring how learners search through the space of possible grammars to identify a particular grammar. In each case we see how attention to the fine structure of grammatical theory along with computational and psycholinguistic considerations contributes to new insights in the study of language acquisition. These articles thus provide a model for how future research in developmental linguistics can reconnect with the fundamental explanatory goals of linguistic theory.

Acknowledgements

The conference that gave rise to the papers in this special issue was funded in part by a grant from the National Science Foundation: BCS-1451584.

References

  • Baker, M. 1995. The Polysynthesis Parameter. Oxford University Press.
  • Becker, M. 2005. There Began to be a Learnability Puzzle, Linguistic Inquiry, 37, 441–456
  • Becker, M. & G. Mitchener 2011. A Computational Model of Learning the Raising-Control Distinction, Research on Language and Computation, 8, 169–207.
  • Berent, I. 2013. The Phonological Mind. Cambridge University Press.
  • Bergelson, E. & W. Idsardi 2009. Structural Biases in Phonology: Infant and Adult Evidence from Artificial Language Learning. BUCLD 33: Proceedings of the 33rd Annual Boston University Conference on Language Development, 85–96.
  • Chomsky, N. 1965. Aspects of the Theory of Syntax. MIT Press.
  • Chomsky, N. 1981. Lectures on Government and Binding. Foris.
  • Clahsen, H. & C. Felser 2006. Grammatical processing in language learners. Applied Psycholinguistics, 27, 3–42.
  • Clark, R. 1989. On the relationship between the input data and parameter setting. In J. Carter & R. M. Dechaine (Eds.), Proceedings of The 19th Annual Meeting of the North East Linguistic Society. Amherst, MA.
  • Clark, R. 1992. The selection of syntactic knowledge. Language Acquisition, 2, 83‐149.
  • Conroy, A., Takahashi, E., Lidz, J., & Phillips, C. 2009. Equal treatment for all antecedents: How children succeed with principle B. Linguistic Inquiry, 40, 446–486.
  • Crain, S. 1991. Language acquisition in the absence of experience. Behavioral and Brain Sciences, 14, 597–650.
  • Crain, S. & R. Thornton. 1998. Investigations in universal grammar: A guide to experiments on the acquisition of syntax and semantics. MIT Press.
  • Dresher, E. & J. Kaye. 1990. A Computational Learning Model for Metrical Phonology. Cognition 34, 137–195.
  • Gibson, E., & Wexler, K. 1994. Triggers. Linguistic Inquiry, 25, 407–454.
  • Gleitman, LR. 1990. The structural sources of verb meanings. Language Acquisition, 1, 3–55.
  • Grodzinsky, Y. & T. Reinhart. 1993. The innateness of binding and of coreference. Linguistic Inquiry, 24, 69–101.
  • Hamburger, H., & Crain, S. 1984. Acquisition of cognitive compiling. Cognition, 17, 85–136.
  • Hoekstra, T. & Hyams, N. 1998. Aspects of Root Infinitives. Lingua 106, 81–112.
  • Lidz, J. & A. Gagliardi. In Press. How nature meets nurture: Universal Grammar and Statistical Learning. Annual Reviews of Linguistics, 1.
  • Lidz, J., S. Waxman & J. Freedman. 2003. What Infants Know about Syntax but Couldn’t have Learned: Experimental Evidence for Syntactic Structure at 18-months. Cognition 89:B65–B73.
  • Lightfoot, D. 1991. How to Set Parameters. MIT Press.
  • Musolino, J., & Lidz, J. 2006. Why children aren’t universally successful with quantification. Linguistics, 44, 817–852.
  • Omaki, A. & J. Lidz. In Press. Linking parser development to the acquisition of syntax. Language Acquisition.
  • Orfitelli, R. & N. Hyams. 2012. Children’s Grammar of Null Subjects: Evidence from comprehension. Linguistic Inquiry 43,4; 563–590.
  • Pater, J. 2008. Gradual learning and convergence. Linguistic Inquiry, 39, 334–345.
  • Pearl, L. 2009. Necessary Bias in Natural Language Learning. PhD Dissertation, University of Maryland.
  • Pearl, L. 2011. When unbiased probabilistic learning is not enough: Acquiring a parametric system of metrical phonology. Language Acquisition, 18, 87–120.
  • Pearl, L. & J. Sprouse 2013. Syntactic islands and learning biases: Combining experimental syntax and computational modeling to investigate the language acquisition problem. Language Acquisition, 20, 23–68.
  • Prince, A. & P. Smolensky. 1993. Optimality Theory. MIT Press.
  • Sakas, W.G. & J.D. Fodor. 2012. Disambiguating Syntactic Triggers. Language Acquisition, 19, 83–143.
  • Snyder, W. 2001. On the nature of syntactic variation: evidence from complex predicates and complex word formation. Language, 77, 324–342.
  • Syrett, K., & Lidz, J. 2011. Competence, performance, and the locality of quantifier raising: Evidence from 4-year-old children. Linguistic Inquiry, 42, 305–337.
  • Tesar. B., & P. Smolensky, 1998. Learnability in optimality theory. Linguistic Inquiry, 29, 229–268.
  • Thornton, R. 1990. Adventures in Long-distance Moving. PhD dissertation. University of Connecticut.
  • Thornton, R., & Wexler, K. 1999. Principle B, VP ellipsis, and interpretation in child grammar. Cambridge, MA: MIT Press.
  • Valian, V. 1991. Syntactic subjects in the early speech of American and Italian children. Cognition, 40, 21–81.
  • Valian, V., S. Solt & J. Stewart 2009, Abstract categories or limited-scope formulae: The case of children’s determiners. Journal of Child Language.
  • Viau, J. & J. Lidz. 2011. Selective learning in the acquisition of Kannada ditransitives. Language, 87, 679–714.
  • De Villiers, J., Roeper, T., & Vainikka, A. 1990. The acquisition of long-distance rules. In L. Frazier & J. De Villiers (Eds.), Language processing and language acquisition (pp. 257–297). Dordrecht: Kluwer.
  • Wexler K, & P. Culicover. 1980. Formal Principles of Language Acquisition. MIT Press.
  • Yang, C. 2002. Knowledge and learning in natural language. Oxford University Press.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.