Grammaticality and Language Modelling

Total Page:16

File Type:pdf, Size:1020Kb

Grammaticality and Language Modelling Grammaticality and Language Modelling Jingcheng Niu and Gerald Penn Department of Computer Science University of Toronto Toronto, Canada fniu,[email protected] Abstract With the PBC in mind, we will first reappraise some recent work in syntactically targeted lin- Ever since Pereira(2000) provided evidence guistic evaluations (Hu et al., 2020), arguing against Chomsky’s (1957) conjecture that sta- that while their experimental design sets a new tistical language modelling is incommensu- high watermark for this topic, their results may rable with the aims of grammaticality predic- not prove what they have claimed. We then tion as a research enterprise, a new area of re- turn to the task-independent assessment of search has emerged that regards statistical lan- language models as grammaticality classifiers. guage models as “psycholinguistic subjects” Prior to the introduction of the GLUE leader- and probes their ability to acquire syntactic board, the vast majority of this assessment was knowledge. The advent of The Corpus of Lin- essentially anecdotal, and we find the use of guistic Acceptability (CoLA) (Warstadt et al., the MCC in this regard to be problematic. We 2019) has earned a spot on the leaderboard conduct several studies with PBCs to compare for acceptability judgements, and the polemic several popular language models. We also between Lau et al.(2017) and Sprouse et al. study the effects of several variables such as (2018) has raised fundamental questions about normalization and data homogeneity on PBC. the nature of grammaticality and how accept- ability judgements should be elicited. All the 1 Background while, we are told that neural language models continue to improve. The three currently most popular means of evalu- That is not an easy claim to test at present, ating a neural language model are: (1) perplexity, however, because there is almost no agreement an information-theoretic measure that was in use on how to measure their improvement when long before neural networks became the preferred it comes to grammaticality and acceptability means of implementing language models; (2) task judgements. The GLUE leaderboard bundles performance profiles, in which derivative aspects CoLA together with a Matthews correlation of a language model’s predictions are embedded coefficient (MCC), although probably because in a so-called “downstream” task, with all other CoLA’s seminal publication was using it to compute inter-rater reliabilities. Researchers aspects of the implementation held constant; and working in this area have used other accuracy (3) targeted linguistic evaluations, the purpose of and correlation scores, often driven by a need which is to demonstrate specific syntactic general- to reconcile and compare various discrete and izations that a candidate model implicitly captures continuous variables with each other. or does not capture. These targeted evaluations The score that we will advocate for in this pa- must take place on a large number of small data per, the point biserial correlation, in fact com- sets in order to control for the syntactic and lexical pares a discrete variable (for us, acceptabil- variations that we witness among sentences in a ity judgements) to a continuous variable (for realistic corpus. us, neural language model probabilities). The The purpose of this paper is ultimately to find only previous work in this area to choose the a task-independent means of testing how well lan- PBC that we are aware of is Sprouse et al. (2018), and that paper actually applied it back- guage model probabilities might serve as gram- wards (with some justification) so that the lan- maticality regression scores. Using evidence from guage model probability was treated as the dis- targeted linguistic evaluations, we argue for the crete binary variable by setting a threshold. point-biserial correlation as at least the basis of 110 Proceedings of the First Workshop on Evaluation and Comparison of NLP Systems (Eval4NLP), pages 110–119, November 20, 2020. c 2020 Association for Computational Linguistics such a task-independent measure, and then use the (1U) The girl said that the mothers saw herself. PBC to examine several neural models along with some important variables that affect both their eval- (2U) The mother said that the girls saw herself. uation and the data that we evaluate on. Borrowing a convention from linguistic theory, where (B) is a baseline grammatical string, (D) is a Marvin and Linzen(2018) coined the use of “mini- distractor, and (U) is an ungrammatical string. This mal pairs” as input to language models in order to set has six strings, but sets in their experiments can test these fine-grained variations. For example: have as many as 48 strings each, with as many as 75 sets in a single experiment, each one having (1) Reflexive pronoun in a sentential complement: a unique target pronoun in all of its strings. Be- a. The bankers thought the pilot embar- cause here it is the context that varies, rather than rassed himself. the pronoun, Hu et al.(2020) must rank the condi- b. *The bankers thought the pilot embar- tional probabilities of the pronoun in these various rassed themselves. contexts, rather than total sentence probabilities. Hu et al.(2020) also evaluate models with ac- (2) Reflexive pronoun across an object relative curacies. Because there are three classes of string, clause: rather than two, a model is said to have made the correct prediction if the ungrammatical data receive a. The manager that the architects like a lower score than both the baseline and distractor doubted himself. data. But because there are more than three strings, b. *The manager that the architects like they do not compare individual scores from the can- doubted themselves. didate model, but rather the three means that result These pairs deal with referential agreement in spe- from averaging the conditional pronoun probabili- cific syntactic environments. If a model assigns the ties of the baseline, distractor and ungrammatical grammatical string in a pair a higher score than the strings, respectively. ungrammatical string, then we say that the model This alternative design not only provided bet- made the correct prediction on that pair. Having ter accuracies than were achieved by Marvin and evaluated the model over a large number of these Linzen(2018), but the inclusion of distractors in pairs, we can compute an accuracy score, relative the design lowers the random baseline from 50% to a 50% random baseline. to 33.3% accuracy. Hu et al.(2020) conclude that Hu et al.(2020) have taken exception to the de- current neural language models are learning more sign of many such evaluations on that grounds that: about the licensing of reflexive anaphora than was (1) a number of English nouns are stereotypically previously thought. gendered, which conditions pronoun choice, and (2) the unigram probabilities of reflexive pronouns 2 Theoretical Exceptions are different, which biases the probabilities that In a typical psycholinguistics experiment, we models assign to sentences that contain them. To would give human subjects a task to perform dur- circumvent these shortcomings, they generalized ing which they would be presented with a stimulus the pairs to larger sets of strings in which multi- that was labelled as either baseline, distractor or ple nouns were used in multiple positions so that ungrammatical. The effect of the stimulus on the lexical choice and order could be permuted across task could be measured by time to completion, the sets. They also introduced distractors, grammatical number of correct tokens retrieved during a fixed strings that contain material irrelevant, or distract- interval of time, etc. Regardless, the task would ing, to the determination of the sentence’s gram- almost certainly be chosen so that samples of its maticality. One set that they use, for example, is: corresponding measure of success would be nor- (1B) The girl said that the mother saw herself. mally distributed. So a within-subjects mean of these quantities is entirely justifiable. (2B) The mother said that the girl saw herself. The situation is somewhat less clear with the (1D) The girls said that the mother saw herself. scores that are returned by a neural language model. Ignoring for the moment that Hu et al.(2020) are (2D) The mothers said that the girl saw herself. interested in conditional pronoun probabilities and 111 not sentence probabilities, the scores are gener- ally not regarded as measures of success on a task per se — there is no actual task here, apart from achieving a high rank in the evaluation. Legiti- mate task performance profiles are defined over separate downstream tasks, such as those in the GLUE leaderboard (Wang et al., 2018). It is rather more difficult to think of downstream tasks that depend on conditional pronoun probabilities, how- ever. Note that for Marvin and Linzen(2018), the Figure 1: Surprisals (negative log probabilities) for ev- ratio of conditional pronoun probabilities of a set ery set in experiment 1b for the GRNN model with her- of stimuli was the same as the ratio of their total self, on the left, and for the TransXL model with them- sentence probabilities because the reflexive pro- selves on the right. noun is always the last word of the sentence, and the contexts preceding the pronouns are always identical. α = 0:05, and marginally successful for an addi- Several papers by Lau et al., culminating in tional 8% at α = 0:1. This means that somewhere Lau et al.(2017), have argued instead that sen- between 20–30% of the sets are provably not nor- tence probabilities can justifiably be interpreted as mal. Homoscedasticity is merely one aspect of gradient grammaticality scores, rejecting the long- normal distributions that can be used to prove that standing assumption in generative linguistics that a distribution is not normal.
Recommended publications
  • The Empirical Base of Linguistics: Grammaticality Judgments and Linguistic Methodology
    UCLA UCLA Previously Published Works Title The empirical base of linguistics: Grammaticality judgments and linguistic methodology Permalink https://escholarship.org/uc/item/05b2s4wg ISBN 978-3946234043 Author Schütze, Carson T Publication Date 2016-02-01 DOI 10.17169/langsci.b89.101 Data Availability The data associated with this publication are managed by: Language Science Press, Berlin Peer reviewed eScholarship.org Powered by the California Digital Library University of California The empirical base of linguistics Grammaticality judgments and linguistic methodology Carson T. Schütze language Classics in Linguistics 2 science press Classics in Linguistics Chief Editors: Martin Haspelmath, Stefan Müller In this series: 1. Lehmann, Christian. Thoughts on grammaticalization 2. Schütze, Carson T. The empirical base of linguistics: Grammaticality judgments and linguistic methodology 3. Bickerton, Derek. Roots of language ISSN: 2366-374X The empirical base of linguistics Grammaticality judgments and linguistic methodology Carson T. Schütze language science press Carson T. Schütze. 2019. The empirical base of linguistics: Grammaticality judgments and linguistic methodology (Classics in Linguistics 2). Berlin: Language Science Press. This title can be downloaded at: http://langsci-press.org/catalog/book/89 © 2019, Carson T. Schütze Published under the Creative Commons Attribution 4.0 Licence (CC BY 4.0): http://creativecommons.org/licenses/by/4.0/ ISBN: 978-3-946234-02-9 (Digital) 978-3-946234-03-6 (Hardcover) 978-3-946234-04-3 (Softcover) 978-1-523743-32-2
    [Show full text]
  • CAS LX 522 Syntax I
    It is likely… CAS LX 522 IP This satisfies the EPP in Syntax I both clauses. The main DPj I′ clause has Mary in SpecIP. Mary The embedded clause has Vi+I VP is the trace in SpecIP. V AP Week 14b. PRO and control ti This specific instance of A- A IP movement, where we move a likely subject from an embedded DP I′ clause to a higher clause is tj generally called subject raising. I VP to leave Reluctance to leave Reluctance to leave Now, consider: Reluctant has two θ-roles to assign. Mary is reluctant to leave. One to the one feeling the reluctance (Experiencer) One to the proposition about which the reluctance holds (Proposition) This looks very similar to Mary is likely to leave. Can we draw the same kind of tree for it? Leave has one θ-role to assign. To the one doing the leaving (Agent). How many θ-roles does reluctant assign? In Mary is reluctant to leave, what θ-role does Mary get? IP Reluctance to leave Reluctance… DPi I′ Mary Vj+I VP In Mary is reluctant to leave, is V AP Mary is doing the leaving, gets Agent t Mary is reluctant to leave. j t from leave. i A′ Reluctant assigns its θ- Mary is showing the reluctance, gets θ roles within AP as A θ IP Experiencer from reluctant. required, Mary moves reluctant up to SpecIP in the main I′ clause by Spellout. ? And we have a problem: I vP But what gets the θ-role to Mary appears to be getting two θ-roles, from leave, and what v′ in violation of the θ-criterion.
    [Show full text]
  • Greek and Latin Roots, Prefixes, and Suffixes
    GREEK AND LATIN ROOTS, PREFIXES, AND SUFFIXES This is a resource pack that I put together for myself to teach roots, prefixes, and suffixes as part of a separate vocabulary class (short weekly sessions). It is a combination of helpful resources that I have found on the web as well as some tips of my own (such as the simple lesson plan). Lesson Plan Ideas ........................................................................................................... 3 Simple Lesson Plan for Word Study: ........................................................................... 3 Lesson Plan Idea 2 ...................................................................................................... 3 Background Information .................................................................................................. 5 Why Study Word Roots, Prefixes, and Suffixes? ......................................................... 6 Latin and Greek Word Elements .............................................................................. 6 Latin Roots, Prefixes, and Suffixes .......................................................................... 6 Root, Prefix, and Suffix Lists ........................................................................................... 8 List 1: MEGA root list ................................................................................................... 9 List 2: Roots, Prefixes, and Suffixes .......................................................................... 32 List 3: Prefix List ......................................................................................................
    [Show full text]
  • Language Structure: Phrases “Productivity” a Property of Language • Definition – Language Is an Open System
    Language Structure: Phrases “Productivity” a property of Language • Definition – Language is an open system. We can produce potentially an infinite number of different messages by combining elements differently. • Example – Words into phrases. An Example of Productivity • Human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features. (24 words) • I think that human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features, which are fascinating in and of themselves. (33 words) • I have always thought, and I have spent many years verifying, that human language is a communication system that bears some similarities to other animal communication systems, but is also characterized by certain unique features, which are fascinating in and of themselves. (42 words) • Although mainstream some people might not agree with me, I have always thought… Creating Infinite Messages • Discrete elements – Words, Phrases • Selection – Ease, Meaning, Identity • Combination – Rules of organization Models of Word reCombination 1. Word chains (Markov model) Phrase-level meaning is derived from understanding each word as it is presented in the context of immediately adjacent words. 2. Hierarchical model There are long-distant dependencies between words in a phrase, and these inform the meaning of the entire phrase. Markov Model Rule: Select and concatenate (according to meaning and what types of words should occur next to each other). bites bites bites Man over over over jumps jumps jumps house house house Markov Model • Assumption −Only adjacent words are meaningfully (and lawfully) related.
    [Show full text]
  • 24.902F15 Class 13 Pro Versus
    Unpronounced subjects 1 We have seen PRO, the subject of infinitives and gerunds: 1. I want [PRO to dance in the park] 2. [PRO dancing in the park] was a good idea 2 There are languages that have unpronounced subjects in tensed clauses. Obviously, English is not among those languages: 3. irθe (Greek) came.3sg ‘he came’ or ‘she came’ 4. *came Languages like Greek, which permit this phenomenon are called “pro-drop” languages. 3 What is the status of this unpronounced subject in pro-drop languages? Is it PRO or something else? Something with the same or different properties from PRO? Let’s follow the convention of calling the unpronounced subject in tensed clauses “little pro”, as opposed to “big PRO”. What are the differences between pro and PRO? 4 -A Nominative NP can appear instead of pro. Not so for PRO: 5. i Katerina irθe the Katerina came.3sg ‘Katerina came’ 6. *I hope he/I to come 7. *he to read this book would be great 5 -pro can refer to any individual as long as Binding Condition B is respected. That is, pro is not “controlled”. Not so for OC PRO. 8. i Katerinak nomizi oti prok/m irθe stin ora tis the K thinks that came on-the time her ‘Katerinak thinks that shek/m /hem came on time’ 9. Katerinak wants [PROk/*m to leave] 6 -pro can yield sloppy or strict readings under ellipsis, exactly like pronouns. Not so OC PRO. 10. i Katerinak nomizi oti prok/m irθe stin ora tis the K thinks that came on-the time her ke i Maria episis and the Maria also ‘Katerinak thinks that shek came on time and Maria does too’ =…Maria thinks that Katerina came on time strict …Maria thinks that Maria came on time sloppy 7 11.
    [Show full text]
  • Final Review: Syntax Fall 2007
    Final Review: Syntax Fall 2007 Jean Mark Gawron San Diego State University December 12, 2007 1 Control and Raising Key: S Subject O Object C Control R Raising For example, SOR = Subject(to)-Object Raising. Example answers: 1.1 seem a. Identify the control type [subject/object]. What NP is understood as controller of the infinitive (does or is expected to do or tries to do or ... the action described by the verb in infinitival form) John tries to go Subject SSR, SC John seems to go Subject SSR, SC John is likely to go Subject SSR, SC John is eager to go Subject SSR, SC Mary persuaded John to go Object SOR, OC Mary expected John to go Object SOR, OC Mary promised John to go Subject SSR, SC The control type of seem is subject! b. Produce relevant examples: (1) a. Itseemstoberaining b. There seems to be a problem c. The chips seems to be down. d. It seems to be obvious that John is a fool. e. The police seem to have caught the burglar. f. The burglar seems to have been caught by the police. c. Example construction i. Construct embedded clause: (a) itrains. Simpleexample;dummysubject * John rains Testing dummy subjecthood ittorain Putintoinfinitivalform ii. Embed (a) ∆ seem [CP it to rain] Embed under seem [ctd.] it seem [CP t torain] Moveit— it seems [CP t to rain] Add tense, agreement (to main verb) d. Other examples (b) itisraining. Alternativeexample;dummysubject *Johnisraining Testingdummysubjecthood ittoberaining Putintoinfinitivalform ∆ seem [CP it to be raining] Embed under seem it seem [CP t toberaining] Moveit— it seems [CP t to be raining] Add tense, agreement (to main verb) (c) Thechipsaredown.
    [Show full text]
  • Experimenting with Pro-Drop in Spanish1
    Andrea Pešková Experimenting with Pro-drop in Spanish1 Abstract This paper investigates the omission and expression of pronominal subjects (PS) in Buenos Aires Spanish based on data from a production experiment. The use of PS is a linguistic phenomenon demonstrating that the grammar of a language needs to be considered independently of its usage. Despite the fact that the observed Spanish variety is a consistent pro-drop language with rich verbal agreement, the data from the present study provide evidence for a quite frequent use of overt PS, even in non-focal, non- contrastive and non-ambiguous contexts. This result thus supports previous corpus- based empirical research and contradicts the traditional explanation given by grammarians that overtly realized PS in Spanish are used to avoid possible ambiguities or to mark contrast and emphasis. Moreover, the elicited semi-spontaneous data indicate that the expression of PS is optional; however, this optionality is associated with different linguistic factors. The statistical analysis of the data shows the following ranking of the effects of these factors: grammatical persons > verb semantics > (syntactic) clause type > (semantic) sentence type. 1. Introduction It is well known that Spanish is a pro-drop (“pronoun-dropping”) or null- subject language whose grammar permits the omission of pronominal 1 A preliminary version of this paper was presented at the international conference Variation and typology: New trends in syntactic research in Helsinki (August 2011). I am grateful to the audience for their fruitful discussions and useful commentaries. I would also like to thank the editors, Susann Fischer, Ingo Feldhausen, Christoph Gabriel and the anonymous reviewers for their detailed and helpful comments on an earlier version of this article.
    [Show full text]
  • The Position of Subjects*
    Lingua 85 (1991) 21 l-258. North-Holland 211 The position of subjects* Hilda Koopman and Dominique Sportiche Department of Linguistics, UCLA, Los Angeles, CA 90024, USA Grammatical theories all use in one form or another the concept of canonical position of a phrase. If this notion is used in the syntax, when comparing the two sentences: (la) John will see Bill. (1 b) Bill John will see. we say that Bill occupies its canonical position in (la) but not in (lb). Adopting the terminology of the Extended Standard Theory, we can think of the canonical position of a phrase as its D-structure position. Since the concept of canonical position is available, it becomes legitimate to ask of each syntactic unit in a given sentence what its canonical position is, relative to the other units of the sentence. The central question we address in this article is: what is the canonical position of subjects1 Starting with English, we propose that the structure of an English clause is as in (2): * The first section of this article has circulated as part of Koopman and Sportiche (1988) and is a written version of talks given in various places. It was given in March 1985 at the GLOW conference in Brussels as Koopman and Sportiche (1985), at the June 1985 CLA meeting in Montreal, at MIT and Umass Amherst in the winter of 1986, and presented at UCLA and USC since. The input of these audiences is gratefully acknowledged. The second section is almost completely new. 1 For related ideas on what we call the canonical postion of subjects, see Contreras (1987), Kitagawa (1986) Kuroda (1988), Speas (1986) Zagona (1982).
    [Show full text]
  • GRAMMATICALITY and UNGRAMMATICALITY of TRANSFORMATIVE GENERATIVE GRAMMAR : a STUDY of SYNTAX Oleh Nuryadi Dosen Program Studi Sa
    GRAMMATICALITY AND UNGRAMMATICALITY OF TRANSFORMATIVE GENERATIVE GRAMMAR : A STUDY OF SYNTAX Oleh Nuryadi Dosen Program Studi Sastra Inggris Fakultas Komunikasi, Sastra dan Bahasa Universitas Islam “45” Bekasi Abstrak Tulisan ini membahas gramatika dan kaidah-kaidah yang menentukan struktur kalimat menurut tatabahasa transformasi generatif. Linguis transformasi generatif menggunakan istilah gramatika dalam pengertian yang lebih luas. Pada periode sebelumnya, gramatika dibagi menjadi morfologi dan sintaksis yang terpisah dari fonologi. Namun linguis transformasi generatif menggunakan pengertian gramatika yang mencakup fonologi. Kalimat atau ujaran yang memenuhi kaidah-kaidah gramatika disebut gramatikal, sedangkan yang berhubungan dengan gramatikal disebut kegramatikalan. Kegramatikalan adalah gambaran susunan kata-kata yang tersusun apik (well-formed) sesuai dengan kaidah sintaksis. Kalimat atau ujaran yang tidak tersusun secara well-formed disebut camping atau ill-formed. Gramatika mendeskripsikan dan menganalisis penggalan-penggalan ujaran serta mengelompokkan dan mengklasifikasikan elemen-elemen dari penggalan-penggalan tersebut. Dari pengelompokan tersebut dapat dijelaskan kalimat yang ambigu, dan diperoleh pengelompokkan kata atau frase yang disebut kategori sintaksis. Kegramatikalan tidak tergantung pada apakah kalimat atau ujaran itu pernah didengar atau dibaca sebelumnya. Kegramatikalan juga tidak tergantung apakah kalimat itu benar atau tidak. Yang terakhir kegramatikalan juga tidak tergantung pada apakah kalimat itu penuh arti
    [Show full text]
  • Word Recognition in Two Languages and Orthographies: English and Greek
    Memory & Cognition 1994, 22 (3), 313-325 Word recognition in two languages and orthographies: English and Greek HELENA-FIVI CHITIRI St. Francis Xavier University, Antigonish, Nova Scotia, Canada and DALE M. WILLOWS Ontario Institute for Studies in Education, Toronto, Ontario, Canada Word recognition processes of monolingual readers of English and of Greek were examined with respect to the orthographic and syntactic characteristics of each language. Because of Greek's direct letter-to-sound correspondence, which is unlike the indirect representation of English, the possibility was raised of a greater influence of the phonological code in Greek word recognition. Because Greek is an inflected language, whereas English is a word order language, it was also possible that syntax might influence word recognition patterns in the two languages differen­ tially. These cross-linguistic research questions were investigated within the context of a letter cancellation paradigm. The results provide evidence that readers are sensitive to both the ortho­ graphic and the linguistic idiosyncracies of their language. The results are discussed in terms of the orthographic depth hypothesis and the competition model. In this research, we examined word recognition in En­ ber (singular, plural), and the case (nominative, genitive, glish and Greek, two languages with different orthogra­ accusative, vocative) of the noun. Verb endings provide phies and syntactic characteristics. We examined the rela­ information about person and number as well as about tionship of orthography to the role of the phonological the voice and the tense of a verb. code in word recognition as it is exemplified in the pro­ Greek words are typically polysyllabic (Mirambel, cessing of syllabic and stress information.
    [Show full text]
  • Pronominal Vs. Anaphoric Pro in Kannada
    Pronominal vs. anaphoric pro in Kannada Anuradha Sudharsan∗ The English and Foreign Languages University Hyderabad, Telangana, India Kannada licenses a pronominal pro and an anaphor pro in root and subordinate clauses. In the subordi- nate clauses, pro’s person feature largely determines its pronominal/anaphoric status. Accordingly, there are four types of pro, each with a distinct referential property. The 1st and 2nd person pros allow pronomi- nal and anaphoric interpretations. In their pronominal reference, they refer to the speaker and the listener, respectively. In their anaphoric reference, they refer to the matrix subject and the object, respectively, irre- spective of the latter’s person feature, which results in a feature mismatch between pro and its antecedent. The 3rd person pro allows only pronominal interpretation. Kannada quotative verbs report direct speech. The null subjects in the embedded clauses of reported speech are basically pronominal because they are ‘copies’ of the pronominal subjects in direct speech. Accordingly, the embedded verbal inflection corre- sponds to inflection in direct speech, which results in a feature clash between pro and its antecedent. The fourth type, a null anaphor, occurs in non-argument endu-clauses. It is bound by an NP in the matrix with which it shares its F-features. It is the semantic relation between the non-argument and main clauses that explains the presence of an anaphoric pro here. Keywords: pro, pronominal, feature clash, anaphor, Kannada 1. Introduction In most studies on the pro-drop parameter, pro was characterized as a pronominal empty category based on evidence from European null subject languages.
    [Show full text]
  • Experimental Methods in Studying Child Language Acquisition Ben Ambridge∗ and Caroline F
    View metadata, citation and similar papers at core.ac.uk brought to you by CORE provided by MPG.PuRe Advanced Review Experimental methods in studying child language acquisition Ben Ambridge∗ and Caroline F. Rowland This article reviews the some of the most widely used methods used for studying children’s language acquisition including (1) spontaneous/naturalistic, diary, parental report data, (2) production methods (elicited production, repeti- tion/elicited imitation, syntactic priming/weird word order), (3) comprehension methods (act-out, pointing, intermodal preferential looking, looking while lis- tening, conditioned head turn preference procedure, functional neuroimaging) and (4) judgment methods (grammaticality/acceptability judgments, yes-no/truth- value judgments). The review outlines the types of studies and age-groups to which each method is most suited, as well as the advantage and disadvantages of each. We conclude by summarising the particular methodological considerations that apply to each paradigm and to experimental design more generally. These include (1) choosing an age-appropriate task that makes communicative sense (2) motivating children to co-operate, (3) choosing a between-/within-subjects design, (4) the use of novel items (e.g., novel verbs), (5) fillers, (6) blocked, coun- terbalanced and random presentation, (7) the appropriate number of trials and participants, (8) drop-out rates (9) the importance of control conditions, (10) choos- ing a sensitive dependent measure (11) classification of responses, and (12) using an appropriate statistical test. © 2013 John Wiley & Sons, Ltd. How to cite this article: WIREs Cogn Sci 2013, 4:149–168. doi: 10.1002/wcs.1215 INTRODUCTION them to select those that are most useful for studying particular phenomena.
    [Show full text]