Note: When clicking on a Digital Object Identifier (DOI) number, you will be taken to an external site maintained by the publisher.
                                            Some full text articles may not yet be available without a charge during the embargo (administrative interval).
                                        
                                        
                                        
                                            
                                                
                                             What is a DOI Number?
                                        
                                    
                                
Some links on this page may take you to non-federal websites. Their policies may differ from this site.
- 
            Abstract Computational models of infant word‐finding typically operate over transcriptions of infant‐directed speech corpora. It is now possible to test models of word segmentation on speech materials, rather than transcriptions of speech. We propose that such modeling efforts be conducted over the speech of the experimental stimuli used in studies measuring infants' capacity for learning from spoken sentences. Correspondence with infant outcomes in such experiments is an appropriate benchmark for models of infants. We demonstrate such an analysis by applying the DP‐Parser model of Algayres and colleagues to auditory stimuli used in infant psycholinguistic experiments by Pelucchi and colleagues. The DP‐Parser model takes speech as input, and creates multiple overlapping embeddings from each utterance. Prospective words are identified as clusters of similar embedded segments. This allows segmentation of each utterance into possible words, using a dynamic programming method that maximizes the frequency of constituent segments. We show that DP‐Parse mimics American English learners' performance in extracting words from Italian sentences, favoring the segmentation of words with high syllabic transitional probability. This kind of computational analysis over actual stimuli from infant experiments may be helpful in tuning future models to match human performance.more » « less
- 
            Abstract Psycholinguistic research on children's early language environments has revealed many potential challenges for language acquisition. One is that in many cases, referents of linguistic expressions are hard to identify without prior knowledge of the language. Likewise, the speech signal itself varies substantially in clarity, with some productions being very clear, and others being phonetically reduced, even to the point of uninterpretability. In this study, we sought to better characterize the language‐learning environment of American English‐learning toddlers by testing how well phonetic clarity and referential clarity align in infant‐directed speech. Using an existing Human Simulation Paradigm (HSP) corpus with referential transparency measurements and adding new measures of phonetic clarity, we found that the phonetic clarity of words’ first mentions significantly predicted referential clarity (how easy it was to guess the intended referent from visual information alone) at that moment. Thus, when parents’ speech was especially clear, the referential semantics were also clearer. This suggests that young children could use the phonetics of speech to identify globally valuable instances that support better referential hypotheses, by homing in on clearer instances and filtering out less‐clear ones. Such multimodal “gems” offer special opportunities for early word learning. Research HighlightsIn parent‐infant interaction, parents’ referential intentions are sometimes clear and sometimes unclear; likewise, parents’ pronunciation is sometimes clear and sometimes quite difficult to understand.We find that clearer referential instances go along with clearer phonetic instances, more so than expected by chance.Thus, there are globally valuable instances (“gems”) from which children could learn about words’ pronunciations and words’ meanings at the same time.Homing in on clear phonetic instances and filtering out less‐clear ones would help children identify these multimodal “gems” during word learning.more » « less
- 
            Abstract To efficiently recognize words, children learning an intonational language like English should avoid interpreting pitch‐contour variation as signaling lexical contrast, despite the relevance of pitch at other levels of structure. Thus far, the developmental time‐course with which English‐learning children rule out pitch as a contrastive feature has been incompletely characterized. Prior studies have tested diverse lexical contrasts and have not tested beyond 30 months. To specify the developmental trajectory over a broader age range, we extended a prior study (Quam & Swingley, 2010), in which 30‐month‐olds and adults disregarded pitch changes, but attended to vowel changes, in newly learned words. Using the same phonological contrasts, we tested 3‐ to 5‐year‐olds, 24‐month‐olds, and 18‐month‐olds. The older two groups were tested using the language‐guided‐looking method. The oldest group attended to vowels but not pitch. Surprisingly, 24‐month‐olds ignored not just pitch but sometimes vowels as well—conflicting with prior findings of phonological constraint at 24 months. The youngest group was tested using the Switch habituation method, half with additional phonetic variability in training. Eighteen‐month‐olds learned both pitch‐contrasted and vowel‐contrasted words, whether or not additional variability was present. Thus, native‐language phonological constraint was not evidenced prior to 30 months (Quam & Swingley, 2010). We contextualize our findings within other recent work in this area.more » « less
- 
            Goldwater, M; Anggoro, FK; Hayes, BK; Ong, DC (Ed.)Before they start to talk, infants learn the form and meaning of many common words. In the present work, we investigated the nature of this word knowledge, testing the specificity of very young infants’ (6-14 months) phonological represen- tations in an internet-based language-guided-looking task using correct pronunciations and initial-consonant mispronunciations of common words. Across the current sample (n=78 out of 96 pre-registered), infants’ proportion looking to the target (named image) versus the distracter was significantly lower when the target word was mispronounced, indicating sensitivity to phonological deviation. Performance patterns varied by age group. The youngest group (6-8 months, n=30) was at chance in both conditions, the middle group (9-11 months, n=21) showed significant recognition of correct pronunciations and a marginal mispronunciation effect, and the oldest age group (12-14 months, n=27) demonstrated the mature pattern: significant recognition and a significant mispronunciation effect. Ongoing work is completing the pre-registered sample size.more » « less
- 
            Children are adept at learning their language’s speech-sound categories, but just how these categories function in their developing lexicon has not been mapped out in detail. Here, we addressed whether, in a language-guided looking procedure, 2-year-olds would respond to a mispronunciation of the voicing of the initial consonant of a newly learned word. First, to provide a baseline of mature native-speaker performance, adults were taught a new word under training conditions of low prosodic variability. In a second experiment, 24- and 30-month-olds were taught a new word under training conditions of high or low prosodic variability. Children and adults showed evidence of learning the taught word. Adults’ target looking was reduced when the novel word was realized at test with a change in the voicing of the initial consonant, but children did not show any such decrement in target fixation. For both children and adults, most learners did not treat the pho- nologically distinct variant as a different word. Acoustic-phonetic variability during teaching did not have consistent effects. Thus, under conditions of intensive short-term training, 24- and 30-month-olds did not differentiate a newly learned word from a variant differing only in consonant voicing. High task complexity during training could explain why mispronunciation detec- tion was weaker here than in some prior studies.more » « less
 An official website of the United States government
An official website of the United States government 
				
			 
					 
					
 
                                     Full Text Available
                                                Full Text Available