Child Consonant Harmony Revisited: The Role of Lexical Memory Constraints and Segment Repetition
Young children often produce non-target-like word forms in which non-adjacent consonants share a major place of articulation (e.g., [gɔgi] "doggy"). Termed child consonant harmony (CCH), this phenomenon has garnered considerable attention in the literature, primarily due to the apparent absence of analogous patterns in mature phonological systems. This study takes a close look at a potential account of CCH that is compatible with findings from adult word learning, serial recall, and phonological typology. According to this account, CCH is a response to memory pressure involved in remembering and retrieving multiple consonantal contrasts within a word. If this is the main motivation behind CCH, we would expect the resulting child forms to be biased toward full assimilation (i.e., consonant repetition) as it allows maximal reduction of phonolexical memory load. To test this prediction, children's productions of target words containing consonants that differ in both major place and manner were analyzed using two data sources: a single session sample from 40 children aged 1-2 years learning English, French, Finnish, Japanese, or Mandarin; and longitudinal samples from seven English-learning children between 1 and 3 years of age. Prevalence of consonant repetitions was robustly evidenced in early child forms, especially in those produced for target words with the structure CVCV(C). The results suggest that early word production is shaped by constraints on phonolexical memory.
Building a Grammatical Network: Form and Function in the Development of Hebrew Prepositions
This study investigates the emergence of prepositions in Hebrew-speaking children aged 2;6-6;0 years, analyzing a peer talk corpus of 75 children across five age groups. Across 45-minute triadic conversations, we examined the distributions, semantic functions, and form-function relations of prepositions. Two results sections are presented. First, using network analysis, we modeled the development of form-function correlations of Hebrew prepositions. Second, we conducted qualitative developmental analyses of the distributions and semantics of all prepositions identified in the study. Our findings reveal that prepositions expressed 22 functions, predominantly grammatical, spatial, and temporal. With age, the use of prepositions increased, abstract functions became more prevalent, and functions were served by a broader range of prepositions. The data suggest the emergence of systematic relations, forming network-based clusters or communities of semantically related functions. This systematic growth of the prepositional category signifies not just lexical but also syntactic development in Hebrew, transitioning from lexicalized preposition-marked verb arguments to diverse, abstract preposition-marked syntactic adjuncts, which enrich clause-level complexity.
Aptitude, Anxiety, and Success in L2 Speech Development: A Longitudinal Study of Chinese EFL College-Level Learners
This study examined the second language (L2) speech development of a group of Chinese English as a Foreign Language (EFL) college-level learners ( = 83) and the association between their aptitude, anxiety, and L2 speech development. The performance of the participants' L2 speech, including speech comprehensibility and accentedness, was evaluated using a sentence reading task and a pair of picture description tasks (pre- and post-test). In addition, students completed the LLAMA tests (language learning and aptitude measurement in adults) and the Second Language Speaking Anxiety Scale (SLSAS). The -test revealed that after one semester of English speaking and pronunciation instruction, the EFL students in the study demonstrated significant gains in their L2 speech performance in terms of accentedness and comprehensibility in both controlled and impromptu tasks. Regression analyses revealed that phonemic coding ability positively predicted gains in comprehensibility but negatively predicted gains in accentedness during sentence reading. In addition, associative memory was positively associated with comprehensibility gains in the picture description task. Anxiety appeared to be a strong predictor of participants' gains in speech comprehensibility and accentedness during the picture description task.
Cross-Linguistic Phonetic Variation in Bilingual Speech: Cantonese /n/ > [l] Merger in Early Cantonese-English Bilinguals
/n/ is merging with /l/ in Cantonese, as well as in several other Chinese languages. The Cantonese merger appears categorical, with /n/ becoming /l/ syllable-initially. This project aims to describe /n/ and /l/ in Cantonese and English speech from early Cantonese-English bilinguals to better understand the status of the merger in Cantonese and its potential for cross-linguistic mutual influence. We examine early bilinguals' ( = 34) speech using the Speech in Cantonese and English (SpiCE) corpus, focusing on pre-vocalic /n/ and /l/ onsets in both languages. Items were auditorily coded for their perceived category identity, and two acoustic measures anticipated to have the potential to differentiate /n/ and /l/ within and across languages were applied. In English, bilinguals maintained a clear contrast between /n/ and /l/ in the auditory coding and in acoustic measurements. In Cantonese, however, there were higher rates of [l] for /n/ items, in line with the merger, and [n] for /l/ items, indicating hypercorrection of the pattern. Across languages, bilinguals produced language-specific /l/s, but there were no acoustic differences between Cantonese and English /n/. The participation of Cantonese /n/ in a sound change does not appear to compromise English /n/s' patterning, suggesting that Cantonese and English /n/ are maintained as distinct categories in the minds of early bilinguals.
Effects of Orthographic Input and Inhibitory Control on Second-Language Speech Production
The present study extends previous research reporting that orthographic forms, such as the use of a single letter or two letters to indicate the same sound, affect sound duration in second-language (L2) production. Native-language (L1) Korean L2 English sequential bilinguals performed a delayed repetition task for word pairs containing the same consonant or vowel spelled with one or two letters. Korean provided an interesting case because (1) it has an alphabetic orthographic system but not a Roman alphabet and thus, there may be no interorthographic interference and (2) it has no phonemic length contrast for vowels, whereas there is some disagreement on the contrastiveness of the consonant length, which can lead to an asymmetry in the grapheme-to-phoneme correspondence between vowels and consonants. The results showed that Korean learners produced the same English vowel with a short duration when spelled with a single letter and with a long duration when spelled with double letters or digraphs composed of two different letters; this variation in duration did not appear when producing English consonants spelled with a single or two letters. This study further examined whether individual differences in inhibitory control influenced the magnitude of orthographic effects in the production of English vowels by Korean learners. Individual differences in inhibitory control were not strongly related to the influence of orthography on vowel production.
Increased Breathiness in Adolescent Kiezdeutsch Speakers: A Marker of Multiethnolectal Group Affiliation?
Kiezdeutsch is a multiethnolectal variety of German spoken by young people from multicultural communities that exhibits lexical, syntactic, and phonetic differences from standard German. A rather salient and pervasive feature of this variety is the fronting of the standard palatal fricative /ç/ (as in "I") to [ɕ] or [ʃ]. Previous perception work shows that this difference is salient and carries social meaning but dependent on the listener group. Further investigations also point to the significance of /ɔɪ/-fronting in production; however, whether this is salient in perception has not yet been investigated. In several (multi)ethnolectal varieties, differences in voice quality compared to the standard have been identified. Therefore, in this study, we present an acoustic comparison of voice quality in adolescent speakers of Kiezdeutsch and standard German, with results showing that Kiezdeutsch speakers produce a breathier voice quality. In addition, we report on a perception test designed to examine the social meaning of voice quality in combination with two segmental cues: coronalization of /ç/ and /ɔɪ/-fronting. The results indicate perceptual gradience for phonetic alternations detected in Kiezdeutsch with coronalization of /ç/ being a highly salient and reliable marker, whereas fronting of /ɔɪ/ and breathy voice do not appear to be clearly enregistered features of Kiezdeutsch by all listeners. Thus, even though we find differences in production, these may not necessarily be relevant in perception, pointing toward enregisterment- like sound change-being a continuous process of forming learned associations through tokens of experiences.
Pronunciation of Vowel Digraphs in Nonwords: A Replication and Extension
In English, the pronunciation of a vowel digraph can vary; for example, is pronounced /ɛ/ in but /i/ in and /eɪ/ in . We investigated participant-level effects on the pronunciation of ambiguous vowel digraphs in nonwords (e.g., ) using standardized test measures of six reading-related skills in 80 young adults. We employed both an established written task and set of nonword stimuli and a spoken version of the task with the same stimuli. We largely replicated the previously reported pattern of preferred nonword pronunciations in both the written and spoken versions of the task. Generalized linear mixed-effects model analyses revealed that individual differences in phonological memory, spelling knowledge, and word reading efficiency contributed to pronunciation choice beyond item-level effects. Overall, taken together with the results of item-level analyses, our findings are consistent with models and theories in which specific reading-related skills, intralexical context, and interlexical pronunciation knowledge influence ambiguous vowel digraph pronunciation in nonword reading in fluently reading young adults.
Segmental Influences on the Perception of High Pitch Accent Scaling in American English
Researchers investigating a broad array of questions in spoken language prosody routinely base their arguments on measurements taken from the F0 contours of representative speech samples. These analyses, however, frequently involve abstracting F0 contours away from the segmental strings that bear them, potentially overlooking in the process the role played by segmental qualities such as sonority or periodicity in the realization of F0 patterns by speakers and their interpretation by listeners. This paper reports the results of two experiments investigating how perception of F0 contours is affected by the segmental string over which those contours are realized. The first focuses on gaps in F0 contours created by voiceless obstruents such as stops and fricatives, while the second investigates F0 intervals spanning lower-sonority voiced segments, such as nasals and voiced fricatives. While these two scenarios might at first seem unrelated, we argue that listeners treat both with a single mechanism in perception, namely, by reducing (potentially to zero) the amount of weight accorded to those portions of the contour for determination of the speaker's intended F0 scaling level. We present an account of both effects within a unified model of F0 scaling perception called TCoG-F, with discussion of its implications for phonetic and phonological intonation research going forward.
Importance of Visual Support Through Lipreading in the Identification of Words in Spanish Language
We sought to examine the contribution of visual cues, such as lipreading, in the identification of familiar (words) and unfamiliar (phonemes) words in terms of percent accuracy. For that purpose, in this retrospective study, we presented lists of words and phonemes (adult female healthy voice) in auditory (A) and audiovisual (AV) modalities to 65 Spanish normal-hearing male and female listeners classified in four age groups. Our results showed a remarkable benefit of AV information in word and phoneme recognition. Regarding gender, women exhibited better performance than men in both A and AV modalities, although we only found significant differences for words but not for phonemes. Concerning age, significant differences were detected in word recognition in the A modality between the youngest (18-29 years old) and oldest (⩾50 years old) groups only. We conclude visual information enhances word and phoneme recognition and women are more influenced by visual signals than men in AV speech perception. On the contrary, it seems that, overall, age is not a limiting factor for word recognition, with no significant differences observed in the AV modality.
An Investigation of Language-Specific and Orthographic Effects in L2 Arabic geminate production by Advanced Japanese- and English-speaking learners
Research has indicated that second-language learners have difficulty producing geminates accurately. Previous studies have also shown an effect of orthography on second-language speech production. We tested whether the existence of a contrast in the first language phonology for length aids the second-language production of the same contrast. Furthermore, we examined the effect of exposure to orthographic input on geminate consonant production in a cross-script context. We tested the production of Arabic geminate-singleton stop consonants [/bː/-/b/, /tː/-/t/, /dː/-/d/, and /kː/-/k/], a nasal stop consonant /mː/-/m/, and an emphatic stop consonant /tˤː/-/tˤ/, as well as the effect of the diacritic used in Arabic to mark gemination in a delayed imitation task and two reading tasks (ortho-with diacritics and ortho-without diacritics). A comparison of the productions of advanced Japanese-speaking learners, English-speaking learners, and an Arabic control group showed that both learner groups were able to produce Arabic geminate stops; however, the Japanese-speaking learners exhibited an advantage over the English-speaking learners in the auditory-only task and in the presence of diacritics, highlighting the fact that orthographic effects may occur in some cross-script contexts.
The Case for a Quantitative Approach to the Study of Nonnative Accent Features
Research with nonnative speech spans many different linguistic branches and topics. Most studies include one or a few well-known features of a particular accent. However, due to a lack of empirical studies, little is known about how common these features are among nonnative speakers or how uncommon they are among native speakers. Moreover, it remains to be seen whether findings from such studies generalize to lesser-known features. Here, we demonstrate a quantitative approach to study nonnative accent features using Dutch-accented English as an example. By analyzing the phonetic distances between transcriptions of speech samples, this approach can identify the features that best distinguish nonnative from native speech. In addition, we describe a method to test hypotheses about accent features by checking whether the prevalence of the features overall varies between native and nonnative speakers. Furthermore, we include English speakers from the United States and United Kingdom and native Dutch speakers from Belgium and The Netherlands to address the issue of regional accent variability in both the native and target language. We discuss the results concerning three observed features. Overall, the results provide empirical support for some well-known features of Dutch-accented English, but suggest that others may be infrequent among nonnatives or in fact frequent among natives. In addition, the findings reveal potentially new accent features, and factors that may modulate the expression of known features. Our study demonstrates a fruitful approach to study nonnative accent features that has the potential to expand our understanding of the phenomenon of accent.
Incorporating Frequency Effects in the Lexical Access of Mandarin Tone 3 Sandhi
Mandarin tone 3 sandhi refers to the phenomenon whereby a tone 3 syllable changes to a tone 2 when followed by another tone 3. This phonological process creates a deviation between the tonal forms realized at morphemic (/tone3-tone3/) and word ([tone2-tone3]) levels, posing questions in terms of how disyllabic tone 3 sandhi words are represented and accessed. The current study conducted three cross-modal lexical decision priming experiments to investigate this issue. Experiment 1 manipulated the frequencies of the initial morpheme and whole word, showing that the higher initial-character frequency against the whole word gives stronger activation to the underlying representation and the lower frequency of the initial character leads to stronger activation of the surface tone. Experiments 2 and 3 operationalized the relative frequency of the initial tone 3 morpheme's realization as a sandhi tone, finding that the competition between the two tonal realizations also influences how T3 sandhi words are accessed. Specifically, the more frequently the T3 morpheme surfaces as a T2 allomorph, the less activated the underlying representation becomes in the mental lexicon. Our results indicate a complex interplay between morpheme, word, and the associated tonal representations in the mental lexicon and that these factors co-determine the lexical access of tone 3 sandhi.
Modeling Lexical Tones for Speaker Discrimination
Fundamental frequency (F0) has been widely studied and used in the context of speaker discrimination and forensic voice comparison casework, but most previous studies focused on long-term F0 statistics. Lexical tone, the linguistically structured and dynamic aspects of F0, has received much less research attention. A main methodological issue lies on how tonal F0 should be parameterized for the best speaker discrimination performance. This paper compares the speaker discriminatory performance of three approaches with lexical tone modeling: discrete cosine transform (DCT), polynomial curve fitting, and quantitative target approximation (qTA). Results show that using parameters based on DCT and polynomials led to similarly promising performance, whereas those based on qTA generally yielded relatively poor performance. Implications modeling surface tonal F0 and the underlying articulatory processes for speaker discrimination are discussed.
Incidental Learning of Collocations Under Different Input Modes and the Mediating Role of Perceptual Learning Style
This study investigated how input modes (reading vs. listening) and learners' perceptual learning style (visual vs. auditory) affected the incidental learning of collocations. A total of 182 college students were first assigned to either a visual or auditory group based on their performance on a perceptual learning style questionnaire. Each style group was subsequently subdivided into three groups who were exposed to a series of texts containing unfamiliar collocation items under one of the input conditions: written input, aural input, or no input. Results of the study indicated that both written and aural input led to gains in collocational knowledge, and aural input was more effective than written input. Furthermore, the study provided empirical evidence that there was a moderating role of perceptual learning style on incidental collocation learning. The auditory learners under aural input showed the highest rate of collocation learning among all treatment subgroups.
Aspiring to Aspirate: L2 Acquisition of English Word-Initial /p/ Over 10 Years
This short report describes a longitudinal examination of the acquisition of English-aspirated stops by an initial cohort of 24 adult Slavic-language (Russian, Ukrainian, and Croatian) speakers. All had arrived in Canada with low oral English proficiency, and all were enrolled in the same language instruction program at the outset. Initial bilabial stops in CVCs were recorded at eight testing times: six during the first year of the study, again at year 7, and finally at year 10. Intelligibility was evaluated through a blind listening assessment of the stop productions from the first seven testing times. Voice onset times (VOT) were measured for /p/ from all eight times. Mean /p/ intelligibility improved-mainly during a proposed Window of Maximal Opportunity for L2 speech acquisition-but remained below 100%, even after 7 years. For some speakers, early /p/ productions were minimally aspirated, with VOT increasing over time but remaining intermediate between L1 English and L1 Slavic-Language values at 10 years. However, inter-speaker variability was dramatic, with some speakers showing full intelligibility throughout the study and others showing many unintelligible productions at all times. Individual learning trajectories tended to be non-linear and often non-cumulative. Overall, these findings point to a developmental process that varies considerably from one learner to another. It also demonstrates the serious drawbacks of relying on group means to characterize the process of L2 segmental learning.
The Articulatory and Acoustic Representation of Second-Language French Vowels
This study examines how L1 English-L2 French learners use L1 articulatory and acoustic categories to produce L2 vowels that are both similar to and different from their L1 vowels. Previous studies examining the relationship between L1 and L2 sound inventories have found that learners reuse L1 phone categories to produce L2 phones that are perceived as similar, but importantly, there is a lack of articulatory data included in these types of studies, which has reinforced the assumption that vowel categories can be solely represented by their acoustic properties. The present study uses ultrasound tongue imaging data and videos of lip rounding in addition to acoustic data to examine how L1 English-L2 French learners produce the French vowels /i y u e ø o/ compared with their English vowels /i u e o/. The results focus on individual paths to category formation to show how learners articulate L2 vowels, and reveal that they tend to reuse L1 tongue body gestures to produce the French vowels /i u e o/, and lip rounding gestures to produce the round vowels /y u o/. This study demonstrates that transfer of articulatory gestures depends on vowel quality and emphasizes the importance of using articulatory data to inform theories of L2 category formation.
Accommodation and Language Contact
The paper introduces the Special Issue on , which originates from the conference Phonetics and Phonology in Europe (PaPE) held at the University of Lecce, Italy, in 2019. It discusses the topics of language contact and speaker accommodation, summarizing the contributions included in the Special Issue, and arguing explicitly in favour of a unitary view of how both temporary and stable changes happen in (part of) the linguistic systems. Accommodation is seen as the same gradual and non-homogeneous process at play in different contact settings. In the introductory sections, a discussion is offered on various situations in which linguistic systems are in contact and on the main factors that may be at play; the following sections offer an overview of the papers included in the Special Issue, which focus on accommodation in L2 and heritage speakers as well as on the time dimension of dialect or language societal contact. Finally, accommodation is discussed as the same process that is at work in any interaction, that may modify temporarily or long-term the system of L2 learners and bilinguals (e.g., immigrants), that usually affects in the long-term the heritage speakers' system, and that only in the long term can lead to language changes involving entire communities.
English Speakers' Perception of Non-native Vowel Contrasts in Adverse Listening Conditions: A Discrimination Study on the German Front Rounded Vowels /y/ and /ø/
Previous research has shown that it is difficult for English speakers to distinguish the front rounded vowels /y/ and /ø/ from the back rounded vowels /u/ and /o/. In this study, we examine the effect of noise on this perceptual difficulty. In an Oddity Discrimination Task, English speakers without any knowledge of German were asked to discriminate between German-sounding pseudowords varying in the vowel both in quiet and in white noise at two signal-to-noise ratios (8 and 0 dB). In test trials, vowels of the same height were contrasted with each other, whereas a contrast with /a/ served as a control trial. Results revealed that a contrast with /a/ remained stable in every listening condition for both high and mid vowels. When contrasting vowels of the same height, however, there was a perceptual shift along the F2 dimension as the noise level increased. Although the /ø/-/o/ and particularly /y/-/u/ contrasts were the most difficult in quiet, accuracy on /i/-/y/ and /e/-/ø/ trials decreased immensely when the speech signal was masked. The German control group showed the same pattern, albeit less severe than the non-native group, suggesting that even in low-level tasks with pseudowords, there is a native advantage in speech perception in noise.
Audiovisual Perception of Lexical Stress: Beat Gestures and Articulatory Cues
Human communication is inherently multimodal. Auditory speech, but also visual cues can be used to understand another talker. Most studies of audiovisual speech perception have focused on the perception of speech segments (i.e., speech sounds). However, less is known about the influence of visual information on the perception of suprasegmental aspects of speech like lexical stress. In two experiments, we investigated the influence of different visual cues (e.g., facial articulatory cues and beat gestures) on the audiovisual perception of lexical stress. We presented auditory lexical stress continua of disyllabic Dutch stress pairs together with videos of a speaker producing stress on the first or second syllable (e.g., articulating or ). Moreover, we combined and fully crossed the face of the speaker producing lexical stress on either syllable with a gesturing body producing a beat gesture on either the first or second syllable. Results showed that people successfully used visual articulatory cues to stress in muted videos. However, in audiovisual conditions, we were not able to find an effect of visual articulatory cues. In contrast, we found that the temporal alignment of beat gestures with speech robustly influenced participants' perception of lexical stress. These results highlight the importance of considering suprasegmental aspects of language in multimodal contexts.
Impact of Japanese L1 Rhythm on English L2 Speech
The study aimed to examine whether L1 speech rhythm affects L2 speech by assessing how the speech rhythm of Japanese L2 English speakers differed from native speakers. We chose Japanese and English because they differ markedly in the phonological properties that likely contribute to speech rhythm. Speech rhythm was measured by the variability of vowel and consonant intervals using rate-normalized rhythm metrics (VarcoV and VarcoC; nPVI-V and nPVI-C) and %V. The study utilized recordings of spoken sentences in English by 10 native Australian English speakers; and in English and also in Japanese by 10 native Japanese speakers (who had limited experience in speaking English). Experiment 1 compared the rhythm of L1 English (by measuring 1,750 vowels and 3,093 consonants from 20 sentences) and L1 Japanese (1,923 vowels and 2,097 consonants from 10 sentences). The results showed that for all measures, Japanese had reduced durational variability in both consonant and vowel intervals compared with English. In Experiment 2, we examined the rhythm characteristics of L1 and L2 English using 40 sentences (including the 20 in Experiment 1). The results showed that vowel and consonant intervals were less variable in L2 (Japanese English) than in L1 (Australian English) speech, mirroring the results of Experiment 1. Overall, the results are consistent with the proposal that L1 (Japanese) speech rhythm influenced L2 (English) speech.
Effects of Speaking Rate Changes on Speech Motor Variability in Adults
The relationship between speaking rate and speech motor variability was examined in three groups of neurotypical adults, = 40; 15 young adults (18-30 years), 13 adults (31-40 years), and 12 middle-aged adults (41-50 years). Participants completed a connected speech task at three speaking rates (habitual, fast, and slow) where kinematic (lower lip movement) and acoustic data were obtained. Duration and variability were measured at each speaking rate. Findings revealed a complex relationship between speaking rate and variability. Adults from the middle age range (31-40 years) demonstrated shorter acoustic and kinematic durations compared with the oldest age group (41-50 years) during the habitual speaking rate condition. All adults demonstrated the greatest variability in the slow speaking rate condition, with no significant differences in variability between habitual and fast speaking rates. Interestingly, lip aperture variability was significantly lower in the youngest age group (18-30 years) compared with the two older groups during the fast speaking rate condition. Differences in measures of acoustic variability were not observed across the age levels. Strong negative correlations between kinematic/acoustic duration and lip aperture/acoustic variability in the youngest age group were revealed. Therefore, while a slow speaking rate does result in greater variability compared with habitual and fast speaking rates, longer durations of productions by the different age groups were not linked to higher spatiotemporal index (STI) values, suggesting that timing influences speech motor variability, but is not the sole contributor.