Introduction: In the dynamic field of phonetic research, finding innovative ways to collect data can be both a challenge and an opportunity. At our P&P group, Cong, Yanyu, and Damar are taking a leap forward with their latest study, “Gamifying Phonetic Data Collection: An Accent Identification and Attitude Study.” This blog post delves into how they are incorporating gamification into phonetics, transforming the way they understand accents and linguistic attitudes.
The Challenge of Traditional Data Collection: Traditionally, collecting phonetic data has been a straightforward yet often tedious task, both for researchers and participants. Recognizing this, Cong and her colleagues wondered: Could the principles of gaming enhance the data collection process in phonetic studies?
The Innovative Approach: The study is pioneering in its approach, merging the realms of gaming with linguistic research. The core aims include:
Evaluating the effectiveness of gamified methods in phonetic data collection.
Investigating the influence of geographical location on accent identification and attitudes.
Gamification in Action: One of the highlights of the project is the use of a game-based study by identifying native English accents used in the speech on the map of the UK. The leader board and the points shown at the end of each block function as the motivations for participants. This innovative tool not only engages participants in a unique way but also offers deeper insights into how accents are perceived and understood in various geographical contexts.
Preliminary Findings and Insights: While it was a small-scale study in which participants were mainly members of our group, initial findings suggest that gamification can significantly improve participant engagement and the quality of data collected. These insights have the potential to reshape how we conduct phonetic research in the future.
Thoughts From the Group:
Ghada: Researchers in the fields of phonetics and phonology have long recognized the accent bias issues in everyday activities. Now, it seems to be the right time to move a step forward and explore what we, as phonoticians or linguists, should do next.
Fengting: In many sociolinguistic studies on language attitudes, as well as perceptual studies on accentual adaptation and generalization, researchers often do not provide details on how they recruit talkers, and standards have not been set for the selection of accented speech used in this research. This raises the question of how we can be sure that we are researching the actual accents that interest us. A further question is: from whom can accented speech be considered a good representation of the accent?
Conclusion: With continuous exploration in the realm of gamified phonetic studies, we’re enthusiastic about the prospects this approach holds. Stay tuned for more updates and detailed findings in the future.
Your thoughts, feedback, and insights are always welcome!
For this week’s Prosody Interest Group (PIG), we are excited to invite Dr. Ricardo Napoleão de Souza from the University of Edinburgh to talk about his recent research. Focusing on “Domain-Initial Strengthening from an Acoustic Perspective,” Dr. de Souza presented groundbreaking findings using data from varieties of English, Spanish, and Portuguese. It’s online and easy to sign up. Everyone interested is absolutely welcome to join!
From August 7th to 11th, our research group embarked on an illuminating expedition to the 20th International Congress of the Phonetic Sciences (ICPHS) 2023. As one of the most significant congresses in the field of phonetics, the ICPHS is a pivotal event that only occurs once every four years. We were proud to have nine active participants from our group in attendance. Of these, three members delivered compelling oral presentations, while the rest engaged the academic community through insightful poster presentations.
Within our research group, the conference served as a testament to our unity and collective strength. Every individual contributed to the tapestry of success, and the supportive atmosphere propelled each of us to excel. As questions flowed from the audience, they illuminated the depth of engagement and curiosity that permeated the conference. More details of what we presented in the congress will be provided in the following. This post also aims to encapsulate our experiences, learnings, and the milestones achieved during this prestigious event.
Objectives and Expectations
The overarching theme of ICPHS 2023 was “Intermingling Communities and Changing Cultures,” which deeply resonates with the emerging dynamics of our interconnected world. Over the past few decades, there has been an unprecedented surge in mobility and interpersonal contacts, disrupting the boundaries of national languages and impacting speech patterns universally.
Our primary objective for attending the congress was twofold. Firstly, we aimed to share our own insights and research findings with a broader academic community. We were particularly eager to contribute to the ongoing dialogue about how modern societal shifts are influencing phonetics and phonology. Secondly, we were excited to learn from other leading researchers in the field. We wanted to grasp what constitutes ‘trendy’ research currently and to understand how the academic discourse in this field is changing and evolving.
Moreover, we were keen to explore potential directions for future research and possible collaborative efforts. Given that the congress serves as a melting pot of ideas and innovations, we were optimistic about forging new academic alliances that could pave the way for co-operative ventures in the years to come.
Highlights and Contributions
Oral Presentations
1.Turnbull Rory: Phonological Network Properties of Non-words Influence Their Learnability
Rory’s study underscored the significance of a word’s phonological neighborhood in phonetic processing, extending this concept to non-words. By analyzing participant responses in an experimental setting, this study demonstrated that non-words with more “neighbors” and well-connected neighbors are learned with higher accuracy, indicating that existing lexicon can significantly influence the acquisition of new words.
2. Du Fengting: Rapid Speech Adaptation and Its Persistence Over Time by Non-Standard and Non-Native Listeners
Fengting’s study delved into the intriguing phenomena of how listeners adapt to accented speech, especially when the talkers share the same, similar or different language backgrounds. Through methodical research, the study revealed that both non-standard and non-native English listeners were more adept at perceiving and adapting to accents the same as their own, but not the similar or different one. Notably, this adaptation was not only immediate but also persisted over a 24-hour period, suggesting intriguing implications for language learning and communication.
This presentation offered an innovative perspective on how incremental cue training could aid in lexical tone learning for non-tonal language speakers. The findings suggest that employing exaggerated contrasts in pitch movements during the initial stages of training can significantly improve the learners’ ability to discern tonal differences, leading to comparable end-stage performance with conventional training methods.
Poster Presentations
1. Kelly Niamh: Interactions of Lexical Stress, Vowel Length, and Pharyngealisation in Palestinian Arabic
Niamh’s research filled a crucial gap in the understanding of lexical stress in Palestinian Arabic. By analyzing the acoustic correlates of lexical stress, she shed light on the nuanced interactions among stress, phonemic length, and pharyngealization, offering a comprehensive phonetic description of stress patterns in this Arabic variety.
2.Zhang Cong, Lai Catherine, Napoleão de Souza Ricardo, Turk Alice and Bogel Tina: Language Redundancy Effects on Fundamental Frequency (f0): A Preliminary Controlled Study
This study investigated the effects of language redundancy on fundamental frequency (f0), supporting the Smooth Signal Redundancy Hypothesis. Their controlled experiments revealed that language redundancy could indeed affect f0, potentially adding another layer to our understanding of prosodic structure.
This paper delved into the intricate relationship between voice onset time (VOT) and fundamental frequency (f0) in Jazani Arabic, proposing that f0 perturbation is predictable from VOT patterns. This lends further evidence to the theory of Laryngeal Realism and offers new insights into phonological representation.
This study questioned whether accent familiarity impacts narrative recall, particularly when listeners are exposed to different accents. Findings indicated a ‘familiarity benefit’ in Tyneside listeners, extending the impacts of accentual familiarity on language perception.
Our group’s robust contributions across diverse areas in phonetics and phonology were met with great interest and sparked important academic discussions, marking a significant footprint in the advancements of the field.
Reflection and Future Directions in Phonetics and Phonology
The congress served as an eye-opening experience that showcased the incredible diversity and depth of current research in phonetics and phonology. With hundreds of insightful studies, the event drew a roadmap for the future of these fields. While each study was a piece of a larger puzzle, the six keynote lectures stood out as beacons guiding the way forward.
Technological Advancements and Precision
Research in phonetics and phonology cannot stand alone; it requires the support of powerful tools to quantify sound features and visualize articulatory processes. Over the past few decades, these research tools have continually evolved, and numerous researchers have actively applied state-of-the-art technology in their studies. John Esling’s focus on the larynx as an articulator hinted at the importance of advanced imaging techniques, opening new avenues for understanding language development and linguistic diversity. Similarly, Paul Boersma’s discussion about the future of Praat emphasized how technology will revolutionize phonetic and phonological models. The convergence between these talks suggests a future where technology plays an increasingly central role in refining our analyses and providing greater computational power for simulating the nuances of speech.
Interdisciplinary Synergies
As the landscape of research in phonetics and phonology broadens, the call for multidisciplinary perspectives becomes increasingly urgent. Ravignani Andrea and Stuart-Smith Jane both underscored the importance of multidisciplinary approaches. While Andrea seeks to combine ethology, psychology, neuroscience, and behavioral ecology to explore the origins of vocal rhythmicity, Stuart-Smith envisions a future where sociophonetic and social-articulatory data deepen our understanding of speech patterns related to identity, social class, and dialect. The common thread here is the necessity for interdisciplinary collaboration to answer complex questions that cannot be addressed by any single field alone.
Embracing Social Responsibility
Perhaps the most poignant insights came from talks focusing on the social aspects of research. Titia Benders emphasized the crucial need for expanding child language acquisition research to lesser-studied languages, not only to understand their unique phonological elements but also to develop inclusive research methods. Pavel Trofimovich, on the other hand, urged for a socially responsible approach to second-language speech research, one that balances academic rigor with meaningful social impact. These talks collectively call for a future where research is not just theoretically robust but also socially responsible, reaching communities and languages that have been traditionally underrepresented.
Together, the keynotes painted a vibrant picture of a future that is technologically advanced, inherently interdisciplinary, and deeply rooted in social responsibility. It is clear that the next wave of research in phonetics and phonology will be as diverse and dynamic as the voices that make up human language itself.
Summary
The insights gained from this year’s congress serve as a valuable roadmap for the direction of phonetic and phonological research, areas that are central to the mission of our research group. From the pivotal role of technology in advancing research methodologies to the importance of interdisciplinary collaboration and social responsibility, the keynotes and studies presented offer a multifaceted view of the field’s future. As our group continues to explore new avenues of research, we are invigorated by the wealth of possibilities that these emerging trends present. They not only affirm the work we are currently undertaking but also challenge us to think about how we can contribute to these evolving dialogues in meaningful ways. Thank you for following along with our coverage of the congress, and stay tuned for upcoming research projects that will reflect these dynamic shifts in the field.
On Tuesday, we were delighted to welcome Dr. Byron Ahn for an in-depth workshop on the use of PoLaR in analyzing prosodic features of speech. The three-hour session delved deep into the intricate layers of intonation.
The workshop began by laying the groundwork. While segments in English (like consonants and vowels) shape the words we say, it’s the suprasegmentals that color how we say them. Prosody, thus, captures the nuances in tone, pitch, duration, and emphasis that breathe life into our words.
What sets PoLaR apart in the realm of prosodic analysis? Its rise in popularity stems from its decompositional and transparent labels, making it easy to grasp and apply. Unlike other systems such as TOBI, PoLaR labels concentrate solely on the foundational elements of prosodic structure, namely boundaries and prominences. This results in a richer phonetic detail about the pitch contour. Additionally, there’s no need for a language-specific phonological grammar with PoLaR, making it versatile and cross-linguistically applicable. Yet, it’s essential to note that PoLaR complements other labeling systems, like ToBI, rather than replacing them.
After providing the essential background introduction, Dr. Ahn guided us through the main tiers of PoLaR labelling, including the Prosodic Structure, Ranges Tier, Pitch Turning Points, and Scaled Levels. The session also touched upon Advanced labels, enabling a systematic tracking of a labeller’s theoretical analysis.
We’d like to express our deepest appreciation to Dr. Ahn for imparting his expertise and to all attendees for their active participation!
We are delighted to update our community on the successful completion of our recent workshop this Monday, titled “Training Your First ASR Model: An Introduction to ASR in Linguistic Research“.
Workshop Overview: The workshop was designed to delve deep into the foundational elements of Automatic Speech Recognition (ASR) and its classical architecture. Focusing on the application of ASR practices in linguistic research, participants were guided through a flexible workflow of automatic forced alignment, demonstrated using various research scenarios. The primary objective of this session was to help our attendees understand the core concepts of ASR and provide them with the necessary tools to utilize ASR in their linguistic research.
Speaker Spotlight: Our workshop was led byDr Chenzi Xu, a Postdoctoral Research Associate at the University of York. Dr. Xu’s current work revolves around the fascinating project “Person-specific Automatic Speaker Recognition.” Concurrently, she is concluding her doctorate at the University of Oxford. Dr. Xu’s remarkable achievements in the field have been recognized with the prestigious Leverhulme Early Career Fellowship, which she will commence at the University of Oxford next year.
We trust that our attendees found the workshop both informative and practical. We appreciate the active participation and look forward to the impact this knowledge will have on our individual linguistic research projects!
We are pleased to announce that the research of seven of our group members has been accepted for presentation at the International Congress of Phonetic Sciences (ICPhS) 2023 in Prague. This is a great achievement for our research team and reflects the hard work and dedication of our members.
Rory Turnbull’s research on: The phonological network properties of nonwords and their impact on learnability has been assigned to the 26th session on Phonetic Psycholinguistics. Fengting Du‘s research on: Rapid speech adaptation and its persistence over time by non-standard and non-native listeners has also been assigned to the 1st session on Speech Perception. Yanyu Li‘s research on: Incremental cue training and its impact on lexical tone learning by non-tonal listeners has also been assigned to the 1st session on Speech Perception.
In addition to these oral presentations, Niamh Kelly‘s research on the analysis of the effects of stress on phonetic realisation will be presented as a poster. Cong Zhang and her colleagues (Ricardo Napoleão de Souza, Tina Bögel, Catherine Lai, Alice Turk) will present their research on Language redundancy effects on f0: A preliminary controlled study as a poster presentation. Abdulrahman Dallak will present a study that he coauthored together with his supervisors (Jalal Al-Tamimi, Ghada Khattab) and their research is: Obstruent voicing and laryngeal feature in Arabic. Finally, the title of Andreas Krug‘s poster will be: The effects of accent familiarity on narrative recall in noise.
All the accepted research papers have received positive and constructive comments from the reviewers, demonstrating the high quality of research being conducted by our group. We are proud of our members’ achievements and their contributions to the field of phonetics and speech science.
We are excited to attend the ICPhS 2023 conference in August and look forward to meeting other researchers who share our interests and passion for this field. We wish all our group members the best of luck in their presentations and are confident that they will represent our research team with distinction.
Being interested in sound systems from the perspective of both production and perception, Niamh Kelly ran a project examining the production of /l/ sounds by bilingual speakers of English and Spanish from the El Paso region, to investigate the effects of language dominance on velarisation patterns. She also ran a pilot study where she looked at the production of /z/ sound of a bilingual across time. Here, we had her give us a presentation of the outputs of her work.
Part 1: A bilingual community on the US-Mexico border: what are they doing with their [l]?
Background information:
Transfer can occur in the productions of multilinguals, where one language influences the other and such effects can go in either direction between the L1 and L2. Sometimes, speakers are found to have productions that are intermediate between the two languages. In some regions, the whole community is bilingual, making it convenient to look at language transfer effects.
Although similar to each other, the /l/ sounds in American English (AmE) and Spanish are not exactly matched up. While /l/ sound in Spanish is realised as fronted (light/clear) /l/, in AmE it is more velarised overall, especially when it is in codas.
The participants in this research lived in a city (El Paso) on the US-Mexico border, which is a bilingual community.
Research questions:
This research asks:
To what extent do balanced bilingual speakers show transfer effects in laterals? That is, are there positional effects in just English or in both English and Spanish?
What effects of vowel height and front/backness have on velarisation in laterals in both languages?
Hypotheses:
Since these speakers are balanced bilinguals, they could be expected to keep their languages separate: English /l/ would be more velarised overall than Spanish /l/, and that in English, coda /l/ would be more velarised than onset /l/, while in Spanish no such difference would occur.
Results:
From the analysis of the participants’ production in both English and Spanish the following results emerged:
English and Spanish were significantly different in both positions. /l/ was more velarised in English than Spanish in both onset and coda position.
/l/ was more velarised in codas than onsets in English while no such positional difference emerged in Spanish.
Next steps:
Research like this and more further research can help in adding to the description of non-mainstream varieties of English and varieties used by multilinguals.
Part 2: A bilingual across time: what happened to his /z/? Acquisition of voicing in English /z/ by an L1 Norwegian speaker in a 25-year period.
Background information:
The English /s/ – /z/ contrast has been found to be difficult to acquire for L2 English speakers who do not have this contrast in their L1. Norwegian-accented English has a lack of voicing in /z/ since Norwegian does not have /z/.
Current study:
This study is a longitudinal study of the L2 English of L1 Norwegian speaker Ole Gunnar Solskjær. Ten interviews from two time periods, 1996-8 and 2021 were examined, focusing on his English productions of /s/ and /z/ and how production patterns change over time. There variables were coded: position in word (medial or final), preceding segment (voiced or voiceless), and morphemic status (morphemic, e.g., ‘goals’ vs stem, e.g., ‘please’).
Results:
In the Early timeframe, 100% of /s/ tokens were voiceless and 93% of /z/ tokens were voiceless. In the Late timeframe, 98.5% of /s/ tokens were voiceless (no significant effect of timeframe) and 46% of /z/ tokens were voiceless (a significant effect of timeframe).
Duration was longer when voiceless (supporting the auditory categorisation) but not affected by position in word.
No difference based on morphemic status.
Discussion and next steps:
More exposure to and practise with the L2 has led to an increase in L2-like voicing productions. OGS is acquiring a new voicing contrast, but has not acquired it completely as only about half of /z/ tokens were voiced. More work can be done to look at other fricatives and also the intermediate time frame.
General conclusion:
Here we find transfer of L1 phonetic and phonological patterns to L2 at the individual level, which can continue even after years of exposure and use.
It also occurs on a larger scale when a community is bilingual.
It is important for linguists to describe non-mainstream varieties.
Rory Turnbull gave us a talk on his research on what influences the phonological structure of the words in a language.
The train of thoughts:
The talk started with narrowing down the research questions: from the big question of why languages are the way they are to specific questions of what influences language structure and makes languages have the words that they have and not other words. His current research is on: what influences the phonological structure of the words in a language? While the typical answer to the question is phonotactics, his response is that ‘some’ functional pressures may also affect the phonological structure.
Prior work:
Rory’s prior work suggests that natural languages have unexpectedly smooth phonological networks where each word is a nod and a link exists between two words if they are phonological neighbours (only differ by the deletion, insertion, or substitution of a single phoneme). It means that some words are alone in the network while some have loads of neighbours. Based on previous findings, he proposed: these ‘extreme’ words (unusually clumpy or sparse in the lexicon) are harder to learn and harder to retain than non-extreme words.
Pilot planning:
A pilot is in the plan which aims to test the nonwords learning of native British English participants. At the end of this session, our group provided feedback on the experimental design and recommendations on literature and related topics.
Cong Zhang gave us a talk on boundary tones in Mandarin.
Background information:
Intonational tunes are generally made up of pitch accents and edge/boundary tones. The intonation in a tonal language is difficult to phonologise as F0 can be modulated by the effects of lexical tone, lexical prominence (stress), tone sandhi, sentence prominence (focus) and sentence type (function), etc. The overarching research question is: Is there intonation in Chinese? If there is, then how to differentiate the lexical-level tone and sentence-level intonation?
The tone systems of Tianjin Mandarin:
Tianjin Mandarin is a dialect spoken by residents in the Tianjin municipality, which is near Beijing.
The data was transcripted based on ToBI, a prosodic transcription system based on AM.
Floating boundary tone: a type of boundary tone that modifies the original lexical tone contour and it deters the falling tones from falling and facilitates the rising tone with rising. Such a tone has no phonetic realisation itself, but has a phonological effect in triggering a higher last tone.
Mean F0 range was used to compare the prosodic patterns of questions (Q) and statements (S) in Tianjin Mandarin.
The main results found: a. Q has a higher register (mean pitch) than S. b. The falling tones (L, HL) have a smaller pitch range in Q than in S, while the rising tones (H, LH) have a larger pitch range in Q than in S.
Functional Principal Component (fPCA) Analysis:
fPCA is used to analyse the boundary between onsets and rhymes in the study
Complex boundary tones can be used in Tianjin Mandarin for diverse communication purposes, including extra emphasis, objection, correction, sajiao (being cute), showing off and sarcasm.
Summary: Tianjin Mandarin has bi-tonal boundary tones.
The issues to be looked at next:
What are the functions? Are some of these functions paralinguistic?
How to elicit data in a more controlled environment?
Do bi-tonal boundary tones only exist in northern Mandarin varieties and speakers?
Today, we have Hiromasa Kotera to present the preliminary work of his PhD study, which is about the retainability and sustainability of phonetic discrimination ability of infants.
Rationale:
The perceptual attunement effect has been hotly discussed in linguistics and child development. It refers to a change in infants’ discrimination ability of language sounds. At first, infants can discriminate all sounds in human languages. However, being exposed much to their native language, infants start not to be able to tell certain sounds apart if they do not exist in their L1. Previous studies have shown that short-term exposure to certain languages with the target sounds can allow infants to regain the ability of phonetic discrimination. German does not have the vowel /æ/, and the sound /ɛ/ in German differs from English.
Research question:
This study is to see if the finding can be generalized for long-term effect by testing German infants’ phonetic discrimination ability of English vowels /æ/ and /ɛ/.
Preliminary research:
Three groups of infants aged 5-6 (n=40), 7-8 (30) and 12-13 (30) months are exposed to American English carrier words of the two test sounds: MAF /mæf/ and MEF /mɛf/. A visual habituation paradigm is employed involving three phases: habituation phase, test phase, and control phase. Infants’ looking time to a checkerboard is measured while listening to MAF or MEF.
Findings:
For infants at 5-6 months old, there is a significant difference existing among conditions. In habituation, infants’ looking time when they heard MEF was longer than when they heard MAF.
As for the other two groups of infants, no significant difference was detected among conditions. This supports the perceptual attunement effect.
Feedback from the group:
It may be worthwhile to consider how well the test sounds recordings represent that category.
It is also worth thinking about the difference between phonetic discrimination and perception.