What do kids know about language that doesn’t come from what they hear or see?

This may seem a very strange question. Surely children can only know about language that which they hear, see or experience from the language users around them? After all, a child only learns the language(s) they are exposed to. No child in the UK, for example, grows up learning Auslan, unless they are exposed to it by an Auslan-signing parent.

But this question is a (less provocative!) rendering of one of the biggest open research questions in child language, and one which has split child language researchers, sometimes acrimoniously, since the mid 20th century.

You may be more familiar with this question framed as the “nature vs. nurture” debate, or the innateness debate. Both of these framings I find very weird. When you drill down, all child language researchers believe that both nature and nurture play roles in the acquisition of language by children. The real debates are around the nature of nature and the nature of the nurture. So let’s dive into exploring the two main approaches to these issues, starting with the group that proposes language-specific innate knowledge.

(Full disclosure first: I was trained and now work in the generativist tradition, I know it better, and I think it is more often misunderstood than the constructivist tradition, so I spend quite a bit more time on what follows describing the generativist point of view. Even so, I hope you find it interesting, and somewhat balanced in terms of critique.)

The Domain-Specific approach (a.k.a. the Generativists)

These are the scholars typically positioned as the “nativists”. They’re responsible for creating All the Acronyms (LAD, UG, PLD MMM*…) and are often accused of not taking the other side seriously, mostly because they don’t engage with their work very often (which is, of course, not good scholarship).

But generativists are very serious about their main claim: that human beings are born with some knowledge about language that they use to make sense of the bewilderingly complex language input they receive. This is known as domain-specific knowledge because, quite simply, it’s knowledge that pertains only to the domain of language, and no other cognitive domain. You may have also heard of it referred to as Universal Grammar – some language knowledge, even rules, that all children can use from birth.

Why make such a claim? Well, in part it is because it’s striking that children learning different languages actually pass through very similar developmental stages – adults do the same too when learning artificial languages, irrespective of their first language. This suggests that we don’t just have general learning biases or preferences, but that some biases exist that apply specifically to linguistic information.

The birth of a language: Nicaraguan Sign Language

One of the most compelling examples of this can be seen in the development of Nicaraguan Sign Language (NSL). NSL grew out of the opening of the first state-run school for deaf people in Nicaragua in the 1970s. Students came from all over the country and did not share a language, using instead a range of “home signs”. But as they interacted in the playground, conventions of use developed among the students – a lexicon with some rules. These rules, however, didn’t really resemble other languages. They did the job, and the children communicated and flourished. But a lot of the signs and structures relied heavily on representing events and things in the world fairly literally, and didn’t make use of many abstract markers in the way that full natural languages typically do (things like tense, aspect, person agreement, etc.).

When new, younger students arrived, they took this fledgling system and (entirely subconsciously) generated a true language grammar for it. The younger students used their innate compulsion to build a linguistic system, using highly arbitrary features like space and handshapes to create a complex system that could express more efficiently complex, abstract idea. This system – NSL – has a grammar that is fully recognisable as a natural language grammar. That’s to say that it is hierarchically structured, it doesn’t “count” (there isn’t a specific type of word that must be the third word in the sentence, for example) and its rules are regular in many of the same ways as any other language.

This brings us to another motivation for the generativist approach to understanding language acquisition: for all the logically possible ways in which languages *could* differ, they tend to structure their constituent parts along very similar lines.

What about the Poverty of…

Note that I haven’t mentioned (directly) a theory often cited as a motivation for a generativist approach: the Poverty of the Stimulus. It’s an idea that grew out of “Plato’s Problem” – the question of how humans extract so much information from limited evidence in their environment. With respect to language specifically, it is used to suggest that children simply don’t get enough linguistic evidence to be able to build a language grammar as quickly and as accurately as they do.

And for sure, it’s amazing that by the age of three, small children can engage in long narratives, ask questions, follow complex instructions and reason about the world, even though they can’t tie their shoelaces and some may still be in nappies. However, linguists who made (and make) this kind of argument are often only interested in syntactic strings, which don’t take into account all the *other* cues and pieces of information that actually accompany those strings when spoken or signed in context – intonation, gesture, context, growing world knowledge. Moreover, if we don’t have theories about children’s ability to reason, remember and recognise patterns across the information they receive, then the Poverty of the Stimulus is pretty vague. How much information is enough, and what information are children actually capable of using?

Speaking of vagueness, I will admit, as a card-carrying generativist, that claims about the content of possible domain-specific knowledge remain bafflingly vague. So let’s examine a couple of the more worked-out versions of Universal Grammar in more detail. Note that I’m skipping over quite a lot of historical detail here to focus in on just two approaches to understanding domain-specific knowledge – if you want to know more, come and do a degree with us in linguistics ;).

Rich UG: Principles and Parameters

In the past, some scholars assumed that Universal Grammar consisted of a very rich store of innate knowledge made up of universal linguistic principles and parameters – possible “settings” for those universals in different languages.

A famous example is the pro-drop parameter. Some languages allow users to “drop” the subject of the sentence (that is, not say it out loud). These languages include Italian, Spanish, American Sign Language, Mandarin Chinese, and many more – around two-thirds of the world’s languages, as far as we know. The exact rules that determine when you can drop a subject might differ, but the fact that you can drop a subject is what links them. Take examples (1-2) from Italian, and compare them with the English translations.

  1. (Io) sono felice.
    “I am happy.”
  2. Piove.
    “It rains.”

In (1), the Italian first person pronominal subject “Io” is optional – “Sono felice” is itself a perfectly grammatical sentence in Italian. But in English, “I” is obligatory – “Am happy” would be considered ungrammatical most of the time. In sentences like (2), there is no subject in Italian, because weather verbs like piovere (‘to rain’) just don’t need subjects at all. But in English, we insert the expletive or pleonastic pronoun “it” to get “it rains”, because the grammar’s need for a subject is so strong that it doesn’t matter whether it actually refers to anything or not. It’s a purely grammatical requirement. Other languages that require the subject to be overt, like English, include French, Indonesian and Mupun, to name a few.**

How does this relate to acquisition? According to a theory that assumes principles and parameters, children will be (subconsciously!) listening out for how their language expresses subjects because they’ll ‘know’ it’s an important point of variation for grammar with a roughly binary set of values – either you can drop your subjects, or you can’t.

A theory like this is useful because it helps make predictions about what children might or might not be doing during language acquisition, which you can then test by looking at language corpora (sets of recordings of children) or by running experiments. Taking pro-drop as an example again, it is a common observation that English children drop subjects in their early speech, until around age 3. A theory like principles and parameters provokes an interesting research question – do English-acquiring children drop subjects because that’s grammatical for them to start with, because pro-drop is a default setting (a language-internal explanation), or for some other reason, like reduced processing ability (a language-external explanation)?

Linguists like Nina Hyams and Virginia Valian (among many others) have looked extensively at English children’s subject drop through this lens. They noticed that English-acquiring children’s subject drop is not like the subject drop of children learning Italian. Others also noticed that English children’s subject drop is not like the subject drop of children learning Mandarin (in which the rules on subject drop are a bit different from Italian ones). Mandarin also permits the dropping of objects, so Mandarin-acquiring children oblige, but English-acquiring children barely drop objects at all. Both Italian- and Mandarin-acquiring children learn adult-like pro-drop settings very early, while English-acquiring children are doing something quite different from both English-speaking adults and children learning pro-drop languages. English-acquiring children tend only to drop subjects when they’re the first thing in the sentence – that is, they don’t tend to drop subjects in more complex utterances like wh-questions or subordinate clauses, where other bits of linguistic structure precede the subject.

So what are English-acquiring children doing? They seem to be just missing out the very edge of their utterance, or in full linguistic parlance, the top-most part of the structure (given that sentences actually have hierarchical, not linear structures. More on that another time). Why might they do this? Linguist Luigi Rizzi made an influential proposal in the early 1990s, refined in the 2020s with colleagues Naama Friedmann and Adriana Belletti, that children learn sentence structure from the ‘bottom up’, with layers of information emerging in sequence, which may be “truncated” if their processing capacities become overwhelmed. According to Friedmann, Rizzi and Belletti’s “growing trees” hypothesis, children learn the structure of a sentence’s verb phrase (with any objects) first, then grow into marking tense and aspect, then into marking the subject, followed by things that mark how the sentence fits into a discourse, for example the wh-word in a question.*** The claim goes that at the point at which English-acquiring children are dropping subjects, they know that they are needed in English, but they’re still working out exactly what they look like and how they relate to the rest of the structure. All this work sometimes leads to truncation of the sentence – or in tree-surgeon parlance, the subject gets lopped off.

Though Principles and Parameters provided interesting ways to approach language acquisition (and language description more generally), it’s long been clear that it’s not a psychologically valid proposal for the domain-specific knowledge children might be born with – it is quite improbable that children’s innate tools for making sense of language would take such a highly specified form.

Parsimonious UG: Merge and Agree

A newer approach to domain-specific knowledge, deriving from an article by Noam Chomsky in 2005, proposes two processes (rather than a number of principles) that drive how children unpick their language input. Those processes are Merge and Agree.

Merge is, briefly, the process of taking two things (words, phrases, clauses…) and combining them to create a new linguistic object. Merge builds hierarchical structures, and we already have plenty of evidence that language is hierarchical, not linear. Grammatical rules reflect this, processing reflects this, and acquisition respects hierarchical structure.

An ‘syntax tree’ capturing hierarchical relationships in the statement “The boy who was tired watched cartoons.” For cool work on how even young children know that the relative clause is contained within the subject, take a read of Crain and Nakayama (1987)

Combining objects means that some relationship is formed between those objects – that they come to share or express some feature that we can use as language users to make meaning. So the language user needs to be able to recognise what features drive the combination of specific objects, as this will help them parse structure accurately (understand) and build new structures (produce language). The process of forming relationships between language objects is called Agree. To apply Agree, the language user also needs to form a set of features that are relevant in their language that can be recognised in the language objects, and which they could use again and again to parse any of the infinitely many sentences they might be exposed to in their input.

Let’s (try to!) get a bit less abstract now and return to the issue of pro-drop, with help from linguist Theresa Biberauer, who looked at pro-drop from the point of view of Merge and Agree in 2018. She argued that children acquiring their first language will be constantly looking for patterns across their language and looking for features – some abstract, some maybe less so – that link the structures in their input. As English-acquiring children very rarely hear sentences without subjects, and receive evidence for subjects that carry no independent meaning (expletive “it”), she argues that they won’t look for features that explain subject patterns because, quite simply, there’s no need to. There’s no ‘pattern’ to explain in that sense – you just need to have a subject, and English-acquiring children will pick this up early on, leaving their subject-dropping in production to be explained by other factors such as processing, or still needing to build the parts of the grammar that will connect up the subject with the main verb, such as auxiliary verbs.

The input to Italian- and Mandarin-acquiring children, however, leads them to pose a question – under what circumstances should I drop subjects? They then need to work out what features correlate with dropped subjects. For Mandarin, children will work out that what really matters are discourse features – subjects (and objects) can be dropped when they are topics that are somehow familiar in the conversation already. For Italian, children must realise that discourse matters – familiar subjects can be dropped – but grammatical agreement matters too. Subjects can be dropped because the verb is marked according to the person and number of the subject in Italian – in example (1) above, the verb form sono is the first person singular form of essere (‘to be’). But verbs aren’t marked for the person and number of the object, so objects can’t be dropped.

Noticing these features isn’t just useful for pro-drop, however! Being able to recognise topichood, subjecthood, person and number is also useful in both Mandarin and Italian for other aspects of grammar, like word order, and also for learning how intonation works. So having an innate compulsion to look for relationships between language objects really will speed acquisition along, because the features that characterise those relationships will pop up again and again in the broader language system.

Just because it ain’t language specific, doesn’t mean it ain’t innate…

You’ll note that, in the case of “parsimonious” UG particularly, there are implicit references to cognitive processes that aren’t domain-specific, like pattern-spotting, pattern-matching, awareness of frequency, and more. This is because generativists also make room in their theories for domain-general learning mechanisms – in other words, cognitive processes that aren’t specialised for language but that are used to learn language. We’ll talk more about these below, as they are the key tools in the theoretical arsenal of “the other side”, but they are crucial for the application of Merge and Agree, and are included in generativist theories as “third factors” – factors that aren’t domain-specific knowledge or the input itself.

This label gives you a bit of a clue – these ‘factors’ aren’t the things that generativists are, generally, most interested in. But lack of interest doesn’t mean they’re ignored altogether. Indeed, some generativists take the interaction between general cognitive and language-specific processes very seriously. Linguists like Jeff Lidz and Annie Gagliardi build domain-general processes into their models of language acquisition in just the way described above – noting that general cognitive processes must apply to language input before language-specific processes, to provide filtered input that children could plausibly assess for where Merge applies, and what Agree processes are at work. They map their model in the diagram below, taken from their 2015 paper:

Figure 1 from Lidz and Gagliardi (2015). The processes in the yellow block are basically domain-general processes. ‘Universal Grammar’ would, according to Chomsky, Biberauer etc. refer to Merge, Agree and a template for linguistic features.

Theresa Biberauer also takes seriously the idea of “maximising minimal means”, that is, using and reusing features and observations you’ve already made. This is a domain-general process that would be useful for making sense of lots of different types of input. But with respect to language acquisition, she argues that it is a crucial third-factor used by children as they apply Merge and Agree to their input, and that it also drives some of the restrictions on variation that we see in natural language.

You’ve mentioned input a lot. I thought generativists weren’t too bothered about it…?

I think the most misunderstood aspect of the generativist approach to language acquisition is generativists’ stance on input. It is true that some generativists will take the “target” for acquisition to be the grammatical rules of the relevant language at a general, population level (i.e. what is generally considered to be grammatical in that language). Generativists are also often interested in comparing acquisition trajectories across languages so, again, they generalise across individuals and consider acquisition at a population level. This is in contrast to constructivists, who typically take into account only the kind of language that the individual child experiences – that is, the language used by their families and caregivers.

It’s pretty obvious how the two ‘targets’ here might differ, and how that would affect predictions that you would make. But input is always important, and generativists don’t assume or predict otherwise.

Given that we’re thinking about input, let’s turn now to our other set of players, for whom input drives not only acquisition, but much of their theorising and reasoning.

The Domain-General crew (a.k.a, amongst other things, usage-based researchers or Constructivists)

It’s a little harder to define a “main” claim of researchers for whom domain-general knowledge is the child’s only tool for making sense of their language input. That’s because there are a lot of different types of domain-general process to investigate – working memory, long-term memory, statistical generalisation, sensitivity to frequency, and so on. But something that is key for constructivists like Ben Ambridge and Elena Lieven, which differentiates them from generativists, is that children go from specific observations about their individual language input – e.g. storing particular phrases that they have heard/seen – to gradually generalising abstract rules. Memory, therefore, plays a major role in a constructivist account, whereas even the “third-factor”-interested generativists rarely, if ever, mention it.

What children can plausibly add to memory, of course, depends on what *exactly* they are exposed to. Sensitivity to frequency is a core theme in constructivist theories – children will appear more adult-like when they can use a stored phrase or string from memory, and less adult-like when the task they’re performing can’t be addressed using stored material. That frequency is important in language development is well established – it plays a role in language change in many levels of language, as well as language use, language processing, and patterns that emerge across languages (recall, this was an argument for domain-specific knowledge too…)

Given their focus on individualised input, constructivists also insist on the importance of understanding language acquisition as happening in interaction. As such, studies of the acquisition of pragmatics are dominated by constructivists, with Catherine Snow and Thea Cameron-Faulkner leading the charge****.

Right. So…null subjects from a usage-based POV?

Sorry lads. That’s what I wanted to write about here, but I cannot find an account of English-acquiring children’s null subjects from the constructivist side of things. I did find an account for Spanish, a pro-drop language like Italian. Hannah Forsythe and colleagues argued that the frequency in input of different types of pronoun affects when they begin to be dropped in adult-like ways – first and second person subjects are more frequent than third person subjects, and children achieve adult-like rates of dropped and overt subjects in first/second person cases first, before abstracting the same rules to third person subjects.

As for English null subjects, my guess, but this is only a guess, is that there are three routes to take to explain them using a constructivist lens.

One is to put all the blame on processing limitations of various types – which is of course, ultimately where we landed in the generativist approach too. Reading around a bit, and talking to colleagues of a constructivist persuasion, this seems like a bit of a cop out.

My excellent colleague Dr Nick Riches suggested a second take: subjects don’t add much grammatical information to an utterance – especially when first and second person subjects are so much more common in child speech than third person ones. This is in contrast with objects, which are important for indicating the argument structure of verbs (and, of course, the less obvious identity of the object). So if children are going to drop any information (because of gradual acquisition, or processing pressures), the subject is the least informationally dense element.

The third route is to assume (as constructivists do) that the child’s grammar is very different from an adults, and then go looking at the rate of utterances that have no subjects in the input *regardless* of the morphological form of the verb. Key here are imperative verbs – these are very frequent in child input and *don’t* require a subject but, in English, are not overtly marked as imperative. They look superficially exactly the same as most verbs in the present tense, and infinitives too. I guess a constructivist could argue that imperative verbs could provide sufficient evidence for certain verbs not needing a subject, at least in very early child language. To argue this you’d have to have a story about how children ignore the specific pragmatic use of imperatives, but maybe that’s doable.

It’s also important to note that usage-based approaches tend to focus, perfectly legitimately, on children’s production in terms of evidence for their grammar. This is of interest to generativists, of course, but generativists are also interested in children’s comprehension. So while both approaches will look at production in their accounts, generativists are also likely to probe children’s grammar by looking experimentally at how they react to grammatical and ungrammatical sentences.*****

What all this illustrates, though, is how the “two sides” here are not two sides of the same coin. At best, we’re looking at the head of a pound coin and the tails of a 50p. The two approaches work within the same field, but often look to answer quite different questions – and they’ll make use of different data to support their own arguments. As it turns out, those data don’t necessarily overlap all that much, *even* in a language as heavily studied as English.

I’m doing my A-levels. How can I make sense of all of this for the exam?

I have a few ideas here, but this is a topic I’m probably going to return to in more depth in a future blog post too. Also, bear in mind that I’m pretty well acquainted with the specifications for some of the large exam boards (particularly AQA and OCR), but I’m still not as steeped in them as trained teachers are! But if you’re interested to know my initial thoughts, then please continue…

Key to the exam first and foremost (and to linguistics in general) is the ability to accurately *describe* children’s production, so above all else, ensure that you are confident in labelling parts of speech and morphemes and that you can spot and describe features like person and number, both for the adult target and the child’s production. Also, don’t focus just on children’s non-adult-like production – give them ‘credit’ for what they *can* do in an adult-like way too, whether that’s producing subjects, using relevant auxiliary verbs, and so on.

Given the direct link that constructivists make between the child’s individualised input and their production, it’s a little easier to see how you could zoom out from transcripts of child language to talk about constructivist approaches to language acquisition. Even then, the role of frequency is hard to address from fragments of transcripts, so to be true to constructivist techniques, you’d have to talk about what you would look for if you had more examples of dialogue from the same child-caregiver pair.

With respect to linking from transcripts to generativist claims, the “growing trees” hypothesis is interesting to link to in terms of evaluating why verb phrase syntax is adult-like earlier than tense and subjects. You can also think about what grammatical features a child needs to work out for different linguistic items that are in the transcript and what interim features they might propose – features are a good way of thinking about and comparing linguistic complexity. As an example, determiners like “the” and “a” differ in terms of definiteness, but “a” is even more complex because it can be used both with specific and non-specific nouns, and is also specified for number (singular only). So there are three features that a child has to learn when trying to pick a determiner, and the determiner options differ in their featural complexity. Imagine what effect that could have on the acquisition of determiners in English.

Notes

*Language Acquisition Device (see ******), Universal Grammar (described in text), Primary Linguistic Data (=input!), Maximising Minimal Means (described in text)…and many more!

**If you’d like to know more about how languages express subjects, take a look at the World Atlas of Language Structures. You can also find lots more information about the structures of hundreds of languages all over the world on this fantastic website.

***There’s a competing hypothesis under development by Martina Wiltschko and my very good colleague Johannes Heim, called the “Inward Growing Spine Hypothesis“, which suggests that structure grows inward from the bottom (the verb phrase) and the very top (discourse-related phrases that call on the addressee to respond, like vocatives). I’m not sure either approach is *quite* right, but Wiltschko and Heim’s idea that some of the very topmost structure does come in early is clear from early child utterances. The question is – which parts of the topmost structure? I’m working on my own contribution to this question – watch this space 😉

****Generativists, it should be noted, have spilled a lot of digital ink on one particular aspect of pragmatic acquisition, namely the topic of scalar implicatures. But that’s a topic for another day.

*****Obviously, you can’t just ask a small child whether they find an utterance grammatical or not – in fact you can’t do this until they’re about 6 or 7 years old (and even then, we’d use more child-friendly wording!). But you can interpret children’s behaviour in response to stimuli in lots of clever ways to probe what they do and don’t know about the adult grammar.

******LAD – Language Acquisition Device. A term from Noam Chomsky’s early work to represent an (unspecified) process or set of processes and knowledge that the child is born with and that specifically concern and work over language input. Not in use by generative linguists today.

And a postscript…

I may have become a bit delirious when initially planning this post (I was at a soft play centre too early on a Sunday morning. It does strange things to you) and I ended up casting the two competing approaches as the Capulets and Montagues in a terrible rewriting of the prologue from Romeo and Juliet. Trigger warning: what follows might constitute cultural vandalism…

Two (main) factions, both alike in indignancy
In Academia, where we set our scene
From old grudges break to new strawmen
Where wilful snubs make wilful claims quite mean.
From forth the messy output of these groups
This messy cross-referenced blog takes shape;
Whose well-intentioned overview does aim
To make clear where their paths each other trace.
Generativists from LAD****** through UG
Move, while constructivist claims also persist.
I cannot claim to end the feud – not me! –
Just reveal of what the arguments consist.
The which, if you with patient thumbs scroll through
What here I’ve missed, I toil now to construe.

The tip of the tongue and the lips and the teeth (or, phonetics for kids)

Strange things are happening in the world of education. To the benefit of absolutely no-one, STEM subjects are frequently held up as the only ones worth studying, while technological advances (and misadventures) demonstrate the need, time and time again, for a nuanced understanding of human beings, how they think, and how they move through the world. In short, we all need more of the arts and humanities.

As a discipline that straddles both STEM and arts and humanities, we have a unique opportunity in linguistics to disrupt the current binary discourse (STEM or bust) and muddy the waters wherever possible. Linguists frequently apply the scientific method to that most human of behaviours, language, but just recently, my fabulous colleague Dr Emma Nguyen and I had the chance to squeeze some awareness of being human into a science space.

The scene: STEAM week at a very lovely local primary school. The brief: chatting to year 3 (ages 8-9) about muscles. The topic? Well, as linguists, we were duty bound to present the group of muscles that gave our field its name* – the tongue. The tongue is crucial for spoken language, but what did our 8 year olds already know about how it shapes language sounds?

By year 3, children in English schools have had a few years of instruction in phonics – a system that teaches the most common associations between sounds (phones) and letters (graphs) to help children ‘decode’ written words. The principle is that by sounding out a word’s constituent graphs and blending the corresponding phones, children can break into written English without needing to rely (too heavily) on memorising full word forms.

However, the phonics schemes we’ve come across as parents or in conversation with schools spend very little time on explicitly describing how, where and with what body parts sounds are made. Some schemes even include mascots that make speech sounds despite having no lips or teeth. Teachers say that the sounds made at the very front of the mouth like /p,b,m/ might elicit some discussion about the lips, and sounds like /f,v/ are talked about in terms of lips and teeth. But sounds at the back of the mouth like /k,g,ŋ/? No-one is sure how to talk about how these sounds come into being.

So the tongue and the back of the mouth are as foreign to our year 3s as the Moon. But unlike the Moon, we don’t need a rocket to find out more about them. Instead, we used a little trick sometimes employed in first year university phonetics classes (that would also buy us a bit of notoriety – more on that later).

That’s how the tongue makes sounds? Sweet!

We ran a little experiment, turning Willy Wonka by handing out lollipops** (or lollipop sticks to anyone who didn’t fancy a sweet). We instructed the children to put the lollipop toward the back of their tongue, then we all read together words with different vowel sounds in them.

This helped demonstrate how open front vowels like /a/, where the lolly (and tongue!) move down and forward in the mouth, differ from closed back vowels like /i/, where the highest point of the tongue is right close to the back part of the hard palate (resulting in a few lollipops stuck to the roofs of mouths).

The kids were amazed at how mobile their tongues were, while we were amazed at how disciplined they were at not chomping the lollipops at the first opportunity. NB: the sweets were not a gratuitous choice! They stick beautifully to your tongue to allow you to focus on how your tongue moves without actually inhibiting it from doing so normally. Until it gets stuck to your palate after a close sound, of course, which will happen if you’re (a) eight years old or (b) narrow-of-mouth, like me.

The experiment demonstrated how place of articulation, or the point at which the tongue, teeth, or lips close or come close together, is a really useful way of distinguishing between sounds, and helps us differentiate words like tap and cap. Tap’s initial /t/ sound is made by the tip of the tongue closing up against the alveolar ridge, just behind the teeth. In contrast, cap’s initial /k/ is made by pushing the back of the tongue right up against the velum, or soft palate. It is the position of the tongue alone*** that prevents misunderstandings about wearing bathroom furniture as a hat.

We also showed the kids some MRI videos from the fabulous website Seeing Speech. These give a fantastic insight into the flexibility and precision of the tongue from a side-on view (pretty hard to achieve any other way), with the added gross-out value of demonstrating just how enormous your tongue actually is.

What other muscles affect how sounds are made?

Some children noticed, though, that place of articulation is not the only thing that differentiates sounds. This led to discussion of another set of muscles and tissues that help us produce speech – the vocal folds. We placed our hands on our throats and made the sounds /s/ and /z/ on repeat. The buzzy feeling of /z/ is due to vibrations in the vocal folds as air is forced between them, and this is called voicing.

Experiencing voicing first hand
Photo credit: Hotspur Primary School

To get a sense of what voicing really looks like, we played another video, this time of vocal folds in action from above, and then we got creative. Using a paper cup and elastic bands, we created our own voice boxes, using straws to blow across the elastic-band-vocal folds to cause them to vibrate, and even stretching them with our fingers to create higher pitched sounds. This activity was inspired by one described on howtosmile.org.

One of the year 3 form teachers is also the school’s music teacher, so we compared out paper cup larynxes to his guitar collection, and I also brought a kazoo along to demonstrate how voicing changes airwaves. But it was the *other* form who gave us an amazing demo of one of their favourite assembly songs to demonstrate pitch changes – a really glorious moment.

Kazoos are a great way of demonstrating how voicing affects the sound waves that we can produce
Photo credit: Hotspur Primary School

Did it all work well?

We thought so! The children took part with enthusiasm, asking great questions about how the vocal cords seemed to change shape and about how speech sounds differ across languages. We taught the sessions on two different days, so we knew we’d had an impact on the first group, as the second already knew some of what we were going to talk about (plus they were well primed for the sweets). We will happily take playground chat as evidence of successful learning!

As for the teachers, they thought that increasing awareness of tongue movements could really help some of the children with language delay. Of course, speech and language therapists employ plenty of exercises that do just this, but we reckon there’s a benefit to be had by sharing all this information with children earlier and more generally, embedding it into phonics learning.

Want to know a bit more?

As you might imagine, we couldn’t do much more than that in a 45-minute session, and we didn’t even get onto manner of articulation****, a third factor in classifying sounds. We didn’t really demonstrate the International Phonetic Alphabet as a method for representing speech sounds, and we’d have loved to hand round kazoos to all (the teachers may have been less pleased with this). If you’re interested in thinking a bit more about place, manner and voicing and how we use these to describe sounds, here are some resources that might be useful (and spoiler: we’re in the process of trying to develop some that will complement traditional phonics programmes too!).

Notes

*From lingua, Latin for ‘tongue’, came the 16th century coinage ‘linguist’, meaning “master of languages” or “one who uses their tongue freely”. ‘Linguistic’, meaning “pertaining to language(s)” was an early 19th century innovation from German linguistisch, fabulously described by the OED as “hardly justifiable etymologically [… but] has arisen because lingual suggests irrelevant associations.” The term ‘linguistics’ for the field of study followed in the mid 19th century, but only really took off (and overtook its main competitor, ‘philology’) in the late 1940s. All info in this note, apart from the Google Ngram, is taken from the wonderful resource etymonline.com: linguist, linguistic, linguistics.

**Maoam Joystixx were far and away the best sweet for this experiment, with Swizzles Matlow Drumsticks as a decent runner-up as a vegetarian/halal option. No, this isn’t an #Ad. Yes, there was an extensive testing period one afternoon ;).

***Yes, OK, context helps too. But sometimes that is absent, and sometimes terrible analogies happen too.

****This is to do with how the articulators (lips, tongue, teeth, etc.) relate to each other, mostly in terms of how close they are. For example, the initial sounds in both tap and sap are unvoiced and made by placing the tip of the tongue at the alveolar ridge, but where there is complete closure of the tongue against the alveolar ridge for /t/, it’s just close to the ridge for /s/, allowing air to hiss through that small gap.

What do kids (and linguists) know about vowels (and poetry)?

This blog post is part reminisce, part information about children’s earliest language experience, and part reflection on an event I ran in November 2023 as part of the Being Human festival. If any of this sounds interesting to you, please read on…

My eldest’s first poem was composed shortly before he turned three.

Now, proud parent I may be, but let’s be clear – I’m not claiming that he pulled his mini IKEA chair up to a tiny bureau, called for his quill, and committed his thoughts to parchment with an inky flourish (in fact, now five years of age, Smidge is still largely a pen-refuser).

Rather, he had discovered the pleasure of pairing words like “dog” and “cot” as part of a developing awareness of assonance – vowels that are shared across words, even if their consonants differ.

From there, he started to exclaim over his breakfast – “Mummy! “Spoon” and “moon” rhyme!” We were then just a hop, skip and a jump from his first couplet, gleefully stitched together and bellowed as we puddle-jumped our way to the bus one autumn morning –

Cosy wosey,
Rosy nosy!

A very cheerful rendition of autumn from a fledgling poet wrapped up in fleece and mittens.

Smidge’s increasingly conscious play with vowels also reminded me of one of my odder teaching moments. I was in my fourth year teaching Child Language Acquisition at the University of Huddersfield. It was always one of my favourite courses to teach, with a different group each year of enthusiastic, thoughtful, funny final year students.

I started from the beginning, presenting some pretty incredible work showing how children in utero can perceive some vowel sounds and tell them apart about 10 weeks before they are due to be born.

The class started looking at me a little strangely. One or two even giggled.

Of course. I was standing (well, perching) in front of them, about 30 weeks pregnant. As I was passing on degree-level knowledge to my students, I was simultaneously passing on foundational knowledge about English intonation and vowels to Smidge there on the inside, invisible and yet very much present on the room with us. Largely silent, but very much listening in.

A heavily pregnant person smiling at the camera on a set of pretty stairs, looking out at the Barcelona skyline.
Me, not stood in front of my Child Language Acquisition class.
Smidge at about 32 weeks gestation, still listening.

Jump now to late 2023. Inspired by my academic knowledge and my experiences as a parent, I’m standing in front of a room full of preschoolers and their parents, at an event called I’ve designed with two colleagues, a linguist (the superb Dr Emma Nguyen) and a poet (the horribly talented Harry Man).

The audience includes nursery friends of Smidge. I ask the children when they learned their first sound. Quick as a flash, Friend-of-Smidge calls out, “Alphablocks!” A gorgeously guileless answer from someone becoming much more consciously aware of the world of sounds around him as he becomes immersed in primary school phonics schemes (on which more another time…)

Two acquisitionists and a poet walk into a children’s story centre…
Photo credit: Luke Waddington

I near blow Friend-of-Smidge’s mind when I tell him that he learned his first sounds from inside his mum’s tummy. But this is mostly just set up: for the rest of the hour with the preschoolers, my colleagues and I play on the children’s earliest linguistic experiences – distinguishing vowels – to create poetry with them. 1960s experimental French poetry, no less.

Univocalisms are poems written using just one of the five (orthographic) vowels – A, E, I, O or U. Of course, each of these letters can represent a range of different phonemes (tall, pan, saga, agar)*, and sometimes different orthographic vowels can represent the same phoneme (feet, mini)**. We also played very fast and loose (for linguists) by allowing Y in all cases. But we only had an hour to work with here, so we felt justified!

We provided some worksheets and some envelopes of words for each vowel and encouraged the children to come up with sentences consisting of two noun phrases (containing an adjective and a noun), linked by a verb – so introducing a little extra linguistic vocabulary along the way. The only other instruction was to go as crazy as possible – we wanted as many fancy prawns to scan Granny’s pants as possible!

A fab “A” creation by Brianna, with some sea creatures for extra pizzazz

And they did an amazing job! Our fantastic student assistant Lainie created a zine of their best efforts, which we sent around to all the participating families and you can view here (spot also my own effort, as well as poems by Emma, Harry and Lainie). Huge thanks to the AHRC and the Being Human Festival for funding our event. We now hope to take it around primary schools in the area to encourage slightly older children, around 5-7 years old, to keep playing with their languages and to sneak a bit of early linguistics into their consciousness too.

Here’s a bit of play with U and O (we weren’t going to be too strict about the “uni” part of the univocalism… This poet, Sam, was only three…

What did we learn from the event? That there’s a lot of mileage in just getting children to make funny sounds (even more so if funny faces are involved), that there are few things funnier than the idea of an eggy bee stretching smelly jelly (I mean, genius) and that we have so, so much more to learn about what children know, implicitly and explicitly, about language in the preschool years.

*t[ɔː]ll, p[æ]n, sag[ə], [eɪ]gar
**f[iː]t, min[i] – give or take a bit of lengthening of the vowel, indicated by “ː”

What do kids know about pronouns?

It’s 7.30pm on a July evening. I’m lying in the dark, small shafts of light filtering into the room between the blackout blinds and the curtains. I can’t move and I’m trying not to breathe too loudly.

I’m trying to get child #2 (we’ll call him Squidge) to sleep. Squidge, not quite 2-and-a-half, is lying next to me and having none of it. He just wants to try out his new linguistic skills.

His eyes are still open and my arm is going numb, so I decide there’s no harm in getting more comfortable. As I shift in his narrow bed, a strand of my hair falls across Squidge’s face.

“Put my hair on your nose,” he says.
“Put your hair on your nose.
Put your hair on my nose.”

I make an immediate mental note of this pronominal poetry for later.

The same week I receive an email from a colleague who listened to my Free Thinking episode on Childhood and Play, asking the following:

I’m particularly interested in the move from third-person to first-person speech in early childhood: primarily (it seems, as a non-expert) structured and modelled by the parent. There’s a move from “where’s Peter’s nose?” and “it’s mummy’s turn…now it’s Peter’s turn”, to “my” and “your” with the timing of the move possibly(?) led by the child.

Do you know if there’s been much research into this phenomenon and what it reflects and depends on?

There are two slightly different questions here, which meet each other in their interest in how children learn pronouns, which are indexical parts of speech; that is, items in language that shift their meaning depending on context.

You and me always…

Let’s take Squidge’s linguistic musings first, because there’s quite a bit more work been done on this than on the second question. Let’s describe what Squidge does first.

In commenting on the scene, Squidge intended to say “You put your hair on my nose.” Abstracting away from the dropping of the first “you” (a common process in early speech), he ‘reverses’ the second-person possessive pronoun “your”, using first-person possessive “my” instead, in reference to the noun “hair”. He then repeats the trick in the opposite direction with reference to the noun “nose”. This is not an isolated incident either – though this moment stood out because of the way in which he noticed his own “error”*, I was quite used by now to him referring to himself using second-person pronouns.

It’s unclear how common reversal of pronouns is in child language. It’s typically associated with children who are autistic, blind, or hard-of-hearing, though very early talkers also seem to reverse pronouns with some frequency, and eldest children seem more prone to do so too.

Squidge doesn’t fall into any of these categories (as far as we know), but there are also plenty of case-studies of typically developing children reversing pronouns, especially swapping out first- for second-person pronouns (the “nose” case), around two years of age. One other way in which he does align with other children in the literature, including some children with autism, is that he is quite an imitative child – impressionistically, he’s a much more enthusiastic (usually immediate) repeater of what he hears than, say, his older brother. So what’s going on?

I am me, so are you…

In general, pronouns present a fascinating example of the difference between children’s production of language (their performance) and their comprehension of it (their competence). While most children use first person pronouns very accurately early on, followed by second and third person pronouns (e.g. he, she, they), they tend to interpret second person pronouns more accurately to start with, followed by first and third person pronouns. In short, they appear to be using first-person pronouns before they’ve fully grasped the full meaning and use-conditions of them.

It’s easy to see why it could be tricky to interpret the difference between “you” and “me/I”, precisely because they shift their referent with each new speaker. Well, sort of. Though the only person who will use “I” to refer to me is me,** I am “you” to a whole range of people. So, plausibly, “you” is just another name for me – at least, this is a hypothesis that a child could make.

That said, that doesn’t seem to be what Squidge is doing, because he’s reversing possessive pronouns (i.e. your and my) as well. Another theory is that some children might understand that there’s a perspective shift involved in pronouns, but not when and how that works. This is also a plausible attempt to make sense of pronouns, given how other parts of language like go and come are more or less acceptable based on perspective.

Whatever a child’s initial hypothesis about you and me, they’ll be “you” to others throughout their life. So what makes a child, especially one who has been prone to pronoun reversal, shift like an adult?

When two become three

The early talker studies suggest a couple of triggers that might lead children out of the pronoun jungle. One is based on the observation that many of the reported pronoun-reversers in the literature are eldest children, who won’t necessarily have been exposed to many “triadic discourses” – that is, conversations involving three (or more) people.

Triadic conversations not only have multiple Is, they will also involve at least two yous that aren’t your own self. It might be a coincidence to share a name with one other person in the room, but add another and it starts to look like something else is going on (unless, like me, that name is Rebecca and you were born in the late 1980s. Then, you’re never alone).

Given that triadic conversations become more of a common experience for children when they experience childcare outside the home, or a family holiday, these things could constitute an external trigger that would make a child rethink how they interpret you. Indeed, Yuriko Oshima-Takane and her colleagues showed that second-born children, who are often in triadic situations from birth, are generally more adult-like in their pronoun use from an earlier age than firstborn children.

Pronouns? They’re child’s play!

Another trigger proposed by Evans and Demuth is a cognitive one. They noticed that a number of the neurotypical pronoun-reversers in the literature stopped reversing pronouns around 2;4 (that’s 2 years 4 months of age). In the case of the child they studied, the shift was very abrupt – she went from reversing nearly 100% of her first-person pronouns to none at all between recording sessions (that’s to say, within about a week).

As a result, they suggest that some internal process of cognitive maturation taking place at around age 2;4 might contribute to the child’s ability to use different linguistic terms to refer to different participants in a conversation. Naomi B. Schiff-Myers, in her observation of her own young pronoun-reverser, Lauren, noticed that Lauren stopped reversing pronouns around the same time that she started to role-play with dolls and invite more imaginative play.

Indeed, two months on from that long evening with Squidge, I can report that he doesn’t pronoun-reverse any more. It’s hard to say exactly when he stopped, but we have also noticed a huge uptick in playing imaginative games, where he takes on a range of roles and interacts more with other ‘players’ than he might have done previously. He seems, at least, to be one more anecdotal brick to add to the ‘cognitive maturation’ wall.

Naming the issue

And so to the second question about self-reference, or more precisely, referring to yourself by your own name. When children do this, effectively referring to themselves in the third-person, how do they switch to more adult-like first person pronouns?

Let’s give an example first. A friend of mine came round for a cup of tea and I witnessed the following exchange between her and her two-year-old daughter Jane (all names are changed):

Friend: Mummy’s drinking her water. Can Jane drink her water?
[hands Jane her water bottle]
Child: Jane’s bottle.

My friend was very consistent in referring to herself as Mummy and her child as Jane at first mention in an utterance, and using third person pronouns thereafter. Jane was similarly consistent in using full names and didn’t produce a single personal pronoun during that visit (though now aged three, she uses them as accurately as any other child).

Interestingly, it seems that the consistent full-namers, like Jane, may be a completely different group of children than the pronoun-reversers like Squidge. Schiff-Myers (mum of pronoun-reverser Lauren) notes, based on another study by Rosalind Charney, that children use their own names to avoid pronoun confusion (a claim corroborated in another, later study), but that this is not a route available to Lauren as she does not refer to herself using her own name (though she demonstrably knows it). I should note here that neither Squidge nor we as caregivers frequently use full names for self-reference either.

Schiff-Myers even goes on to suggest that using a child’s name instead of pronouns might be of clinical use for children who have delayed language, though this ignores the fact that using a personal name in place of pronouns consistently is marked in typical adult speech and might not come naturally to some caregivers.

How common is it for caregivers to use personal names in place of pronouns? It has been reported that personal names are used much more frequently in child-directed speech than in adult-directed speech – Adi-Bensaid and colleagues report this for caregiver and child names in Hebrew child-directed speech. It is likely that at least some of this uptick in personal name use is in place of pronouns, though other studies find that personal names lag behind far pronouns when used as subjects and objects of sentences in child-directed speech.

Essentially, not much work has been done on how caregivers refer to themselves and to their children, and whether idiosyncratic differences in self-naming affect how children acquire pronouns. We also don’t know much about how full-namer children transition out of full-naming themselves, as this route to pronoun acquisition hasn’t really been studied.

That said, I would expect the triggers that nudge pronoun-reversers towards adult-like pronoun use to have a similar effect on full-namers. As full-namers experience triadic conversations with new conversational partners, they will likely be exposed to people who don’t use personal names in place of pronouns. They will therefore, like the pronoun-reversers, gain more evidence that “you” shifts with the addressee and “I” with the speaker.

The effect of a step in cognitive development is harder to hypothesise about, as we might expect full-namers to continue role-playing with full names. But perhaps, again, role-play with other players would likely lead them to be exposed to different ways of speaking than the child-directed speech they’re used to.

At least now I know what to propose to this year’s students by way of dissertation topics!

Notes

*I don’t really like to refer to children’s non-target-like utterances as “errors”, because according to their grammar at the time, it might not be an error at all. However, “error” is a lot shorter, quicker to read and more transparent than “non-target-like utterance”, so I will scare-quote it instead.


** Forgive me this tortuous locution – I was trying to avoid saying “Only I can call myself “I”. While on one reading (the referential reading), this much shorter statement is true, it’s also untrue on another reading, the bound variable reading, which garnered much attention in formal linguistics following an influential article by the legendary Angelika Kratzer in 2009. They say formal linguists are nerdy…they’re not wrong.

References

Charney, Rosalind. 1980. Speech roles and the development of personal pronouns. Journal of Child Language 7, 509-528. DOI: https://doi.org/10.1017/S0305000900002816

Dale, Philip & Catherine Crain-Thoreson. 1993. Pronoun reversals: Who, when, and why? Journal of Child Language, 20(3), 573-589. doi:10.1017/S0305000900008485

Evans, Karen E. and Katherine Demuth. 2012. Individual differences in pronoun reversal: evidence from two longitudinal case studies. Journal of Child Language 39(1), 162-191. DOI: https://doi.org/10.1017/S0305000911000043

Kratzer, Angelika. 2009. Making a Pronoun: Fake Indexicals as Windows into the Properties of Pronouns. Linguistic Inquiry 40(2), 187–237. doi: https://doi.org/10.1162/ling.2009.40.2.187

Laakso, Aarre, & Linda B. Smith. 2007. Pronouns and verbs in adult speech to children: A corpus analysis. Journal of Child Language, 34(4), 725-763. doi:10.1017/S0305000907008136

Oshima-Takane, Yuriko, Elizabeth Goodz and Jeffrey L. Derevensky. 1996. Birth order effects on early language development: do secondborn children learn from overheard speech? Child Development 67(2), 621-634. Stable URL: https://www.jstor.org/stable/1131836

Schiff-Myers, Naomi. B. 1983. From pronoun reversals to correct pronoun usage: A case study of a normally developing child. Journal of Speech and Hearing Disorders 48, 385–94. DOI: https://doi.org/10.1044/jshd.4804.394

What do kids know about accents and dialects?

One of the things I was most excited about when moving to Newcastle was being surrounded by the fabulous, famous, accent (indeed, accents) that you hear on the banks of the Tyne.

Don’t get me wrong, I have a massive amount of affection for the variety of Yorkshire accents that had been my soundtrack for over a decade, and I was learning so much more about the nuances of West Yorkshire accents living and playing in Leeds. But the Geordie accent was part of my childhood – my dad’s family are from North Tyneside – and I was so intrigued to see what our then 10-month-old would pick up from the local accent, which differs in so many ways from my generic Northern and his dad’s generic Midlands.

The result was, and continues to be, fascinating. At nursery three days a week with carers principally from Newcastle and South Tyneside, our wee one (we’ll call him Smidge here) developed a weekday vs weekend split, particularly with respect to the vowel in words like “book”.

What would you call these?
Image credit: CC-BY-SA 2010 See-ming Lee 李思明 SML

In many British English accents, the vowel in “book” is what we call a near-close, near-back rounded vowel – listen and look here at my wonderful colleague, Prof. Ghada Khattab, producing it. To break it down, that means that the tongue is nearly touching the roof of the mouth (near-close) fairly near the back of the hard palate (near-back) and the lips are rounded (…well, rounded). Have a look at this MRI of the tongue in action during the production of vowels if you want to visualise this a bit more clearly. This vowel is represented by the symbol /ʊ/ in the International Phonetic Alphabet, so the word “book” is typically pronounced in English as /bʊk/.

Not so in Tyneside English, especially south of the river in South Tyneside, where [bu:ks]* are enjoyed. The vowel [u:] is a close back rounded lengthened vowel, so the tongue is very close** to the very back of the roof of the mouth, near the soft palate, the lips are rounded and the sound lasts a little longer.

Back to Smidge and his reading matter. He’s about 22 months old and he’s been at nursery Monday to Wednesday for about 2 months since it reopened after the first Covid lockdown (he’d managed all of about 6 weeks at nursery prior to it). Despite this brief and relatively late exposure to Tyneside English, he’s asking for [bu:ks] at the start of the week and [bʊks] by the end of the weekend…just in time to read [bu:ks] again with his carers the next week.

One of my colleagues in theoretical phonetics and phonology was quite surprised that such a young child had two such clear variants for the same vowel with, apparently, set days of the week for the use of each.

But for an acquisitionist it’s not quite so surprising, especially as it’s not really to do with the days of the week.

We know from studies of multilingual children that they distinguish their languages very early on – for example, they’re aware which of their conversational partners uses each of their languages and they choose their words according to the context. They also demonstrate their identity and social place by manipulating their language in their earliest school years.

Would you “aloha” or “ahoj”? The kids will know…
Image from The Connection Between Multilingualism and Business Success (speexx.com)

It can be harder to see this in monolingual multidialectal children, simply because it can be less obvious when a child is using a particular accent or dialect, especially if their dialects share a number of features. However, we know that children acquire from their parents the stylistic and linguistic rules for certain dialect features from as young as 3 years old, especially ones that users are consciously aware of. At an even younger age, around two-and-a-half, children are sufficiently aware of the social status of different accents that they start to use the more prestigious dialect, even in places where the less prestigious dialect is used, like the home. And a little later, around age 4, children can identify different dialects and group language users according to the vowels they use in a range of words – that is, they’d be able to identify users of [bu:k] and [tɪʔkɪʔ] as a group distinct from users of [bʊk] and [tɪkɪʔ], assuming they had experience of Tyneside English.

So children know a *lot* more about language, and how we all use it, than they’re often given credit for.

Back to Smidge: what does he do now? He’s starting school in September and he is most commonly a [buk] reader – his vowel may be slightly shorter than that of a true Tynesider, but he’s definitely backer and closer than me and his dad. Of course, this isn’t “inappropriate” use or use of the “wrong” dialect, because the change in the vowel doesn’t obscure the meaning of the word for us [bʊk]-lovers. Moreover, we would *never* correct his pronunciation of a word like this, because we both believe that all accents and dialects are valid, and we want Smidge to develop his own style and to use it confidently. Ach, maybe that sounds very worthy, but it’s true – kids are aware enough to know when your corrections are based on understanding and when they’re based on prejudice. They’ve been around the block far too long to be hoodwinked.

* The lengthening diacritic in [u:] isn’t quite as we’d write it in the IPA – rather than a colon with two dots, it looks more like two triangles stacked on top of each other, pointing towards each other. But WordPress’s typefaces won’t play today – take a look at Watt and Allen’s (2003) illustration of Tyneside English for the accurate diacritics.

** Obviously the tongue in [u] isn’t fully closed against the roof of the mouth – a defining characteristic of a vowel is that there is no full closure between any articulators. If the tongue fully met the roof of the mouth here, you’d end up with a /k/ or /g/ (if you discount the lip rounding!).

A Bender-style note: The studies cited in the multilingualism paragraph concern communities using English/French and Light Warlpiri/Warlpiri/English. In the multidialectal paragraph studies concern communities using Buckie English/Standard Scots English, Dutch/Limburgish, and Yorkshire/Standard Southern British English.

And an Easter Egg: If you read the dialect-grouping study (Jeffries 2019), you’ll know that “[s]timuli for the experiment were recorded from one bidialectal speaker, a 25-year-old female, who was able to switch between two different accents and produce vowel pronunciations typical of both Yorkshire and the South East of England.” It was me!