The writing on a recently-discovered tablet has linguists baffled, and GPT passes the Turing Test

Here’s what happened this week in language and linguistics.

The writing on a recently-discovered tablet has linguists baffled, and GPT passes the Turing Test

Welcome to this week’s edition of Discovery Dispatch, a weekly roundup of the latest language-related news, research in linguistics, interesting reads from the week, and newest books and other media dealing with language and linguistics.

💡
If you don't want this weekly digest, you can manage your email subscriptions here. There you can unsubscribe from Discovery Dispatch and still stay subscribed to the newsletter (if you'd like).

📢 Updates

Announcements and what’s new with me and Linguistic Discovery.

Me posing with fellow content creators Dimitri (@dkdynamite1) and Christine (@christines_snapss) at the YouTube Creator Collective event in Chicago.

This week I attended one of YouTube’s Creator Collective events in Chicago, and had a blast meeting tons of other creators and talking shop about newsletters vs. videos, shortform vs. longform, how they’re monetizing, all the jazz. I came with a few new friends and very encouraged to continue my content creation journey! Thank you again to all of you who have supported me with either your precious time and attention, or your financial support along the way. It means a lot to have all of you reading and enjoying this newsletter.

So let’s dive in! Here’s what’s new in the world of linguistics this week:

📰 In the News

Language and linguistics in the news.

A stencil of the recently-discovered Bashplemi tablet covered in undeciphered inscriptions.

A stone tablet was discovered near Lake Bashplemi, Georgia in 2021, bearing an inscription in a script unknown to linguists, historians, and paleographers. There are 60 symbols total, 39 of which are unique. The tablet is dated from sometime in the 1st millennium BCE, which is around the same time that the Phoenician alphabet was spreading into Europe and becoming the Greek alphabet. This script, however, bears little resemblance to Phoenician.

Mystery language on ancient stone tablet stumps archeologists
The Bashplemi inscription possibly shares similarities with over 20 other alphabets.

🗞️ Current Linguistics

Recently published research in linguistics.

Inuit languages really do have more words for snow

I’ll have an entire post about this next week, but a new study looks at which areas of vocabulary different languages tend to have more words for, and it seems that Inuit languages really do have more words for snow:

Do Inuit languages really have many words for snow? The most interesting finds from our study of 616 languages
Can you guess which languages score the highest for terms relating to ‘love’, ‘death’, ‘canoe’ or ‘sheep’? We made a tool to help you explore our extensive dataset.

A new corpus of spoken American English

One of the most important tools with which linguists study how people use language is a corpus (plural: corpora)—a dataset consisting of a collection of texts in a language. (A text in linguistics is any cohesive stretch of discourse, whether that be a written piece of text, a recorded conversation, a poem, a video of sign language users, a traditional oral narrative, etc.) Linguists use large corpora to study how grammatical patterns shift over time, or how certain linguistic features correlate with age, region, or other demographic factors.

One of the most famous corpora is the Google Books corpus, which aims to be a corpus of the text of every book ever published. You can explore its data and examine trends using Google’s Ngram Viewer. More recently, large language models (LLMs) used in generative AI are being trained on absolutely massive corpora of billions upon billions of words. The Common Corpus, for example, consists of 500 billion words in various languages. Historically it had been impractical to create corpora of this size, so these mega-corpora are unlocking new research opportunities for linguists.

The Google Books Ngram Viewer, showing the frequency of the terms “Frankenstein”, “Sherlock Holmes”, and “Albert Einstein” from 1900 until today.

A major limitation of these corpora, however, is that they’re all text. Spoken corpora are much more difficult to compile and there are very few of them in existence. (And don’t even get me started on sign language corpora.) Researchers would have to recruit numerous participants, record them for hours, and spend an unseemly amount of time transcribing the audio. It often takes about 1 hour to transcribe just 1 minute of natural conversation, due to the fact that speech is riddled with restarts, overlap, hesitations, and talking over each other. Corpus linguists often transcribe at a fine level of detail too, indicating filler works like um and uh, intonational cues, length of pauses, and more—each of which provides helpful insights into cognition, grammar, and discourse. While modern speech recognition technology is making the task of transcribing audio significantly easier, it’s still a time-intensive task.

The onerous nature of creating spoken corpora is a shame, because speech is the natural modality for language. Natural conversation gives linguists the best insight into how people use language in real time, whereas writing is more artificial and planned, shackled to its own set of conventions that diverge drastically from how we speak.

So I was thrilled in 2023 when fellow linguistics communicator and now Ph.D.-holder Dr. Elizabeth Hanks (congrats Lizzy! 🎓🎉) told me about her work to create the largest-ever spoken corpus of American English, known as the Lancaster-Northern Arizona Corpus of Spoken American English (LANA-CASE). She asked if I would use my large social media presence to help recruit participants to be recorded for the corpus, and naturally I said yes, eager to support such a great research project. I put together an entire video series on corpus linguistics, which you can watch here:

Now, the data collection is finished, and we’re already beginning to see some early research that this new corpus has enabled! As reported in The Guardian, it seems that a growing number of Britishisms have made their way into American English in recent years. Interestingly, it’s Gen Z that’s the vanguard in adopting these words. You can see the complete top 10 in the article:

Bonkers for Britishisms: the UK terms Americans have embraced
Researchers have catalogued the British words and phrases most used in US conversation, sparking delight and frustration

In any case, I’m proud to have played a tiny role in bringing this corpus to life, and am excited to see the interesting new research it will make possible!

GPT passes the Turing Test

A new study (not yet peer-reviewed) claims that GPT-4.5 passes the Turing Test 73% of the time. In a Turing Test, a human reads a transcript of a conversation between another human and a machine and tries to determine which of the conversationalists is the human and which is the machine. The machine passes if the human cannot reliably tell them apart.

An AI model has finally passed an authentic Turing test, scientists say
GPT-4.5 has successfully convinced people it’s human 73% of the time in an authentic configuration of the original Turing test.
Large Language Models Pass the Turing Test
We evaluated 4 systems (ELIZA, GPT-4o, LLaMa-3.1-405B, and GPT-4.5) in two randomised, controlled, and pre-registered Turing tests on independent populations. Participants had 5 minute conversations simultaneously with another human participant and one of these systems before judging which conversational partner they thought was human. When prompted to adopt a humanlike persona, GPT-4.5 was judged to be the human 73% of the time: significantly more often than interrogators selected the real human participant. LLaMa-3.1, with the same prompt, was judged to be the human 56% of the time -- not significantly more or less often than the humans they were being compared to -- while baseline models (ELIZA and GPT-4o) achieved win rates significantly below chance (23% and 21% respectively). The results constitute the first empirical evidence that any artificial system passes a standard three-party Turing test. The results have implications for debates about what kind of intelligence is exhibited by Large Language Models (LLMs), and the social and economic impacts these systems are likely to have.

Do our brains try to predict what’s coming next in a sentence? It may depend on the language you speak

Do our brains wait til the end of a sentence before trying to interpret it, or do they attempt to predict what’s coming next in the sentence? Previous neurological studies using English suggest that the brain waits for the sentence to end before interpreting everything, but a new study with Dutch suggests that Dutch speakers do in fact try to predict where the sentence is going. This means that different languages may foster different cognitive strategies for interpreting sentences. The reason for this may be that in Dutch sentences verbs come at the end, while in English sentences verbs are in the middle of the sentence. Since verbs are the pillar around which the rest of a clause is structured, it makes sense that Dutch speakers would be thinking ahead to what the verb might be.

Do You Think Differently in Different Languages?
The brain’s response to information depends on language’s grammatical structure
Language-specific neural dynamics extend syntax into the time domain
Comprehending spoken language requires the encoding of information beyond what is presented in the raw speech signals. Using MEG recordings during natural story listening, this study shows that the human brain projects its knowledge of syntax onto speech in a predictive and language-dependent way.

The Pangloss Collection

A new collection of digital corpora of (mostly) endangered or underdocumented languages called The Pangloss Collection has just been released, along with an interface allowing users to listen to and read all the texts in the collection:

Pangloss Collection | Pangloss | Home
Project MUSE - The Pangloss Collection: Opening up research data on endangered and underdocumented languages

Interjections

Most people think of interjections like um, uh, oh, and uh-huh as meaningless filler words, but any linguist with some background in discourse and information flow knows that these are actually crucial parts of natural conversation, signaling to the speaker that you’re listening, or that you need to correct something you just said, or that you want the conversational floor next. They do a huge amount of social work in regulation conversation.

Knowable Magazine has an interview with Mark Dingemanse about his recent review article, “Interjections at the heart of language”, about just how important these types of words—what I’d call discourse markers—actually are.

Huh? The valuable role of interjections
Utterances like um, wow and mm-hmm aren’t garbage — they keep conversations flowing
Interjections at the Heart of Language | Annual Reviews
Interjections, the words that come between sentences, are easily overlooked and usually treated as peripheral to the language sciences. This review surveys work from disparate disciplines that suggests an inversion of perspective: from interjections as marginal items to interjections at the heart of language. Around one out of every seven turns in conversation is an interjection, and the most common ones are not the involuntary exclamations that typically feature in examples; instead, they form a small set of agile and adaptive interactional tools that streamline everyday language use. Continuers like mmhm help people co-construct complex interactional structures, repair initiators like huh? help people calibrate mutual understanding on the fly, and change-of-state tokens like oh display knowledge as it evolves in interaction. Interjections emerge as words that help us talk and think, scaffolding the complexity of language as we know it. The review critically considers received views of interjections as primitive grunts, affect bursts, or symptoms of strain and provides a number of alternative ways of thinking about interjections.

📃 This Week’s Reads

Interesting articles I’ve come across this week.

Cognates with different meanings

Here’s a fun article from the Duolingo blog about why cognates can have the opposite or drastically different meanings across languages. I like that it acknowledges that there are several pathways by which this situation can come about. Here’s one illustration:

Proto-Germanic *slimbaz ‘oblique, crooked’ English: oblique, crooked → slanting → slight → thin Dutch: oblique, crooked → morally crooked → sly → clever
Dear Duolingo: Why can words have opposite meanings across languages?
Some words have the same spelling in different languages but completely different meanings—it’s an accident of history!

What’s the difference between a loan word and a loan translation (also called a calque)? Ryan Starkey of Starkey Comics has a great infographic with the answer:

How to prepare for your bilingual baby
Growing up with more than one language has obvious advantages for children.
How Much Does Our Language Shape Our Thinking?
English continues to expand into diverse regions around the world. The question is whether humanity will be homogenized as a result.
From Igbo to Angika: how to save the world’s 3,000 endangered languages
With half of all languages predicted to die out in decades, activists are turning to online tools to preserve them
Dear Duolingo: Why don’t some words change over time?
Language is always changing—but some words can resist for thousands of years. What are they, and how do they do it?

📚 Books & Media

New (and old) books and media touching on language and linguistics.

I’m excited for this one, y’all! A new book about Proto-Indo-European (PIE) called, appropriately, Proto. It’s pretty cool that our understanding of PIE and the origins of the early Indo-European peoples has continued to evolve sufficiently in over the last two decades that we’re still getting updated looks at the latest science like this. The last popular look at Indo-European, The horse, the wheel, and language, is now almost 20 years old! So this new book is most welcome.

Amazon
Amazon

I should also mention a similar book coming out right around the same time, which I’ve mentioned in a previous issue of the digest: The Indo-Europeans rediscovered: How a scientific revolution is rewriting their story.

Amazon

The King’s Letters is a South Korean historical drama about the invention of the Hangul writing system by Sejong the Great in 1443. Hangul is a unique writing system which combines phonetic units into syllable blocks, making it a combination of a featural alphabet and a syllabary.

Amazon

I hope you enjoyed this week’s issue of Discovery Dispatch! Have a wonderful week!

~ Danny

💡
The Amazon links on this site are affiliate links, which means that I earn a small commission from Amazon for purchases made through them (at no extra cost to you).

If you'd like to support Linguistic Discovery, purchasing through these links is a great way to do so! I greatly appreciate your support!

Check out my entire Amazon storefront here.