Marianne de Heer Kloots's Avatar

Marianne de Heer Kloots

@mdhk.net

Linguist in AI & CogSci πŸ§ πŸ‘©β€πŸ’»πŸ€– PhD student @ ILLC, University of Amsterdam 🌐 https://mdhk.net/ 🐘 https://scholar.social/@mdhk 🐦 https://twitter.com/mariannedhk

1,277
Followers
518
Following
111
Posts
06.08.2023
Joined
Posts Following

Latest posts by Marianne de Heer Kloots @mdhk.net

Preview
Shifting neural code powers speech comprehension Dynamic coding helps explain how the brain processes multiple features of speechβ€”from the smallest units of sounds to full sentencesβ€”simultaneously.

thank you @claudia-lopez.bsky.social for this clear and thoughtful coverage in @thetransmitter.bsky.social on some of our recent work on dynamic neural codes in speech processing! 🧠 πŸŒ€ ✨ www.thetransmitter.org/language/shi...

06.03.2026 22:28 πŸ‘ 21 πŸ” 1 πŸ’¬ 1 πŸ“Œ 0
Post image

πŸ“’ PhD position in the NeuroAI of Language

Why can LLMs predict brain activity so well? We're hiring a PhD student to find out -- AI interpretability meets neuroimaging
Deadline March 20
Please RT πŸ™
πŸ‘‡
mpi.nl/career-education/vacancies/vacancy/fully-funded-4-year-phd-position-neuroai-language

05.03.2026 13:34 πŸ‘ 45 πŸ” 35 πŸ’¬ 2 πŸ“Œ 1

I did a MSc project on seal pup vocalization timing in long rehab centre recordings years ago, may still have some relevant refs scripties.uba.uva.nl/search?id=re... and I believe Koen de Reus' PhD thesis (to be defended this week) has a bunch more up-to-date work on this! www.mpi.nl/events/imprs...

18.02.2026 10:11 πŸ‘ 3 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

Our paper has been accepted to EACL 2026!πŸŽ‰ We systematically evaluate several vision-language (VLMs) and language-only models, measuring their alignment with brain responses to concept words. Our results show that vision-language models offer a promising tool to model human concept processing

23.01.2026 12:02 πŸ‘ 14 πŸ” 4 πŸ’¬ 1 πŸ“Œ 0
Post image

Thrilled to announce the 1st Workshop on Computational Developmental Linguistics (CDL) at ACL 2026 πŸŽ‰ A new venue at the intersection of development linguistics Γ— modern NLP, spearheaded by @fredashi.bsky.social @marstin.bsky.social, and and outstanding team of colleagues!

A thread 🧡

20.01.2026 11:26 πŸ‘ 22 πŸ” 9 πŸ’¬ 3 πŸ“Œ 1
4th Dutch Speech Tech Day We are thrilled to announce that the 4th Dutch Speech Tech Day will take place on Monday, February 2, 2026, at Beeld & Geluid in Hilversum!

If you're in the Netherlands or nearby, check out the Fourth Dutch Speech Tech Day. I'll be there.
www.eventbrite.com/e/4th-dutch-...

13.01.2026 09:58 πŸ‘ 3 πŸ” 1 πŸ’¬ 0 πŸ“Œ 0
Commentary title: 
Linguists should learn to love speech-based deep learning models 

Authors: 
Marianne de Heer Kloots, Paul Boersma, Willem Zuidema

Abstract: 
Futrell and Mahowald present a useful framework bridging technology-oriented deep learning systems and explanation-oriented linguistic theories. Unfortunately, the target article's focus on generative text-based LLMs fundamentally limits fruitful interactions with linguistics, as many interesting questions on human language fall outside what is captured by written text. We argue that audio-based deep learning models can and should play a crucial role.

Commentary title: Linguists should learn to love speech-based deep learning models Authors: Marianne de Heer Kloots, Paul Boersma, Willem Zuidema Abstract: Futrell and Mahowald present a useful framework bridging technology-oriented deep learning systems and explanation-oriented linguistic theories. Unfortunately, the target article's focus on generative text-based LLMs fundamentally limits fruitful interactions with linguistics, as many interesting questions on human language fall outside what is captured by written text. We argue that audio-based deep learning models can and should play a crucial role.

'Tis the season to preprint BBS commentaries; I'm happy to share ours too! πŸŽ„βœ¨

The textual basis of current LLMs causes trouble, but linguistically relevant insights *can* be found in systems modelling the more natural form of human spoken language: the speech signal itself. arxiv.org/abs/2512.14506

17.12.2025 15:21 πŸ‘ 27 πŸ” 10 πŸ’¬ 1 πŸ“Œ 1

New book! I have written a book, called Syntax: A cognitive approach, published by MIT Press.

This is open access; MIT Press will post a link soon, but until then, the book is available on my website:
tedlab.mit.edu/tedlab_websi...

24.12.2025 19:55 πŸ‘ 122 πŸ” 41 πŸ’¬ 2 πŸ“Œ 3
Preview
How cognitive science can contribute to AI: methods for understanding #2 in a series on cognitive science and AI

New post! Last week I shared why I thought cognitive (neuro)science hasn’t contributed as much as one might hope to the design of AI systems; this week I'm sharing my thoughts on how methods and principles from these fields *have* been useful in my work. infinitefaculty.substack.com/p/how-cognit...

23.12.2025 17:09 πŸ‘ 42 πŸ” 8 πŸ’¬ 2 πŸ“Œ 3
Poster title: Does multimodal pre-activation influence linguistic expectations in LLMs and humans?

Authors: Sasha Kenjeeva, Giovanni Cassani, Noortje Venhuizen, Afra Alishahi

Poster title: Does multimodal pre-activation influence linguistic expectations in LLMs and humans? Authors: Sasha Kenjeeva, Giovanni Cassani, Noortje Venhuizen, Afra Alishahi

Poster title: Generalizing Without Evidence: How Transformer Models Infer Syntactic Rules From Sparse Input

Authors: Mark van den Hoorn, Raquel G. Alhama

Poster title: Generalizing Without Evidence: How Transformer Models Infer Syntactic Rules From Sparse Input Authors: Mark van den Hoorn, Raquel G. Alhama

Poster title: Dependency Length, Syntactic Complexity & Memory: A Reading Time Benchmark for Sentence Processing Modeling

Authors: Nina Nusbaumer, Corentin Bel, Iria de-Dios-Flores, Guillaume Wisniewski, Benoit CrabbΓ©

Poster title: Dependency Length, Syntactic Complexity & Memory: A Reading Time Benchmark for Sentence Processing Modeling Authors: Nina Nusbaumer, Corentin Bel, Iria de-Dios-Flores, Guillaume Wisniewski, Benoit CrabbΓ©

Poster title: 
The success of Neural Language Models on syntactic island effects is not universal: strong wh-island sensitivity in English but not in Dutch

Authors: Michelle Suijkerbuijk, Naomi Tachikawa Shapiro, Peter de Swart, Stefan L. Frank

Poster title: The success of Neural Language Models on syntactic island effects is not universal: strong wh-island sensitivity in English but not in Dutch Authors: Michelle Suijkerbuijk, Naomi Tachikawa Shapiro, Peter de Swart, Stefan L. Frank

Cool posters from day 2!

@sashakenjeeva.bsky.social openreview.net/forum?id=Vtd...

github.com/markvandenho... openreview.net/forum?id=rX3...

@nina-nusbaumer.bsky.social openreview.net/forum?id=GRz...

www.ru.nl/personen/sui... openreview.net/forum?id=NcJ...

19.12.2025 16:03 πŸ‘ 4 πŸ” 1 πŸ’¬ 1 πŸ“Œ 1
Post image Post image

Today we are time-travelling with @stefanfrank.bsky.social 😎

19.12.2025 12:18 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0
Beyond surprisal: GPT-derived attention metrics offer additional... The N400 component of the EEG signal is a well-established neural correlate of real-time language comprehension, sensitive to a range of lexical and contextual variables. While earlier studies have...

I’m really enjoying the Computational Psycholinguistics Meeting in Utrecht! cpl2025.sites.uu.nl 🧠

Yesterday MSc student Sven Terpstra (co-supervised w/ @wzuidema.bsky.social) presented his project on predicting the N400 with GPT-derived metrics beyond surprisal openreview.net/forum?id=MAl...

19.12.2025 12:18 πŸ‘ 9 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0

This is in response to the recent target article by @futrell.bsky.social & @kmahowald.bsky.social, a wonderful read! Thanks to the authors for writing such a clear and inspiring piece to start this discussion. doi.org/10.1017/S014...

17.12.2025 15:21 πŸ‘ 1 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
Commentary title: 
Linguists should learn to love speech-based deep learning models 

Authors: 
Marianne de Heer Kloots, Paul Boersma, Willem Zuidema

Abstract: 
Futrell and Mahowald present a useful framework bridging technology-oriented deep learning systems and explanation-oriented linguistic theories. Unfortunately, the target article's focus on generative text-based LLMs fundamentally limits fruitful interactions with linguistics, as many interesting questions on human language fall outside what is captured by written text. We argue that audio-based deep learning models can and should play a crucial role.

Commentary title: Linguists should learn to love speech-based deep learning models Authors: Marianne de Heer Kloots, Paul Boersma, Willem Zuidema Abstract: Futrell and Mahowald present a useful framework bridging technology-oriented deep learning systems and explanation-oriented linguistic theories. Unfortunately, the target article's focus on generative text-based LLMs fundamentally limits fruitful interactions with linguistics, as many interesting questions on human language fall outside what is captured by written text. We argue that audio-based deep learning models can and should play a crucial role.

'Tis the season to preprint BBS commentaries; I'm happy to share ours too! πŸŽ„βœ¨

The textual basis of current LLMs causes trouble, but linguistically relevant insights *can* be found in systems modelling the more natural form of human spoken language: the speech signal itself. arxiv.org/abs/2512.14506

17.12.2025 15:21 πŸ‘ 27 πŸ” 10 πŸ’¬ 1 πŸ“Œ 1
Preview
Why Do Humans Have Linguistic Intuition? | Cadernos de LinguΓ­stica

Why do humans have linguistic intuition? And why should you care?

A short thread about my new paper in @cadlin.bsky.social

This work has the most original insight I've ever had, a genuinely new idea about the nature of language

cadernos.abralin.org/index.php/ca...

1/20

15.12.2025 16:14 πŸ‘ 57 πŸ” 20 πŸ’¬ 3 πŸ“Œ 1
Preview
What’s Surprising About Surprisal - Computational Brain & Behavior In the computational and experimental psycholinguistic literature, the mechanisms behind syntactic structure building (e.g., combining words into phrases and sentences) are the subject of considerable...

Many studies of naturalistic comprehension report that surprisal (often LLM derived) explains more of the variance in data than other predictors. Why is this? And why can it be problematic for our conclusions?

A 🧡 of takeaways from our paper doi.org/10.1007/s421... with @andreaeyleen.bsky.social

17.11.2025 17:12 πŸ‘ 23 πŸ” 8 πŸ’¬ 1 πŸ“Œ 1
Preview
Why isn’t modern AI built around principles from cognitive science? First post in a series on cognitive science and AI

Why isn’t modern AI built around principles from cognitive science or neuroscience? Starting a substack (infinitefaculty.substack.com/p/why-isnt-m...) by writing down my thoughts on that question: as part of a first series of posts giving my current thoughts on the relation between these fields. 1/3

16.12.2025 15:40 πŸ‘ 117 πŸ” 34 πŸ’¬ 4 πŸ“Œ 5

I had a wonderful time getting to know @gronlp.bsky.social last week while discussing linguistic structure and learning trajectories in speech models! ✨ Many thanks for the invite @frap98.bsky.social, already looking forward to catching up again soon :)

01.12.2025 17:13 πŸ‘ 13 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0
The title page

The title page

🚨NEW PUBLICATION ALERT!🚨
The 'Design Features' of Language Revisited (w/ @mperlman.bsky.social @glupyan.bsky.social Koen de Reus & @limorraviv.bsky.social)
Feature Review out now in #OpenAccess in @cp-trendscognsci.bsky.social! #language #linguistics
Paper: doi.org/10.1016/j.ti...

25.11.2025 19:48 πŸ‘ 102 πŸ” 32 πŸ’¬ 3 πŸ“Œ 2

Interested in the evolution of human language? Check out our new paper in @science.org where we synthesize latest findings and outline a multifaceted, bio-cultural approach for studying how language evolved. Super proud of this work, and hoping it leads to exciting new research! tinyurl.com/ykacvanp

21.11.2025 09:47 πŸ‘ 41 πŸ” 15 πŸ’¬ 2 πŸ“Œ 1

The Multilingual Minds & Machines Meetings call for abstracts is now open! Everything you need to know is here -> mmmm2026.github.io

18.11.2025 10:05 πŸ‘ 10 πŸ” 7 πŸ’¬ 0 πŸ“Œ 1
Preview
Human cortical dynamics of auditory word form encoding We perceive continuous speech as a series of discrete words, despite the lack of clear acoustic boundaries. The superior temporal gyrus (STG) encodes …

happy to share our new paper, out now in Neuron! led by the incredible Yizhen Zhang, we explore how the brain segments continuous speech into word-forms and uses adaptive dynamics to code for relative time - www.sciencedirect.com/science/arti...

07.11.2025 18:16 πŸ‘ 48 πŸ” 17 πŸ’¬ 2 πŸ“Œ 1
PNAS Proceedings of the National Academy of Sciences (PNAS), a peer reviewed journal of the National Academy of Sciences (NAS) - an authoritative source of high-impact, original research that broadly spans...

Delighted to share our new paper, now out in PNAS! www.pnas.org/doi/10.1073/...

"Hierarchical dynamic coding coordinates speech comprehension in the brain"

with dream team @alecmarantz.bsky.social, @davidpoeppel.bsky.social, @jeanremiking.bsky.social

Summary πŸ‘‡

1/8

22.10.2025 05:21 πŸ‘ 93 πŸ” 35 πŸ’¬ 2 πŸ“Œ 5
Post image

🌍Introducing BabyBabelLM: A Multilingual Benchmark of Developmentally Plausible Training Data!

LLMs learn from vastly more data than humans ever experience. BabyLM challenges this paradigm by focusing on developmentally plausible data

We extend this effort to 45 new languages!

15.10.2025 10:53 πŸ‘ 44 πŸ” 16 πŸ’¬ 1 πŸ“Œ 4
Preview
Learning without training: The implicit dynamics of in-context learning One of the most striking features of Large Language Models (LLM) is their ability to learn in context. Namely at inference time an LLM is able to learn new patterns without any additional weight updat...

Interesting paper suggesting a mechanism for why in-context learning happens in LLMs.

They show that LLMs implicitly apply an internal low-rank weight update adjusted by the context. It’s cheap (due to the low-rank) but effective for adapting the model’s behavior.

#MLSky

arxiv.org/abs/2507.16003

06.10.2025 13:30 πŸ‘ 59 πŸ” 19 πŸ’¬ 1 πŸ“Œ 5
Preview
PhD Position: Accented Speech Processing | Radboud University Do you want to work as a PhD: Accented Speech Processing at the Faculty of Arts? Check our vacancy!

PhD Position: Accented Speech Processing - Apply now!

Come work with Mirjam Broersma, @davidpeeters.bsky.social, and me at the Centre for Language Studies, Radboud University in the Netherlands.

Application deadline: 19 October 2025

For more information, see
www.ru.nl/en/working-a...

02.10.2025 14:35 πŸ‘ 26 πŸ” 31 πŸ’¬ 0 πŸ“Œ 0

Huge congrats to the envisionBOX team for the Open Science award nomination! πŸŽ‰

My tutorial on speech analysis tools in Python from the Unboxing Multimodality summer school (github.com/mdhk/unboxin...) is now also available at envisionbox.org

Thanks for the invitation & this great initiative! πŸ‘

02.10.2025 17:18 πŸ‘ 10 πŸ” 3 πŸ’¬ 0 πŸ“Œ 0
Post image Post image Post image Post image

The π—œπ—Ÿπ—–π—• π—¦π˜‚π—Ίπ—Ίπ—²π—Ώ 𝗦𝗰𝗡𝗼𝗼𝗹 in Marseille went beyond all my expectations! πŸ’―

A week has already flown by since I had one of the most formative experiences of my PhD so far. πŸ‘©β€πŸŽ¨

12.09.2025 09:52 πŸ‘ 6 πŸ” 2 πŸ’¬ 1 πŸ“Œ 0

Thanks to all co-authors in the Dutch SSL training team @hmohebbi.bsky.social @cpouw.bsky.social @gaofeishen.com @wzuidema.bsky.social + Martijn Bentum

And to @itcooperativesurf.bsky.social (EINF-8324) for granting me the resources that enabled this project πŸ‘©β€πŸ’»βœ¨

27.08.2025 14:31 πŸ‘ 0 πŸ” 0 πŸ’¬ 0 πŸ“Œ 0

Check out the paper for more details:
πŸ“„ arxiv.org/abs/2506.00981

Or the model, dataset and code released alongside it:
πŸ€— huggingface.co/amsterdamNLP...
πŸ—ƒοΈ zenodo.org/records/1554...
πŸ” github.com/mdhk/SSL-NL-...

We hope these resources help further research on language-specificity in speech models!

27.08.2025 14:31 πŸ‘ 2 πŸ” 0 πŸ’¬ 1 πŸ“Œ 0