Advertisement ยท 728 ร— 90

Posts by Sam Boeve

Preview
In-Mind: Psychology for You! (@in-mindmagazine.bsky.social) ๐Ÿ”” ๐—ก๐—ฒ๐˜„ ๐—ฎ๐—ฟ๐˜๐—ถ๐—ฐ๐—น๐—ฒ: ๐—” ๐—ป๐—ฒ๐˜„ ๐—ฝ๐—ฒ๐—ฟ๐˜€๐—ฝ๐—ฒ๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—ผ๐—ป ๐—น๐—ฎ๐—ป๐—ด๐˜‚๐—ฎ๐—ด๐—ฒ ๐—ฎ๐—ป๐—ฑ ๐—ฐ๐—ผ๐—ด๐—ป๐—ถ๐˜๐—ถ๐—ผ๐—ป ๐Ÿ”” How can computers help us understand language learning? ๐Ÿค– From readability scores toโ€ฆ

๐Ÿ“Š ๐—›๐—ถ๐—ด๐—ต๐—น๐—ถ๐—ด๐—ต๐˜๐˜€ ๐—ณ๐—ฟ๐—ผ๐—บ ๐Ÿฎ๐Ÿฌ๐Ÿฎ๐Ÿฑ:

๐Ÿ“‘ 33 articles were published on In-Mind.โ€จ
๐Ÿ“Œ The most interacted article on Bluesky was โ€œLanguage models: A new perspective on language and cognitionโ€ by @boevesam.bsky.social.
๐Ÿ‘‰ Read the post here again:โ€จbsky.app/profile/in-m...

3 months ago 1 1 1 0

๐Ÿ“‘ ๐—ช๐—ต๐—ฎ๐˜ ๐˜„๐—ฒ ๐—ฝ๐˜‚๐—ฏ๐—น๐—ถ๐˜€๐—ต๐—ฒ๐—ฑ
In 2025, we published 33 articles and 12 blog posts.
๐Ÿ“Œ The most interacted article on Bluesky was โ€œLanguage models: A new perspective on language and cognitionโ€ by @boevesam.bsky.social
๐Ÿ‘‰ Read it again here: โ€จbsky.app/profile/in-m...

3 months ago 3 1 1 0

๐Ÿ‘๏ธ ๐Ÿ“– ๐Ÿ‘๏ธ
@boevesam.bsky.social
made this interactive visualisation to get a feeling for word predictability:
๐Ÿ”— wordpredictabilityvisualized.vercel.app

Curious how these predictability indices were obtained? Find out in our new paper!
๐Ÿ”— doi.org/10.3758/s134...

#Reading #LargeLanguageModels #MECO

7 months ago 3 1 0 0
Word Predictability Visualization App

Want to explore word predictability yourself on a sample of each corpus used in this work, check out this app:

wordpredictabilityvisualized.vercel.app

7 months ago 0 0 0 0
Preview
GroNLP/gpt2-small-dutch ยท Hugging Face Weโ€™re on a journey to advance and democratize artificial intelligence through open source and open science.

Modelling reading times in Dutch?:

gpt2-small-dutch (huggingface.co/GroNLP/gpt2-...) or gpt2-medium-dutch-embeddings (huggingface.co/GroNLP/gpt2-...) are great options.

7 months ago 0 0 1 0
Post image

3. Predictability effects are also logarithmic in Dutch, corroborating effects found in English (= linear effect of surprisal):

For very unpredictable words, a decrease in predictability has a much larger slowing-down effect on reading times than the same decrease for highly predictable words.

7 months ago 0 0 1 0
Post image

2. Language-specific models are generally better than multilingual ones (multilingual models are shown in blue in the figure below).

7 months ago 0 0 1 0
Advertisement
Post image

Key findings ๐Ÿ“

1. Smaller Dutch models often predict reading times better (= inverse scaling trend) ~ in line with evidence of English models.

But, with more context (in a book reading corpus), larger models catch up.

7 months ago 0 0 1 0

Large language models are powerful tools for psycholinguistic research.

But, most evidence so far is limited to English.

How well do Dutch open-source language models fit reading times using their word predictability estimates?

7 months ago 0 0 1 0
Preview
A systematic evaluation of Dutch large language modelsโ€™ surprisal estimates in sentence, paragraph and book reading - Behavior Research Methods Studies using computational estimates of word predictability from neural language models have garnered strong evidence in favour of surprisal theory. Upon encountering a word, readers experience a pro...

๐Ÿšจ B&B is open for business ๐Ÿšจ

Not a new career move, the next Boeve & Bogaerts paper is out in Behavior Research Methods!
doi.org/10.3758/s134...
@bogaertslab.bsky.social

7 months ago 2 0 1 1

โœจPlayback at #teap2025 (Part 1)
Thanks to the amazing speakers and the audience for the successful symposium on #languagemodels in #psycholinguistics!

Katharina Menn, @hannawoloszyn.bsky.social, @boevesam.bsky.social, Marco Marelli, Fritz Gรผnther, @benjamingagl.bsky.social

1 year ago 8 1 0 0
Post image

Proud PI! ๐Ÿ‘ On #TeaP2025 two lab members presented their work:

Haoyu Zhou in the symposium #StatisticalLearning and its Role in #Language and #Reading acquisition.

@boevesam.bsky.social in the symposium From Babies to Semantics: Leveraging #LanguageModels for #Psycholinguistic Research.

1 year ago 8 5 1 0
Preview
A Systematic Evaluation of Dutch Large Language Modelsโ€™ Surprisal Estimates in Sentence, Paragraph, and Book Reading A psychometric evaluation of Dutch large language models. Hosted on the Open Science Framework

Overall, our results provide a psychometric leaderboard of Dutch large language models, ideal for researchers interested in effects of predictability in Dutch.

Check out our full dataset and code here:
osf.io/wr4qf/

1 year ago 0 0 0 0

Finally, we found a linear link between surprisal and reading times except for the GECO corpus where a non-linear link between surprisal and reading times fitted the data best.

A challenge to the notion of an universal linear effect of surprisal.

1 year ago 0 0 1 0

Second, smaller Dutch models showed a better fit to reading times than the largest models, replicating the inverse scaling trend seen in English.
However, this effect varied depending on the corpus used.

1 year ago 0 0 1 0
Advertisement

First, across three eye-tracking corpora, we found that in each case, a Dutch LLMs' surprisal estimates outperformed the multilingual model (mGPT) and the N-gram model in predicting reading times.

1 year ago 0 0 1 0

3.

Does surprisal still show linear link with reading times when estimated with a Dutch-specific language model as opposed to a multilingual model?

1 year ago 1 0 1 0

2.

Do these Dutch-specific LLMs show a similar inverse scaling trend as English models?

That is, do the smaller transformer models' surprisal estimates account better for reading times than those of the very large models?

1 year ago 0 0 1 0

1.

What is the best computational method for estimating word predictability in Dutch?

We compare 14 Dutch large language models (LLMs), a multilingual model (mGPT) and an N-gram model in their ability of explaining reading times.

1 year ago 0 0 1 0

The effect of word predictability on reading times is well established for English but not so much for Dutch.

We adressed this and asked three questions:

1 year ago 0 0 1 0
Preview
A Systematic Evaluation of Dutch Large Language Modelsโ€™ Surprisal Estimates in Sentence, Paragraph, and Book Reading A psychometric evaluation of Dutch large language models. Hosted on the Open Science Framework

Ending the year on a high note with the submission of a new preprint:

A Systematic Evaluation of Dutch Large Language Modelsโ€™ Surprisal Estimates in Sentence, Paragraph, and Book Reading

Preprint: dx.doi.org/10.13140/RG....
OSF: osf.io/wr4qf/

1 year ago 3 0 1 0