Advertisement · 728 × 90

Posts by Zoe Purcell 🕵🏻‍♀️

a man hanging from a scale where on one side there are really heavy stereotypes, and on the other side very light base-rates, to illustrate base-rate neglect

a man hanging from a scale where on one side there are really heavy stereotypes, and on the other side very light base-rates, to illustrate base-rate neglect

Are you thinking in vain? New paper with @zoepurcell.bsky.social, @luciecharlesneuro.bsky.social, @wimdeneys.bsky.social & @kobedesender.bsky.social

𝘛𝘩𝘪𝘯𝘬𝘪𝘯𝘨 𝘪𝘯 𝘝𝘢𝘪𝘯: 𝘈𝘯 𝘌𝘷𝘪𝘥𝘦𝘯𝘤𝘦 𝘈𝘤𝘤𝘶𝘮𝘶𝘭𝘢𝘵𝘪𝘰𝘯 𝘈𝘤𝘤𝘰𝘶𝘯𝘵 𝘰𝘧 𝘉𝘪𝘢𝘴𝘦𝘥 𝘙𝘦𝘢𝘴𝘰𝘯𝘪𝘯𝘨

🔗 osf.io/preprints/ps...

A 🧵

1 day ago 3 1 1 0
title of the paper in behavior research methods

title of the paper in behavior research methods

Our paper with @zoepurcell.bsky.social , @luciecharlesneuro.bsky.social and @wimdeneys.bsky.social on using LLMs to estimate belief strength in reasoning is out in Behavior Research Methods.

If you're interested in reasoning biases & LLMs as measurement tools, check it out: rdcu.be/eZXGK (free pdf)

3 months ago 8 1 1 1
Post image

Many people worry that AI-assisted writing feels “less authentic” and may erode trust.
So we tested it in two preregistered, incentivized trust-game experiments (N = 1,637), which just got published in iScience (Cell Press)
More details 👇

3 months ago 3 1 1 0

Our paper with @zoepurcell.bsky.social, @luciecharlesneuro.bsky.social, and @wimdeneys.bsky.social
has been accepted at Behavior Research Methods! 🥳

Here is the updated preprint: osf.io/preprints/ps...

Also, the baserater package is now on CRAN: cran.r-project.org/package=base...

#psynomBRM

4 months ago 11 3 0 0
Abstract

Accurately quantifying belief strength in heuristics-and-biases tasks is crucial yet methodologically challenging. In this paper, we introduce an automated method leveraging large language models (LLMs) to systematically measure and manipulate belief strength. We specifically tested this method in the widely used “lawyer-engineer” base-rate neglect task, in which stereotypical descriptions (e.g., someone enjoying mathematical puzzles) conflict with normative base-rate information (e.g., engineers represent a very small percentage of the sample). Using this approach, we created an open-access database containing over 100,000 unique items systematically varying in stereotype-driven belief strength. Validation studies demonstrate that our LLM-derived belief strength measure correlates strongly with human typicality ratings and robustly predicts human choices in a base-rate neglect task. Additionally, our method revealed substantial and previously unnoticed variability in stereotype-driven belief strength in popular base-rate items from existing research, underlining the need to control for this in future studies. We further highlight methodological improvements achievable by refining the LLM prompt, as well as ways to enhance cross-cultural validity. The database presented here serves as a powerful resource for researchers, facilitating rigorous, replicable, and theoretically precise experimental designs, as well as enabling advancements in cognitive and computational modeling of reasoning. To support its use, we provide the R package baserater, which allows researchers to access the database to apply or adapt the method to their own research.

Abstract Accurately quantifying belief strength in heuristics-and-biases tasks is crucial yet methodologically challenging. In this paper, we introduce an automated method leveraging large language models (LLMs) to systematically measure and manipulate belief strength. We specifically tested this method in the widely used “lawyer-engineer” base-rate neglect task, in which stereotypical descriptions (e.g., someone enjoying mathematical puzzles) conflict with normative base-rate information (e.g., engineers represent a very small percentage of the sample). Using this approach, we created an open-access database containing over 100,000 unique items systematically varying in stereotype-driven belief strength. Validation studies demonstrate that our LLM-derived belief strength measure correlates strongly with human typicality ratings and robustly predicts human choices in a base-rate neglect task. Additionally, our method revealed substantial and previously unnoticed variability in stereotype-driven belief strength in popular base-rate items from existing research, underlining the need to control for this in future studies. We further highlight methodological improvements achievable by refining the LLM prompt, as well as ways to enhance cross-cultural validity. The database presented here serves as a powerful resource for researchers, facilitating rigorous, replicable, and theoretically precise experimental designs, as well as enabling advancements in cognitive and computational modeling of reasoning. To support its use, we provide the R package baserater, which allows researchers to access the database to apply or adapt the method to their own research.

1/10

🚨 New preprint: Using Large Language Models to Estimate Belief Strength in Reasoning 🚨

When asked: "There are 995 politicians and 5 nurses. Person 'L' is kind. Is Person 'L' more likely to be a politician or a nurse?", most people will answer "nurse", neglecting the base-rate info.

A 🧵👇

6 months ago 13 3 1 1

Hahaha yeh I even remembered the damn thing to thank them in my talk and then ofc forgot to do it in the panic!

9 months ago 0 0 0 0
Post image

Great week at the CSBBCS/EPS conference! Speaking alongside two of my academic heroes and mentors was such an honour and a really special moment for me! Huge thanks to CSBBCS and Aimee Surprenant for the invitation and opportunity, and all those who attended our symposium 🙏 @exppsychsoc.bsky.social

9 months ago 10 1 1 0

w/ @lauracharbit.bsky.social, Grégoire Borst, @lapsyde.bsky.social and Anne-Marie Nussberger @mpib-berlin.bsky.social. Now in @cognitionjournal.bsky.social 🎉

10 months ago 0 0 0 0
Advertisement
Post image

Do AI builders hold different values from AI users?

We show that AI builders and men are more utilitarian and less supportive of pro-diversity outputs, highlighting ongoing concerns about workforce diversity and whose values are shaping AI.

tinyurl.com/AIcognit

10 months ago 8 6 1 0
Post image

New preprint: “Folk Thinking, Fast and Slow: Intuitive Preference for Deliberation in Humans and Machines”

Pop culture often praises intuition (“Blink”, Steve Jobs). But do we really trust it? Across 13 studies, we find a strong intuitive preference for deliberation.

tinyurl.com/8r54dmyn (1/6)

1 year ago 14 10 1 1

Come hear about moral psychology and AI at 11am today at #SPSP2025!

@awad.bsky.social @zoepurcell.bsky.social Anne-Marie Nussberger

1 year ago 8 5 0 1

New pre-print with @jfbonnefon.bsky.social !

Using Generative AI to Increase Skeptics’ Engagement with Climate Science

Available at: doi.org/10.31234/osf...

1 year ago 10 7 0 0
Postdoctoral Researcher (2011) - Birkbeck, University of London Birkbeck

📣🧠📣 I’m hiring !! 📣🧠📣

We’re looking for a postdoc to join The Uncertainty Lab at @birkbeckpsychology.bsky.social You’ll lead fMRI work on a new project studying how communication with others alters private metacognition of our own minds.

cis7.bbk.ac.uk/vacancy/post...

1 year ago 54 48 5 6

Let's give this a try ;-) go.bsky.app/TBS14hQ

1 year ago 7 2 3 0
Post image

📢Our paper on AI-mediated communication (out now!) shows that:
1) People expect others to use AI tools more than themselves
2) People are more accepting of open than secret AI use (at least, explicitly)
bit.ly/zpurc
👥 w/ Mengchen Dong, Anne-Marie Nussberger, Nils Kobis, & Maurice Jakesch.

1 year ago 6 2 1 0

Thanks for coming along!

1 year ago 1 0 0 0
Advertisement

We're currently running some follow up studies but should have something available this year :) I'll post it here once we do

1 year ago 2 0 1 0
Post image Post image

Today at IMBES my favourite talk was by @zoepurcell.bsky.social unpicking the relationships between partisan identity, prior beliefs, and reasoning. Partisan identity doesn’t make reasoning less accurate, but it does make the experience of reasoning subjectively more difficult. #academicsky

1 year ago 2 1 2 0
Post image Post image

Was thrilled to present our work at #ICT2024! I was overwhelmed by the support for this new angle and am now riding a new wave of motivation for this project -- thank you all✨ @wimdeneys.bsky.social @kobedesender.bsky.social

1 year ago 2 0 0 0

Thanks Nick! Was a pleasure to meet you in person :)

1 year ago 0 0 1 0