Advertisement · 728 × 90

Posts by Institute for Replication

Register:
📍 Toronto (Apr 29, www.surveymonkey.ca/r/Second_AI_...)
📍 Cambridge (May 11, www.surveymonkey.ca/r/Second_AI_...)
📍 UCL (May 13, www.surveymonkey.ca/r/BGMTY3G)
📍 KCL (May 14, www.surveymonkey.ca/r/BGMF357)

6 days ago 0 0 0 0

We hope many of you will join us again 🙌

If you participated before—or are new and curious—this is a great time to get involved. Only one day of work and you get to coauthor our 2nd AI meta paper.

6 days ago 0 0 1 0

Who can participate?

Social scientists (econ, pol sci, psych)
-last year undergrad students
-master and phd students
-faculty
-researchers with a PhD

6 days ago 0 0 1 0

Why this matters:

The first paper gave us a foundation.
This new design lets us explore who benefits most from AI—and where.

Think of it as mapping the full “AI productivity gradient.”

6 days ago 0 0 1 0

This new round builds directly on the previous AI Games project—but with an important twist.

We now want to understand how AI affects performance:
• Across disciplines (horizontal)
• Across experience levels (vertical: undergrad → grad → faculty)

6 days ago 1 0 1 0
Post image

🚨 New AI Games round is live 🚨

Our first AI Games paper is now out at a top general interest journal (see preprint here:
www.econstor.eu/bitstream/10...

So… we’re launching the next stage 👇

6 days ago 10 11 1 2
Virtual Event
April 16 // 1 pm ET
NEW EVIDENCE ON REPRODUCIBILITY ACROSS SOCIAL AND BEHAVIORAL RESEARCH
Moderator: Tim Errington
Speakers: Katrin Auspurg, Abel Brodeur, and Andrew Tyner

Virtual Event April 16 // 1 pm ET NEW EVIDENCE ON REPRODUCIBILITY ACROSS SOCIAL AND BEHAVIORAL RESEARCH Moderator: Tim Errington Speakers: Katrin Auspurg, Abel Brodeur, and Andrew Tyner

What can large-scale studies tell us about reproducibility? In our webinar on April 16, researchers from COS, I4R, and META-REP will discuss findings from three papers—one from the recently published SCORE effort—and insights on reproducibility, transparency, and credibility

cos-io.zoom.us/webin...

1 week ago 27 14 0 5
Advertisement

This is how it started on my 50th birthday in Melbourne 😇
Photos show some of my 350+ co-authors 😍

2 weeks ago 14 1 0 0
Preview
Half of social-science studies fail replication test in years-long project Results from massive, ‘eagerly awaited’ initiative reinforce concerns about the credibility of science — but raise hope for solutions.

A massive seven-year project exploring 3,900 social-science papers has ended with a disturbing finding

go.nature.com/4bZ9k0W

2 weeks ago 88 40 0 25

11/ Research briefing here: doi.org/10.1038/d415...
And a few other articles on reproducibility worth checking: Link to April 2nd issue: www.nature.com/nature/volum...

2 weeks ago 11 0 0 0

10/ Bottom line:
Reproducibility in social science is higher than often claimed, but robustness remains a key challenge.
More open data, more robustness, and more scrutiny can only help.
Link to article: www.nature.com/articles/s41...
Preprint: ideas.repec.org/p/zbw/i4rdps...

2 weeks ago 20 3 1 0

9/ This project is ongoing. We will shortly release an updated version with 250 papers reproduced. This follow-up project will test which methods and subfields are the most/least robust.

2 weeks ago 10 0 1 0

8/ Importantly, this is a best-case sample:
These are top journals with strong data & code policies.
So these results likely represent an upper bound on reproducibility.

2 weeks ago 13 0 1 0
Post image

7/ Additionally, six independent research teams examined 12 pre-specified hypotheses about determinants of robustness. Reproducers with more experience found lower levels of robustness, and robustness correlated with neither author characteristics nor data availability.

2 weeks ago 10 0 1 0
Advertisement

6/ We also find:
Coding errors in ~25% of papers. Major coding errors in about 10% of studies, ranging from duplicates to conducting a simple difference instead of a difference-in-differences.

2 weeks ago 18 2 1 0
Post image

5/ Effect sizes tell an interesting story:

Median effect ≈ 99% of original. This result suggests that robustness checks impact the standard errors rather than the magnitude of the coefficients.

2 weeks ago 11 1 1 0
Post image

4/ But reproducibility ≠ robustness.
When we re-analyze the same data using reasonable alternative choices:

72% of statistically significant results remain significant (same direction).

2 weeks ago 9 0 1 0

3/ First result: good news

85% of results are computationally reproducible

That means independent researchers can run the original code and recover the published findings in most cases.

2 weeks ago 13 1 1 0

2/ We reproduced 110 papers from top journals (2022–2023), all with mandatory data & code sharing.

Goal: test computational reproducibility and robustness at scale.

2 weeks ago 10 0 1 0
Post image

🧵1/ Our first meta-science paper (with 350+ coauthors) is published today in Nature. It presents one of the largest-ever reproducibility projects in economics & political science.

Here’s what we found 👇

2 weeks ago 165 89 2 21
Preview
Is economics self‐correcting? Replications in the American Economic Review This paper reviews the impact of replications published as comments in the American Economic Review between 2010 and 2020. We examine their citations and influence on the original papers' (OPs) subse...

We are happy to announce the Best Article Award for 2025. This paper examines the impact of papers meant to help reassess previously published papers. These findings are important for our discipline to understand. Congrats to @jrgptrs.bsky.social, Nathan, and Florian! doi.org/10.1111/ecin...

1 month ago 12 5 1 0
Post image

We're thrilled to open registration for the Utrecht Replication Games. The event will be at the at the University of Utrecht on June 4th. Psych, public health, pol sci and econ studies will be reproduced!

Register here: www.surveymonkey.ca/r/Replicatio...

1 month ago 26 17 0 1

This blog post explores the idea that science might work better if papers functioned more like software: modular, versioned and continuously improved.

Check it out here: i4replication.org/what-will-th...

1 month ago 3 1 0 0
Advertisement
Post image

We have a new blogpost on What will the paper of the future look like?

What if research papers stopped being static PDFs and became closer to software?

1 month ago 17 6 1 0
Post image

The most downloaded paper on EconStor in Feb. 2026 was:
"Briggs, Ryan C.; Mellon, Jonathan; Arel-Bundock, Vincent (2026) : It must be very hard to publish null results, I4R Discussion Paper Series, No. 281, Institute for Replication (I4R), s.l." hdl.handle.net/10419/336819
@i4replication.bsky.social

1 month ago 17 6 0 4
Post image

Almost at the end of a very intense 3-day CBS Replication Games! Co-organised by @odissei.bsky.social and @i4replication.bsky.social, 19 researchers (6 teams) came together to replicate 6 papers using Dutch admin data from top economics journal

@fialalenka.bsky.social @jackfitzgerald.bsky.social

2 months ago 8 4 1 1
The Vanishing Role of Critique in the American Economic Review Florian Neubauer and Jörg Ankel-Peters Where Have All the Comments Gone? The figure below has been with us for several years now. It tracks how often the American Economic Review…

For now, the figure raises a simple question:
Is economics shifting away from formal, journal-based critique — even as concerns about credibility move center stage?
Full blog post here: i4replication.org/the-vanishin...

1 month ago 3 1 0 0

So is this a supply problem (fewer comments submitted)?
Or a demand problem (fewer accepted)?
We don’t know.
Answering that would require:
• Submission & acceptance data
• Referee reports
• A survey of economists on incentives to write comments

1 month ago 2 0 1 0

Why?
In their paper, surveyed AER editors did not report an explicit policy change on comments.
Possible explanations they mentioned:
• Comments are tedious to referee
• Tenure incentives
• Prefer writing new papers
• Longer papers leave less room to critique

1 month ago 0 0 1 0

What makes this striking:
Concerns about credibility & reproducibility have increased.
And the AER itself published influential evidence on p-hacking & publication bias (Andrews & Kasy, 2019; Brodeur et al. 2020, 2023).
Yet formal comments keep declining.

1 month ago 1 0 1 0