Skip to main content
Planet Money

Don't hate the replicator, hate the game

36 min episode · 2 min read
·

Episode

36 min

Read time

2 min

AI-Generated Summary

Key Takeaways

  • P-Hacking Detection: Researchers manipulate datasets by running hundreds of analytical variations until results cross the 5% statistical significance threshold, a practice called p-hacking. Abel and colleagues identified this by scraping published journals and finding an abnormal clustering of results just above that threshold, published in their 2016 paper "Star Wars: The Empirics Strike Back."
  • Enforcement Probability Over Severity: Academic behavior changes not based on punishment severity but on the perceived likelihood of getting caught. Abel's strategy deliberately creates the impression that any paper could be audited at any time, mirroring how IRS audit rates influence taxpayer compliance even when actual enforcement remains relatively rare.
  • Replication Games Structure: Teams of three to five researchers receive a paper's full data and code package, then spend seven hours completing two phases: first, confirming the original code reproduces identical numerical results; second, running robustness checks by altering model specifications, excluded variables, or dataset subsets to test whether conclusions survive reasonable analytical alternatives.
  • Scale Through Crowdsourcing: Abel built a replication infrastructure essentially at zero cost by offering co-authorship credit on a published meta-paper as compensation. Recruiting researchers through social media and academic networks, he scaled from an unplanned 80-person Oslo event in October 2022 to over 50 games replicating approximately 300 papers within roughly two years.
  • Institutional Legitimacy as a Tool: To obtain data and code from reluctant researchers, Abel created the Institute for Replication with a professional website and recruited prominent economists to its board. This manufactured credibility generated email responses that informal individual requests had failed to produce, demonstrating that perceived institutional authority unlocks cooperation in academic settings.

What It Covers

University of Ottawa economist Abel Brodeur created the Replication Games, a hackathon-style event where teams of social scientists spend seven hours checking whether recently published academic papers hold up under scrutiny, addressing a two-decade replication crisis affecting psychology, medicine, and economics research worldwide.

Key Questions Answered

  • P-Hacking Detection: Researchers manipulate datasets by running hundreds of analytical variations until results cross the 5% statistical significance threshold, a practice called p-hacking. Abel and colleagues identified this by scraping published journals and finding an abnormal clustering of results just above that threshold, published in their 2016 paper "Star Wars: The Empirics Strike Back."
  • Enforcement Probability Over Severity: Academic behavior changes not based on punishment severity but on the perceived likelihood of getting caught. Abel's strategy deliberately creates the impression that any paper could be audited at any time, mirroring how IRS audit rates influence taxpayer compliance even when actual enforcement remains relatively rare.
  • Replication Games Structure: Teams of three to five researchers receive a paper's full data and code package, then spend seven hours completing two phases: first, confirming the original code reproduces identical numerical results; second, running robustness checks by altering model specifications, excluded variables, or dataset subsets to test whether conclusions survive reasonable analytical alternatives.
  • Scale Through Crowdsourcing: Abel built a replication infrastructure essentially at zero cost by offering co-authorship credit on a published meta-paper as compensation. Recruiting researchers through social media and academic networks, he scaled from an unplanned 80-person Oslo event in October 2022 to over 50 games replicating approximately 300 papers within roughly two years.
  • Institutional Legitimacy as a Tool: To obtain data and code from reluctant researchers, Abel created the Institute for Replication with a professional website and recruited prominent economists to its board. This manufactured credibility generated email responses that informal individual requests had failed to produce, demonstrating that perceived institutional authority unlocks cooperation in academic settings.

Notable Moment

During the very first replication game in Oslo, the opening paper contained a dataset where 75% of subjects shared identical age, gender, and location — a massive copy-paste duplication error. Since the paper studied inequality, having nearly identical subjects statistically eliminated the variation driving its core findings entirely.

Know someone who'd find this useful?

You just read a 3-minute summary of a 33-minute episode.

Get Planet Money summarized like this every Monday — plus up to 2 more podcasts, free.

Pick Your Podcasts — Free

Keep Reading

More from Planet Money

We summarize every new episode. Want them in your inbox?

Similar Episodes

Related episodes from other podcasts

This podcast is featured in Best Finance Podcasts (2026) — ranked and reviewed with AI summaries.

You're clearly into Planet Money.

Every Monday, we deliver AI summaries of the latest episodes from Planet Money and 192+ other podcasts. Free for up to 3 shows.

Start My Monday Digest

No credit card · Unsubscribe anytime