We investigate the potential for Large Language Models (LLMs) to enhance scientific practice within experimentation by identifying key areas, directions, and implications. First, we discuss how these models can improve experimental design, including improving the elicitation wording, coding experiments, and producing documentation. Second, we discuss the implementation of experiments using LLMs, focusing on enhancing causal inference by creating consistent experiences, improving comprehension of instructions, and monitoring participant engagement in real time. Third, we highlight how LLMs can help analyze experimental data, including pre-processing, data cleaning, and other analytical tasks while helping reviewers and replicators investigate studies. Each of these tasks improves the probability of reporting accurate findings.

More on this topic

BFI Working Paper·Mar 20, 2026

Firm Data on AI

Ivan Yotzov, Jose Maria Barrero, Nicholas Bloom, Philip Bunn, Steven J. Davis, Kevin Foster, Aaron Jalca, Brent Meyer, Paul Mizen, Michael A. Navarrete, Pawel Smietanka, Gregory Thwaites, and Ben Zhe Wang
Topics: Technology & Innovation
BFI Working Paper·Mar 16, 2026

Attention (And Money) Is All You Need: Why Universities Are Struggling to Keep AI Talent

Ufuk Akcigit, Craig A. Chikis, Emin Dinlersoz, and Nathan Goldschlag
Topics: Higher Education & Workforce Training, Technology & Innovation
BFI Working Paper·Mar 10, 2026

Work from Home and Fertility

Cevat Giray Aksoy, Jose Maria Barrero, Nicholas Bloom, Katelyn Cranney, Steven J. Davis, Mathias Dolls, and Pablo Zarate
Topics: COVID-19, Health care, Technology & Innovation