Categories
Nevin Manimala Statistics

Quantifying large language model usage in scientific papers

Nat Hum Behav. 2025 Aug 4. doi: 10.1038/s41562-025-02273-8. Online ahead of print.

ABSTRACT

Scientific publishing is the primary means of disseminating research findings. There has been speculation about how extensively large language models (LLMs) are being used in academic writing. Here we conduct a systematic analysis across 1,121,912 preprints and published papers from January 2020 to September 2024 on arXiv, bioRxiv and Nature portfolio journals, using a population-level framework based on word frequency shifts to estimate the prevalence of LLM-modified content over time. Our findings suggest a steady increase in LLM usage, with the largest and fastest growth estimated for computer science papers (up to 22%). By comparison, mathematics papers and the Nature portfolio showed lower evidence of LLM modification (up to 9%). LLM modification estimates were higher among papers from first authors who post preprints more frequently, papers in more crowded research areas and papers of shorter lengths. Our findings suggest that LLMs are being broadly used in scientific writing.

PMID:40760036 | DOI:10.1038/s41562-025-02273-8

By Nevin Manimala

Portfolio Website for Nevin Manimala