A paper authored by AI has just passed peer review. The scientific community is not ready
The arrival of AI-generated research papers marks a turning point that could radically accelerate discovery, or drown it in automated mediocrity.
By Jacek Krywko edited by Eric Sullivan

AI can generate research infinitely faster than humans can read it, threatening to bury an already strained peer review system under a mountain of automated submissions.
Vince Talotta/Toronto Star via Getty Images
SScience has always relied on the curious human mind to formulate a hypothesis, design an experiment, analyze the results, and present the case to peers. Over the centuries we have built better tools such as electron microscopes, particle accelerators and supercomputers, but the heart of scientific discovery has remained stubbornly human. Today, for the first time, this loop begins with a new type of mind.
Until now, scientists have often benefited from the help of artificial intelligence to solve a predefined and precise task, such as protein folding, says Jeff Clune, professor of computer science at the University of British Columbia. “We say that AI must become the scientist,” he says.
In a recent Nature studyClune and his colleagues unveiled AI Scientist, an AI system that wrote a paper without human intervention and passed peer review for a workshop at the International Conference on Learning Representations 2025 (ICLR), a leading venue in the field of machine learning. The document was poor, according to Clune and other experts. But its existence marks a turning point that the scientific community is only beginning to grasp: AI has quickly moved from an aid to scientists to a try to be one.
On supporting science journalism
If you enjoy this article, please consider supporting our award-winning journalism by subscribe. By purchasing a subscription, you are helping to ensure the future of impactful stories about the discoveries and ideas shaping our world today.
The AI Scientist includes several modules. After researchers give him a general topic, he reviews the available literature and generates hypotheses. “We just give it a general direction like, ‘Come up with something interesting to study about how AI learns,’” Clune says. The system then evaluates and refines these ideas, filtering out those that are not new. From there, other modules plan and run experiments, analyze and plot the data, and finally write the paper. It even conducts its own internal peer review process to find flaws in its papers, Clune says. (The system builds on existing foundation models such as Anthropic’s Claude Sonnet or OpenAI’s GPT-4o; the team’s contribution is the pipeline that orchestrates these models).
To see if The AI Scientist’s results could meet human standards, the team submitted three papers generated by it to the I Can’t Believe It’s Not Better (ICBINB) workshop at ICLR 2025. One was accepted. (Conference organizers gave permission for AI-generated papers to be submitted, and all papers by AI scientists were removed from the conference after the review process.)
The team behind AI Scientist admits that the bar for this workshop was lower than that of a keynote conference publication. “Would a mediocre graduate student get one in three papers accepted at an institution that accepts 70 percent of papers? Of course!” says Jodi Schneider, associate professor of information sciences at the University of Wisconsin-Madison, who was not involved in Clune’s study.
The AI papers “are OK but not great,” Clune says. To him, some of the AI’s ideas seemed genuinely creative, but the system struggled to execute. “The logic, writing and thinking throughout the document did not fit together perfectly,” he notes. Other issues included crazy referencesduplicate figures and a lack of methodological rigor.
Overall, Clune and colleagues’ new study received a lukewarm reception. “The approach is agentic and without any real novelty,” says Maria Liakata, professor of natural language processing at Queen Mary University of London, who was not involved in the work.
There is, however, one metric in which the AI scientist vastly outperformed human researchers: He produced a formally passable paper on machine learning in 15 hours, at a cost Clune estimated at around $140. Compare that with the ability of a graduate student, who might take a full semester to write their first accepted workshop paper, according to Schneider.
As costs fall and production speeds increase, AI-authored papers introduce the scientific community to an immediate challenge. “AI-authored papers will likely make things worse,” warns Yanan Sui, associate professor at Tsinghua University in China and lead chair of the ICLR 2026 workshop.
To guard against this flooding, prominent sites have started setting limits. “There are strict rules for the main conference that do not allow purely AI-written papers to be submitted,” says Sui. The trade-off, for now, is forced transparency: authors using AI must clearly state how it was used. Sui admits, however, that journals and conferences tools are usually missing to reliably detect AI-generated contributions.
In the meantime, tools for writing these contributions independently have already begun to proliferate. Intology claimed his AI Zochi passed peer review for the main proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (although human researchers were involved in areas such as verifying results before submission and communicating with peer reviewers). Another group called Autoscience Institute declared that its AI system created papers that were accepted at ICLR workshops before the AI scientist.
“We’re not going to be able to take away the power to generate scientific papers about AI,” says Aaron Schein, a data scientist at the University of Chicago and one of the organizers of the ICBINB workshop. “This technology is only going to get better. I don’t think there’s anything that can be done about it.”
But what if one day AI-generated logs stopped being mediocre?
Clune sees the transition taking place in two phases. “In the very short term, you’re going to have a lot of trash and garbage, and peer review systems are going to have to deal with that,” he says. But eventually, he says, AI systems will soon be available.it’s better in science than human researchers. “I predict that AI Scientist actually marks the dawn of a new era of rapid scientific progress,” says Clune, imagining humans reduced to curators watching AI perform scientific wonders.
Liakata, however, believes that we humans still have something to do. “I believe the future is not one of completely autonomous scientific discovery, but rather advanced human-agent interaction where humans can examine and contribute to the process,” she says.






























