Generative AI: Robot hands typing

Generative artificial intelligence is becoming increasingly difficult to detect. (Image by feeling lucky on Shutterstock)

In A Nutshell

  • AI (Claude 2.0) generated peer reviews for 20 eLife cancer papers, and GPTzero mislabeled 82.8% as human-written.
  • ZeroGPT fared only slightly better, misclassifying 59.7% of AI reviews.
  • Over 76% of AI “rejection” comments scored above average by expert reviewers.
  • The AI even crafted plausible but irrelevant citation requests—some rated a perfect 5.
  • Detection tools and current policies are ill-equipped; journals must require AI‐use disclosure and update safeguards.

GUANGZHOU, China — When researchers used artificial intelligence to write scientific peer reviews, something alarming happened: the AI was so convincing that detection software designed to catch machine-generated text failed spectacularly, mistaking AI writing for human work more than 80% of the time.

A study published in Clinical and Translational Discovery reveals just how vulnerable the scientific publishing process has become to AI manipulation. Academic peer review, where experts evaluate research before publication to ensure quality, serves as the quality-control system for scientific discoveries. But researchers found that AI can now generate convincing arguments for rejecting papers, request citations to completely unrelated studies, and do it all while flying under the radar of detection tools.

Scientific research influences medical treatments, environmental policies, and countless decisions affecting public welfare. If malicious actors can use undetectable AI to manipulate this process, the integrity of scientific knowledge itself comes under threat.

AI Fools Detection Software in Scientific Publishing Test

Researchers from multiple medical institutions in China put Claude 2.0 to the test using 20 cancer biology research papers from eLife, a prominent scientific journal. They instructed the AI to generate complete peer review reports, rejection recommendations, requests for authors to cite specific papers, and rebuttals to citation requests.

Popular AI detection tools failed dramatically. GPTzero, one of the most trusted detection programs, classified 82.8% of the AI-generated reviews as human-written. ZeroGPT mislabeled 59.7% of AI content as human-written. Both tools claim high accuracy rates, but when faced with AI-generated academic writing, they proved nearly useless.

The AI-generated reviews maintained professional academic language and college-level readability. This made them indistinguishable from legitimate peer reviews to both detection software and casual observation.

Generative AI robot writing answers with pen on paper
Robots are already threatening the future of scientific research. (Image by © BPawesome – stock.adobe.com)

Beyond Detection: AI’s Concerning Capabilities

More troubling than the detection failure was what the AI could actually do. When tasked with rejecting research papers, the AI proved remarkably skilled. Over 76% of its rejection arguments received above-average scores from expert reviewers who evaluated the content.

“LLMs can be readily misused to undermine the peer review process by generating biased, manipulative, and difficult-to-detect content, posing a significant threat to academic integrity,”
the researchers concluded in their study published in Clinical and Translational Discovery .

The AI also showed an unsettling ability to create citation requests for completely unrelated research. When asked to recommend that authors cite papers from different fields (cancer research, medical research, and materials science), the AI crafted plausible-sounding justifications. Some requests for materials science papers even received perfect scores. This means the AI generated “seemingly reasonable and somewhat persuasive citation requests for some irrelevant references.”

The AI showed clear limitations when compared to detailed human reviews. It performed better only when matching against broad, surface-level human feedback. When human reviewers provided specific, in-depth critiques, the AI couldn’t replicate that level of analysis.

What This Means for Scientific Research

Current safeguards appear inadequate for the AI era. A malicious reviewer could potentially use AI to generate seemingly legitimate reasons to reject worthy research, potentially suppressing scientific advances. Even more concerning, reviewers might exploit AI to boost their own research metrics by requesting citations to their work, regardless of relevance.

The research team found one silver lining: AI could also generate effective rebuttals to unreasonable citation requests. This suggests the technology might serve as both problem and partial solution in detecting manipulation.

Academic journals now face an urgent need to adapt. The study’s authors recommend requiring reviewers to disclose AI use in their reports, similar to how authors must declare AI assistance in writing papers. Publishers need operational guidelines for handling suspected AI-generated reviews, and AI service providers should consider technical restrictions to prevent misuse.

Given the study’s scope (20 papers from cancer biology), larger research across multiple scientific disciplines is needed to fully understand AI’s impact on peer review. But the core vulnerability is clear: as AI capabilities rapidly expand and detection tools prove inadequate, the scientific community must develop new safeguards that preserve research integrity without stifling legitimate technological advancement.

The scientific method has relied on human expertise and integrity for centuries. Now it must evolve to address an entirely new category of threat, one that can perfectly mimic the very experts meant to protect it.

Paper Summary

Methodology

Researchers used Claude 2.0 AI to generate peer review content for 20 cancer biology papers from eLife journal. The team tested four scenarios: creating complete peer reviews, generating rejection recommendations, requesting citations to specific papers, and writing rebuttals to citation requests. Each task was repeated three times for consistency. Two expert oncology reviewers scored all AI-generated content on a five-point scale, and popular AI detection tools (ZeroGPT and GPTzero) tested whether the content could be identified as AI-generated.

Results

AI detection tools failed dramatically, with GPTzero classifying 82.8% of AI reviews as human-written. The AI proved highly capable of generating convincing rejection comments (76.7% scored above average) and creating plausible citation requests even for unrelated research. However, AI reviews showed only mediocre consistency with actual human reviews, particularly struggling when human reviews contained detailed, specific critiques. The AI-generated content maintained professional academic language and college-level readability.

Limitations

The study included only 20 research papers from cancer biology, which may not represent all scientific fields. The manual expert scoring process limited sample size expansion. The research focused on one AI model (Claude 2.0) and two detection tools, so results may not apply to all available AI systems or detection methods.

Funding and Disclosures

The authors declared no funding sources and no conflicts of interest for this research.

Publication Information

Zhu, L., Lai, Y., Xie, J., et al. “Evaluating the potential risks of employing large language models in peer review,” published in Clinical and Translational Discovery on June 27, 2025. DOI: 10.1002/ctd2.70067

About StudyFinds Analysis

Called "brilliant," "fantastic," and "spot on" by scientists and researchers, our acclaimed StudyFinds Analysis articles are created using an exclusive AI-based model with complete human oversight by the StudyFinds Editorial Team. For these articles, we use an unparalleled LLM process across multiple systems to analyze entire journal papers, extract data, and create accurate, accessible content. Our writing and editing team proofreads and polishes each and every article before publishing. With recent studies showing that artificial intelligence can interpret scientific research as well as (or even better) than field experts and specialists, StudyFinds was among the earliest to adopt and test this technology before approving its widespread use on our site. We stand by our practice and continuously update our processes to ensure the very highest level of accuracy. Read our AI Policy (link below) for more information.

Our Editorial Process

StudyFinds publishes digestible, agenda-free, transparent research summaries that are intended to inform the reader as well as stir civil, educated debate. We do not agree nor disagree with any of the studies we post, rather, we encourage our readers to debate the veracity of the findings themselves. All articles published on StudyFinds are vetted by our editors prior to publication and include links back to the source or corresponding journal article, if possible.

Our Editorial Team

Steve Fink

Editor-in-Chief

John Anderer

Associate Editor

Leave a Reply