Skip to content

ICML Catches 497 Papers Cheating on AI Peer Review

6 min read

ICML Catches 497 Papers Cheating on AI Peer Review
Photo by J MAD on Pexels

The Conference That Trapped Its Own Reviewers

When the International Conference on Machine Learning published its review statistics on March 18, 2026, one number stood out: 497 desk rejections, all tied to a single policy violation. Reviewers who had pledged not to use language models had used them anyway — and ICML had proof. This wasn’t a blanket crackdown based on suspicion. It was a targeted operation using cryptographic-style watermarking, and it worked.

The rejections affect roughly 2% of all ICML 2026 submissions. Behind each rejected paper is a reciprocal reviewer — someone who agreed to review other researchers’ work in exchange for having their own paper considered. When that reviewer violated the terms, ICML’s policy was clear: the paper goes too. 398 reviewers triggered the rule; 497 papers paid for it.

How the Watermark Worked

ICML’s detection method is technically elegant. Before submission PDFs were distributed to reviewers, the organizers embedded invisible instructions into each document — text visible to a language model parsing the PDF, but not to a human reading it. Each paper received two unique trigger phrases, sampled from a dictionary of 170,000 options. The probability of any given phrase pair appearing by chance is less than one in ten billion.

The trap was simple: if a reviewer fed the PDF into an LLM and pasted the output as their review, the hidden instructions would nudge the model to include those two telltale phrases. Frontier models complied in over 80% of pre-deadline test runs. Every flagged review then received manual human verification before any action was taken — a step ICML says prevented false positives from skewing the results.

In total, 795 reviews from 506 unique reviewers were flagged — approximately 1% of all reviews submitted. Fifty-one reviewers had more than half their reviews flagged; for them, all reviews were deleted and they were removed from the reviewer pool entirely. The family-wise error rate — the statistical probability of even one false positive across all detections — was calculated at 0.0001.

Two Policies, One Choice

What makes the ICML case unusual is the explicit opt-in structure. For the first time, the conference offered reviewers two distinct tracks. Policy A reviewers committed to no LLM use at all. Policy B reviewers were allowed to use AI to help understand papers and polish their writing. Both groups knew the rules upfront and made an active choice.

The 398 violators had selected Policy A — the stricter option — and then ignored it. This is not a case of ambiguous rules or grey-area usage. These reviewers explicitly agreed to a prohibition, then violated it in ways detectable by the conference’s tooling. The enforcement mechanism was already in place before a single PDF was distributed.

The dual-policy framework itself represents a notable shift in how academic conferences handle AI. Rather than issuing a blanket ban (widely ignored) or offering blanket permission (potentially degrading review quality), ICML created a structured choice with accountability. It’s an experiment in governance as much as in technology.

Why This Matters Beyond ICML

AI use in peer review is not a new problem, but the scale is escalating. A 2025 study found that 21% of peer reviews across multiple AI conferences showed strong signals of AI generation. A separate publisher retracted 8,000 fraudulent articles in 2023, many linked to paper mills using automated tools. ICML’s intervention is significant not because the violations are surprising, but because the conference actually caught them — and enforced consequences.

The academic community has watched AI reshape research workflows over the past two years. Tools like Elicit, Semantic Scholar, and Consensus have made literature review dramatically faster — legitimately so. But the peer review process occupies a different role: it’s the quality gate that separates preprints from published science. When reviewers use LLMs to generate evaluations they then submit as their own considered judgment, the gate stops functioning.

This connects to a broader pattern documented at vortx.ch: AI can already write papers that pass peer review. If AI can write the paper and AI can write the review, the entire publication pipeline becomes an automated loop with no human quality signal left. ICML’s watermarking approach is, for now, one of the few enforcement tools that can break that loop.

What Happens to the 497 Papers

The desk rejections are not provisional. Papers submitted by reciprocal reviewers who violated Policy A are out of ICML 2026. Authors of affected papers — who may not have known their designated reviewer was using an LLM — bear the consequence regardless. ICML’s ethics policy, which all authors agreed to, states explicitly that a submitted paper bears responsibility for its reciprocal reviewer complying with conference policy.

For papers that had already received a full set of legitimate reviews before the violations were detected, the situation is more complex. Those papers are also desk rejected, even if the remaining reviews were valid. The conference has indicated that area chairs may need to source additional reviewers for submissions that lost valid reviews due to the cleanup.

The broader question — what to do about reviews that were influenced by LLMs without being fully AI-generated — remains unresolved. The watermarking method is designed to catch the clearest case: reviewer feeds PDF to LLM, pastes output. It won’t catch a reviewer who reads an LLM’s summary and then writes their own analysis. That’s a harder problem, and one ICML hasn’t claimed to solve.

The Enforcement Gap in Academic AI Governance

ICML’s watermarking experiment reveals something important: the main reason AI policy violations in peer review go uncaught is not that detection is impossible, but that conferences rarely invest in detection infrastructure. Building a dictionary of 170,000 phrases and embedding two-phrase signatures in each PDF is not technically complex. It requires planning and commitment, not exotic technology.

The harder institutional question is whether other conferences will follow. NeurIPS, ICLR, CVPR, and ACL each process tens of thousands of reviews per cycle. Implementing watermarking at that scale is feasible; deciding to do so requires organizational consensus on what AI-assisted reviewing means for the field. ICML has now forced that conversation with real data: 497 rejections, 398 violators, one conference willing to act on its own rules.

For researchers, the lesson is direct. Conferences that implement detection will catch violations. The assumption that LLM use in reviewing is undetectable — which many researchers appear to have been operating under — is no longer safe. And for the broader research community, including the AI labs whose models are cited in every second paper at these venues, the integrity of peer review is not a secondary concern. It’s the foundation that makes benchmark results and published findings worth citing at all. See also: how AI tools are reshaping legitimate academic research workflows.

Further Reading

Don’t miss on Ai tips!

We don’t spam! We are not selling your data. Read our privacy policy for more info.

Don’t miss on Ai tips!

We don’t spam! We are not selling your data. Read our privacy policy for more info.

Enjoyed this? Get one AI insight per day.

Join engineers and decision-makers who start their morning with vortx.ch. No fluff, no hype — just what matters in AI.