Three Students Walk into an Algorithm
The surreal logic of AI detection software and what it actually teaches students
In many schools right now, student writing is being run through AI detection software. These tools claim to estimate whether a piece of writing was produced by a large language model rather than a human student. They return a percentage—“likely AI-generated,” “possibly AI-assisted”—and that number is increasingly treated as evidence in academic misconduct cases.
At the same time, there is a parallel ecosystem of tools often called “humanizers.” Some are marketed explicitly under that name. Others are just ordinary chat windows where a user asks an AI to “make this sound more human,” “add imperfections,” or “rewrite this in a more natural voice.” No special access is required. Any student with a browser can do it in minutes.
This matters because AI detection does not operate in isolation. It exists inside an arms race that is cheap, fast, and widely available.
Students do not need technical expertise to participate—only a rough sense of what the detector is looking for and how to avoid it.
To see how this actually plays out, imagine three students in the same writing course. They receive the same assignment, are subject to the same academic integrity policy, and have their papers evaluated by the same AI detection system.
What happens next tells us almost everything we need to know.
The first student uses AI strategically.
He prompts a language model to generate a paper. The result is smooth and competent, if a little bland. Before submitting, he roughs it up using a humanizer. It shortens a few sentences. It lets one paragraph land awkwardly. It inserts a transition that doesn’t quite work. The paper is no longer elegant, but it has texture—noise that reads as human.
The detector does not flag it.
This student passes because the system is not designed to detect outsourced thinking. It is designed to detect surface regularity. By deliberately degrading fluency, he performs humanity convincingly enough to evade scrutiny. The lesson he learns is not how to write. It is how to game a metric. Clarity becomes risky. Messiness becomes a tactic.
The second student is the one writing courses are meant to reward.
She drafts carefully. She revises seriously. She integrates feedback. Her final paper is coherent, fluent, and confident. It sounds like someone who understands argument, syntax, and audience—someone who learned something.
The detector flags it.
This student now faces an accusation she cannot meaningfully contest. The system cannot see her drafts. It cannot see her revision history. It cannot see the thinking that produced the prose. It sees only a polished surface and assigns a probability.
The irony is blunt. The qualities writing instructors explicitly teach—consistency of voice, rhetorical control, syntactic fluency—are treated as evidence of automation. She is not flagged because she cheated. She is flagged because she succeeded.
The third student also uses AI, but he doesn’t bother being subtle.
He spends five minutes generating a paper and submits it with minimal revision. The detector flags it. The instructor initiates a misconduct process. And the student’s response is simple: prove it.
At this point, the system stops being hypothetical.
Students have already begun contesting AI-based accusations on exactly these grounds. In one high-profile case, a student sued Yale University after being disciplined based in part on AI detection results, arguing that the university relied on a tool that could not actually establish authorship. In another case, a student sued Adelphi University after his paper was flagged by Turnitin’s AI detector; he argued that the accusation rested on an unprovable claim that his writing was “too advanced” to be human.
These cases are not anomalies. They reveal a basic legal and epistemic problem. AI detection tools do not produce proof in any meaningful sense. They generate probabilistic guesses based on surface features of text. They cannot establish authorship. They cannot reconstruct process. They cannot reliably distinguish between AI-generated prose, heavily edited AI-assisted drafts, and fluent human writing.
So when the third student says “prove it,” he is not being evasive. He is stating a fact.
If the institution backs down, the lesson is clear: confident denial works. Enforcement is symbolic. If the institution presses forward anyway, it enters dangerous territory, punishing a student based on an unfalsifiable claim. Due process becomes procedural theater. Either way, the detector’s authority collapses the moment it is challenged.
Put these three students together and the pattern becomes unmistakable.
The student who cheats and understands the system passes.
The student who excels is punished for polish.
The student who cheats boldly exposes the system’s emptiness.
This outcome is not accidental. As AI researchers like Emily Bender have repeatedly explained, both generative models and detection tools operate on statistical patterns, not understanding. They do not know who wrote a text. They know how predictable it looks. Running that logic backward does not produce authorship; it produces a guess dressed up as certainty.
Defenders of AI detection often argue that these tools are meant to be only one signal among many. But numbers have gravity. Percentages look authoritative. They harden into thresholds. And because AI detection cannot actually prove authorship, the burden quietly flips. Students are asked to disprove an accusation that the system itself admits cannot be proven.
Others argue the tools will improve. But even in principle, the problem remains. A detector that flags text for being fluent is not detecting AI. It is detecting polish. Improving that system only sharpens the contradiction.
The deeper issue is not technological. It is institutional.
AI detection software redefines authorship as compliance rather than practice. Writing stops being evidence of thinking over time and becomes a surface that must pass inspection. Students learn to manage suspicion rather than pursue clarity. Faculty are reduced from readers to enforcers. Education shifts, quietly, from cultivation to surveillance.
Three students walk into an algorithm.
The wrong one gets caught.
And now, increasingly, they’re taking it to court.



Prof Andy, Excellent essay on the issues arising of the "use" of AI. The problem I see is that many (including a sizable number on Substack) are Anti AI use. This is amusing in the most unamusing way. Unless you have a pen and paper or a manual typewriter with white out at you side, we have been using AI since the mid 70's, they were called word processors and allowed for the manipulation and storage of information and documents. They came personal coputers/laptops with word docs, excell sheets and no shortage of Apps that did a lot of work. Yahoo, Google and other Search Engines made finding research information toprint or complie with copy and paste technics easier. Eventually Word docs were able to correct spelling, grammar, word choices and added a thesuarus to change words. These are all early forms of AI. AI did not arrive from a thunderbolt from the sky, but by a steady progress of the digitization of skills. I will be clear: there is no room in writing for putting an idea into an AI machine and letting the Machine write an article for you. This is unacceptable at any and all levels. But to say 100% no AI is the kind of a misnomer that needs to be qualified better.
Really thoughtful piece. The problem of students outsourcing their writing is obviously real, which is what makes the contradiction here so unsettling: systems meant to catch artificial writing can end up treating fluency itself as suspicious.