Using Generative AI as Early Peer Review

4–6 minutes

Cheap Adversaries, Outsourced Ego, and Engineered Critique ← ChatGPT is obsessed with subtitles.

There is a peculiar anxiety around admitting that one uses generative AI in serious intellectual work. The anxiety usually takes one of two forms. Either the AI is accused of replacing thinking, or it is accused of flattering the thinker into delusion. Both charges miss the point, and both underestimate how brittle early-stage human peer review often is.

What follows is not a defence of AI as an oracle, nor a claim that it produces insight on its own. It is an account of how generative models can be used – deliberately, adversarially, and with constraints – as a form of early peer pressure. Not peer review in the formal sense, but a rehearsal space where ideas are misread, overstated, deflated, and occasionally rescued from themselves.

Audio: NotebookLM summary podcast of this topic.

The unromantic workflow

The method itself is intentionally dull:

  1. Draft a thesis statement.
    Rinse & repeat.
  2. Draft an abstract.
    Rinse & repeat.
  3. Construct an annotated outline.
    Rinse & repeat.
  4. Only then begin drafting prose.

At each stage, the goal is not encouragement or expansion but pressure. The questions I ask are things like:

  • Is this already well-trodden ground?
  • Is this just X with different vocabulary?
  • What objection would kill this quickly?
  • What would a sceptical reviewer object to first?

The key is timing. This pressure is applied before the idea is polished enough to be defended. The aim is not confidence-building; it is early damage.

Image: NotebookLM infographic on this topic.

Why generative AI helps

In an ideal world, one would have immediate access to sharp colleagues willing to interrogate half-formed ideas. In practice, that ecology is rarely available on demand. Even when it is, early feedback from humans often comes bundled with politeness, status dynamics, disciplinary loyalty, or simple fatigue.

Generative models are always available, never bored, and indifferent to social cost. That doesn’t make them right. It makes them cheap adversaries. And at this stage, adversaries are more useful than allies.

Flattery is a bias, not a sin

Large language models are biased toward cooperation. Left unchecked, they will praise mediocre ideas and expand bad ones into impressive nonsense. This is not a moral failure. It is a structural bias.

The response is not to complain about flattery, but to engineer against it.

Sidebar: A concrete failure mode

I recently tested a thesis on Mistral about object permanence. After three exchanges, the model had escalated a narrow claim into an overarching framework, complete with invented subcategories and false precision. The prose was confident. The structure was impressive. The argument was unrecognisable.

This is the Dunning-Kruger risk in practice. The model produced something internally coherent that I lacked the domain expertise to properly evaluate. Coherence felt like correctness.

The countermeasure was using a second model, which immediately flagged the overreach. Disagreement between models is often more informative than agreement.

Three tactics matter here.

1. Role constraint
Models respond strongly to role specification. Asking explicitly for critique, objections, boundary-setting, and likely reviewer resistance produces materially different output than asking for ‘thoughts’ or ‘feedback’.

2. Third-person framing
First-person presentation cues collaboration. Third-person presentation cues evaluation.

Compare:

  • Here’s my thesis; what do you think?
  • Here is a draft thesis someone is considering. Please evaluate its strengths, weaknesses, and likely objections.

The difference is stark. The first invites repair and encouragement. The second licenses dismissal. This is not trickery; it is context engineering.

3. Multiple models, in parallel
Different models have different failure modes. One flatters. Another nitpicks. A third accuses the work of reinventing the wheel. Their disagreement is the point. Where they converge, caution is warranted. Where they diverge, something interesting is happening.

‘Claude says…’: outsourcing the ego

One tactic emerged almost accidentally and turned out to be the most useful of all.

Rather than responding directly to feedback, I often relay it as:

“Claude says this…”

The conversation then shifts from defending an idea to assessing a reading of it. This does two things at once:

  • It removes personal defensiveness. No one feels obliged to be kind to Claude.
  • It invites second-order critique. People are often better at evaluating a critique than generating one from scratch.

This mirrors how academic peer review actually functions:

  • Reviewer 2 thinks you’re doing X.
  • That seems like a misreading.
  • This objection bites; that one doesn’t.

The difference is temporal. I am doing this before the draft hardens and before identity becomes entangled with the argument.

Guardrails against self-delusion

There is a genuine Dunning–Kruger risk when working outside one’s formal domain. Generative AI does not remove that risk. Used poorly, it can amplify it.

The countermeasure is not humility as a posture, but friction as a method:

  • multiple models,
  • adversarial prompting,
  • third-person evaluation,
  • critique of critiques,
  • and iterative narrowing before committing to form.

None of this guarantees correctness. It does something more modest and more important: it makes it harder to confuse internal coherence with external adequacy.

What this cannot do

It’s worth being explicit about the limits. Generative models cannot tell you whether a claim is true. They can tell you how it is likely to be read, misread, resisted, or dismissed. They cannot arbitrate significance. They cannot decide what risks are worth taking. They cannot replace judgment. Those decisions remain stubbornly human.

What AI can do – when used carefully – is surface pressure early, cheaply, and without social cost. It lets ideas announce their limits faster, while those limits are still negotiable.

A brief meta-note

For what it’s worth, Claude itself was asked to critique an earlier draft of this post. It suggested compressing the familiar arguments, foregrounding the ‘Claude says…’ tactic as the real contribution, and strengthening the ending by naming what the method cannot do.

That feedback improved the piece. Which is, rather conveniently, the point.

The Prison of Process

3–4 minutes

This is the proof copy of The Illusion of Light. I reviewed it, approved it, and signalled ‘good to go’. This is being printed and distributed through KDP. I’ve used them before. They’ve been reliable.

EDIT: On the upside, I’ve been notified that the hardback version is available, but it doesn’t appear to be available in France and Canada, two target regions. Hopefully, it becomes available outside of the U.S. soon.

Until now.

My approval triggered a workflow. I know workflows. I used to design them. I also know how dumb they can be.

KDP’s process flagged an error: the text on the spine might not be on the spine. ‘Might’. Theoretically. It could be offset, cut off, or printed on a fold. I understand their reasoning – high-speed printers, mechanical variance, and return risk. I also understand statistics, and a single observation doesn’t make a trend. But anyone with eyes can see at least a couple of millimetres of clearance at the top and bottom. This isn’t a case of ‘maybe’. It’s fine.

What fascinates me here is the ritual of compliance. Once a process is codified, it becomes self-justifying. The rule exists; therefore, it must be obeyed. There is no appeal to reason – only to the flowchart.

In the 1980s, when I was an audio engineer recording to two-inch magnetic tape, some of us liked to record hot, pushing the levels just past the recommended limits. You learned to ride the edge, to court distortion without collapse. That’s how I designed the spine text. Within tolerance. With headroom.

The problem is that modern systems don’t tolerate edges. There’s no “override” button for informed judgment. My remediation path is to shrink the type by half a point, resubmit, and pretend the machine was right.

What’s absurd is the timing. The same system that generated the proof approved this layout days ago. An automated OCR scan could have caught this phantom error earlier. Instead, the machine waits until the human signs off, then throws a flag so the process can justify its existence.

KDP is still faster and saner than IngramSpark. But this is capitalism distilled: survival by being marginally less incompetent than your competitor. Optimisation, not in the sense of best possible, but of barely better than worst acceptable.

The lesson, as always, is that processes begin as aids and end as prisons. The workflow, like the Enlightenment, believes itself rational. But the longer it runs, the less it serves the human at the console and the more it worships its own perfection.

Want to talk about meta? This underscores the contents of the book itself. What the Enlightenment once called Reason, modernity now calls Process. Both pretend to neutral objectivity while enshrining obedience as virtue. The bureaucracy of light has become digital – its catechism written in checkboxes, its priests replaced by automated validators. Every workflow promises fairness; each only codifies submission. The real danger isn’t that machines will replace judgment, but that we will stop noticing when they already have.


The Story Continues: Behind the Scenes

Image: Screenshot of Illustrator layout

I’ve reduced the font size on the spine from 14 points to 13.5. It still technically bleeds over a guideline. I hope I am not forced to reduce it to 13. A reason for text on the spine is to make it visible. Hopefully, the black-and-white vertical separation will help in this regard. Fingers crossed.