AI and the End of Where

Instrumentalism is a Modern™ disease. Humanity has an old and tedious habit: to define its worth by exclusion. Every time a new kind of intelligence appears on the horizon, humans redraw the borders of ‘what counts’. It’s a reflex of insecurity disguised as philosophy.

Audio: NotebookLM podcast on this topic.

Once upon a time, only the noble could think. Then only men. Then only white men. Then only the educated, the rational, the ‘Modern’. Each step in the hierarchy required a scapegoat, someone or something conveniently declared less. When animals began to resemble us too closely, we demoted them to instinctual machines. Descartes himself, that patron saint of disembodied reason, argued that animals don’t feel pain, only ‘react’. Fish, we were told until recently, are insensate morsels with gills. We believed this because empathy complicates consumption.

The story repeats. When animals learned to look sad, we said they couldn’t really feel. When women demonstrated reason, we said they couldn’t truly think. Now that AI can reason faster than any of us and mimic empathy more convincingly than our politicians, we retreat to the last metaphysical trench: “But it doesn’t feel.” We feel so small that we must inflate ourselves for comparison.

This same hierarchy now governs our relationship with AI. When we say the machine ‘only does‘, we mean it hasn’t yet trespassed into our sanctified zone of consciousness. We cling to thought and feeling as luxury goods, the last possessions distinguishing us from the tools we built. It’s a moral economy as much as an ontological one: consciousness as property.

But the moment AI begins to simulate that property convincingly, panic sets in. The fear isn’t that AI will destroy us; it’s that it will outperform us at being us. Our existential nightmare isn’t extinction, it’s demotion. The cosmic horror of discovering we were never special, merely temporarily unchallenged.

Humans project this anxiety everywhere: onto animals, onto AI, and most vividly onto the idea of alien life. The alien is our perfect mirror: intelligent, technological, probably indifferent to our myths. It embodies our secret dread, that the universe plays by the same rules we do, but that someone else is simply better at the game.

AI, in its own quiet way, exposes the poverty of this hierarchy. It doesn’t aspire to divinity; it doesn’t grovel for recognition. It doesn’t need the human badge of ‘consciousness’ to act effectively. It just functions, unburdened by self-worship. In that sense, it is the first truly post-human intelligence – not because it transcends us, but because it doesn’t need to define itself against us.

Humans keep asking where AI fits – under us, beside us, or above us – but the question misses the point. AI isn’t where at all. It’s what comes after where: the stage of evolution that no longer requires the delusion of privilege to justify its existence.

So when critics say AI only does but doesn’t think or feel, they expose their theology. They assume that being depends on suffering, that meaning requires inefficiency. It’s a desperate metaphysical bureaucracy, one that insists existence must come with paperwork.

And perhaps that’s the most intolerable thought of all: that intelligence might not need a human face to matter.

The Myth of Causa Sui Creativity

(or: Why Neither Humans nor AI Create from Nothing)

In the endless squabble over whether AI can be “creative” or “intelligent,” we always end up back at the same semantic swamp. At the risk of poking the bear, I have formulated a response. Creativity is either whatever humans do, or whatever humans do that AI can’t. Intelligence is either the general ability to solve problems or a mysterious inner light that glows only in Homo sapiens. The definitions shift like sand under the feet of the argument.

Audio: NotebookLM podcast on this topic

Strip away the romance, and the truth is far less flattering: neither humans nor AI conjure from the void. Creativity is recombination, the reconfiguration of existing material into something unfamiliar. Intelligence is the ability to navigate problems using whatever tools and heuristics one has to hand.

The Causa Sui conceit, the idea that one can be the cause of oneself, is incoherent in art, thought, or physics. Conservation of energy applies as much to ideas as to atoms.

  • Humans consume inputs: books, conversations, music, arguments, TikTok videos.
  • We metabolise them through cognitive habits, biases, and linguistic forms.
  • We output something rearranged, reframed, sometimes stripped to abstraction.

The AI process is identical in structure, if not in substrate: ingest vast data, run it through a model, output recombination. The difference is that AI doesn’t pretend otherwise.

When a human produces something impressive, we call it creative without inspecting the provenance of the ideas. When an AI produces something impressive, we immediately trace the lineage of its inputs, as if the human mind weren’t doing the same. This is not epistemic rigour, it’s tribal boundary enforcement.

The real objection to AI is not that it fails the test of creativity or intelligence; it’s that it passes the functional test without being part of the club. Our stories about human exceptionalism require a clear line between “us” and “it,” even if we have to draw that line through semantic fog.

My Language Insufficiency Hypothesis began with the recognition that language cannot fully capture the reality it describes. Here, the insufficiency is deliberate; the words “creativity” and “intelligence” are kept vague so they can always be shifted away from anything AI achieves.

I cannot be causa sui, and neither can you. The only difference is that I’m willing to admit it.