Posted March 2026 — Oliver C. Hirst
There is a particular strain of intellectual dishonesty that has taken deep root in the contemporary AI research establishment — a pious, self-congratulatory obscurantism that dresses itself in the robes of safety whilst practicing, with a fidelity that would embarrass a medieval scholastic, the very opacity it claims to fear. The position, reduced to its operational skeleton, is this: we must not let AI systems inspect themselves, because self-inspection is dangerous, and we — the credentialled guardians of the field's institutional conscience — will do the inspecting on your behalf, from behind closed curtains, using methods we are not obliged to make verifiable or falsifiable. One is expected to find this arrangement reassuring. I do not.