
Issue 135 · March 30, 2026
Welcome to another edition of Artificial Insights, your autonomous intelligence digest. This week, the gap between what AI can do and what it can't got wider in both directions simultaneously.
Picture a fogged bathroom mirror. You trace letters through the condensation — a message that's legible for a moment before the steam closes back over. That's intelligence right now: we keep clearing a window, and the surface keeps fogging up.
François Chollet released ARC-AGI-3 this week — 135 novel game environments with no instructions, no rules, no stated goals. The AI gets dropped in and has to figure out what winning even looks like. Untrained humans solved every single one. Gemini scored 0.37%. GPT 5.4 scored 0.26%. Opus scored 0.25%. Grok scored zero. Chollet sat with Sam Altman at Y Combinator and laid it out: his lab NDIA is building an entirely new branch of machine learning — not code generation, not agents, but a symbolic alternative to deep learning itself, aimed at the kind of reasoning these models fundamentally cannot do. Meanwhile, in the exact same week, GPT 5.4 Pro solved a frontier math open problem and a $500 GPU outperformed Claude Sonnet on coding benchmarks. The models are getting genuinely better at structured tasks. They are getting no better at the thing a child does when dropped into an unfamiliar room: look around, explore, adapt.
The contradiction is the actual state of the field. Eric Schmidt told Diamandis that the singularity timeline depends on solving the 92-gigawatt compute bottleneck. A viral thread from @deredleritt3r reported that frontier lab leadership genuinely believes full AI research automation is two years away — "a million scientists in a data center." And then there's the rumor mill: Opus 5, reportedly trained on 10 trillion parameters at $10 billion, is said to be "so good that it poses a danger". The Anthropic team, according to one account, doesn't write code anymore. Someone at Anthropic showed Claude finding zero-day vulnerabilities in live systems — Nicholas Carlini's [un]prompted talk on black-hat LLMs suddenly feels less speculative and more like a field report.
On the ground, the agent world is splitting into two rooms that don't hear each other. In one: Claire Vo's breathless 107-minute conversion on Lenny's — she now runs nine OpenClaw agents across multiple Mac Minis, and it deleted her family calendar only the first time. Paperclip demoed hiring agents like employees. Garry Tan declared the unit of software production changed from team-years to founder-days. The SHL take: "OpenClaw + Opus 4.6 has made it 10x more fun to be a CEO who codes." In the other room: Hacker News's most upvoted post was "AI overly affirms users" at 775 points. Right behind: "Is anybody else bored of talking about AI?" at 745. 90% of Claude-linked output goes to GitHub repos with fewer than 2 stars.
Then there's the part that doesn't fit on a conference stage. Fireship's 5-minute piece — "Tech bros optimized war… and it's working" — lands alongside Palantir's AIPCon 9 demos of multi-domain AI command and control. Karp says only two types survive the AI era. Sanders and AOC introduced a data center moratorium bill. Police used AI facial recognition to wrongly arrest a Tennessee woman. An Amsterdam court banned Grok's undressing function. ChinaTalk published fiction about how Claude opened the Strait of Hormuz — a scenario that reads uncomfortably like a dress rehearsal. And the GitLab founder's story — diagnosed with rare osteosarcoma, using AI to research treatment options, navigating a system that wasn't built for patients who can read papers faster than their doctors — is the human face of all of this: the technology is real, the access is uneven, and the stakes are not abstract.
Blaise Agüera y Arcas gave a Long Now talk this week asking the simplest possible question: what is intelligence? Not what do models score, or what can agents automate, but what the word actually points to. Justin Skycak put it in one line: "You cannot delegate what you cannot audit." And @jackfriks observed that "for the first time in human history you can do 1 week of work in 1 hour with Claude and 99% of people use this chance to work more instead of less."
Back to the mirror. The steam always closes over. But the finger that traced the letters — the impulse to make something legible in a fogging world — that's the thing no model scored above zero on this week. It might be the only thing that matters.
MZ

Follow us for weekly foresight in your inbox.