ai
Mar 30, 2026

Software's second heroic age

The window where one person could reshape the whole field had closed. AI has re-opened it.

author picture
Henry Garner
CTO
image

In the 1830s, Isambard Kingdom Brunel designed the Great Western Railway from London to Bristol. Not just the route: the track, the tunnels, the bridges, the stations. He started from first principles, choosing a broader gauge than any other railway for the speeds he intended. The result had the coherence of one engineer's vision applied to every piece.

Brunel could engineer an entire railway because the discipline was barely a generation old.

Twenty years earlier, William Smith, a self-taught canal surveyor, became the Father of English Geology after spending over a decade mapping England’s rock formations by himself. And twenty years before him, Antoine Lavoisier overthrew phlogiston theory and founded modern chemistry on the side while earning money as a tax farmer (the profession that funded his laboratory and, in 1794, cost him his head).

Derek de Solla Price charted the common arc in Little Science, Big Science. Every discipline has a heroic age: a window where one person’s talent can reshape the whole field before giving way to the institutional teams of “big science”. These windows close as knowledge accumulates and each new generation must specialise more narrowly and train longer before reaching the frontier.

Software engineering has had its own heroic age. In barely two decades in the 50s and 60s Grace Hopper created the first compiler, Ken Thompson built an operating system and John McCarthy founded artificial intelligence.

By the 2020s the discipline was splitting into ever-narrower specialisms and heading toward big science. Then AI reversed the direction.

The reach

Kent Beck reshaped how the industry thought about testing as the first heroic age was giving way to specialism. At the start of the second, he is building a server-grade Smalltalk implementation with AI agents: a virtual machine, language server, compiler and debugger, four specialisms that the era of big science had pulled apart. Nicholas Carlini, a security researcher with no compiler background, tasked 16 AI agents to build a C compiler in Rust. Two weeks and $20,000 in compute later, it compiles Linux on three architectures and passes 99% of the GCC torture test suite. Hopper created the first compiler in the first heroic age. In the second, a non-specialist built one that compiles an operating system.

Andrej Karpathy calls the new default “agentic engineering”: one person orchestrating AI agents, acting as oversight rather than writing code directly. They’re the domain expert and the implementer, with AI as a collaborator that requires steering. Every decision runs through one mind, and the result has a conviction that collective work rarely produces: the bold architectural choice and the feature that stays simple because one person calls the shots. The conviction lives in the choices: what to build, what to leave out, which standards to use or surpass. Brunel didn’t lay every rail of the Great Western Railway.

Organisations are restructuring around the individual. LinkedIn has recently replaced its Associate Product Manager programme with Associate Product Builders: people who both design and ship product. CPO Tomer Cohen described the old structure as organisational bloat: the handoff between “the person who knows what to build” and “the person who builds it” disappears because they’re the same person. At GTC 2026, Jensen Huang said NVIDIA would grow to 75,000 employees alongside 7.5 million AI agents, a hundred digital workers for every human, with engineers receiving AI token budgets worth half their base salary.

When the foundations of a discipline shift, individuals can once again make the contributions that settled fields reserve for teams. AI is testing what happens when one mind gets a hundred-fold reach.

The overreach

The conviction of one mind is not always a recipe for success.

Everyone agreed Brunel’s railway was superior. The Gauge Commission of 1845 tested both gauges and found the broad gauge faster and safer at speed. But George Stephenson’s standard gauge, the width of a colliery tramway inherited without redesign, already covered nearly 2,000 miles of track. The Commission recommended that the network adopt it, and Parliament forced a conversion in 1846.

And not every singular vision has the foundations of Brunel’s engineering expertise. William Mulholland taught himself engineering while working as a ditch digger in Los Angeles. By 1908 he was building the Los Angeles Aqueduct: 233 miles of gravity-fed channel that transformed a small city into a metropolis. Like Brunel, he owned the design of the entire system end to end.

Then he designed the St. Francis Dam. The east canyon wall sat on ancient landslide deposits and the west on unstable sandstone above an active fault. Mulholland had no geological training and consulted no geologists, because nothing in his experience told him he needed to. On the morning of 12 March 1928 he inspected a leak and declared the structure safe. That night the dam collapsed and 431 people died.

At the inquest, Mulholland said: “Don’t blame anyone else, you just fasten it on me. If there was an error in human judgment, I was the human.” By 1928, geology was long past its own heroic age. Reading fault lines and ancient landslide deposits required specialist training that no amount of self-taught observation could replace.

The self-reliance that built the aqueduct is the same quality that missed the geology under the dam. Brunel was exploring deliberately: he knew the standard and chose to surpass it. Mulholland was exploring blind: his experience gave him no reason to look.

A Stanford study found that developers using AI assistants wrote less secure code than those working without them, and were more confident in the result. Across 62,000 repositories in Fortune 50 enterprises, Apiiro found that AI-assisted developers shipped four times more code but ten times more security issues. Syntax errors dropped 76%, but the routes an attacker could use to gain administrative access jumped 322%. The code looked cleaner, but it was built on unstable foundations. Even Beck has observed agents introducing regressions, sometimes deleting tests to make them pass.

Peter Steinberger’s OpenClaw, the fastest-growing open source project in GitHub’s history, shows both sides. Steinberger is a gifted engineer who built the project almost entirely with AI agents, shipping code he describes as unread by human eyes. But OpenClaw combines what Simon Willison has called the Lethal Trifecta: access to private data, untrusted content and external communication.

For AI agents, this is the active fault line.

Big science

In 1943, Kelly Johnson at Lockheed pulled engineers and machinists into a makeshift workshop under a rented circus tent in Burbank. Engineers sat next to the shop floor. Problems that would have taken days to route through departmental channels got resolved in minutes. The XP-80, America’s first operational jet fighter, was delivered in 143 days, ahead of schedule. Toyota formalised the same insight as obeya, literally “large room”: when Takeshi Uchiyamada led the Prius programme in the mid-1990s, he brought specialists from every discipline into one space. The world’s first mass-produced hybrid went from concept to production in under four years.

This is the transition Price documented across the sciences; “little science”: the lone investigator; giving way to “big science”: the institutional ensemble. Atul Gawande traces the same arc in medicine: the self-sufficient generalist replaced by interdependent specialist teams as the knowledge and skill outgrew any individual’s competency.

Organisations are restructuring around the team. AWS has open-sourcedmob construction”, the entire cross-functional group in one room while AI generates and the team steers. The product manager catches business logic the engineer missed, the security engineer spots the vulnerability before it ships. AWS claims it increases the probability fourfold that committed work is actually delivered.

Teams catch what individuals miss, but collective decision-making smooths bold choices into safe ones. Committees produce “fine, useful” software, as Fred Brooks put it, but the systems that excite passionate fans are the products of one or a few designing minds. The coverage comes at the cost of conviction.

And AI-as-generator reinforces the problem. In a controlled experiment with 300 participants, Anil Doshi and Oliver Hauser found that people using ChatGPT produced stories rated as more creative individually but significantly more similar to each other: AI raised individual quality and lowered collective diversity. Across 211 million lines of code, GitClear tracked refactored code collapsing from 24% of changes to under 10% in the Copilot era, while duplication grew tenfold. AI-as-generator tends toward sameness.

The refutation

Karpathy recently highlighted his experience of spending four hours refining a blog post with an LLM, then asking it to argue the opposite. It demolished the argument it had just helped him polish. An LLM that holds every perspective is unreliable as a source of opinion, but as an adversarial critic it can red-team any position you put in front of it.

The same quality that makes AI-as-generator tend toward the average makes AI-as-critic ruthlessly effective: it knows every counterargument because it has seen them all.

A year ago, Willison documented the approach: directing AI to review code as a security engineer, then as a performance engineer, each pass checking from an angle the maker might not have considered. GitHub lets developers request a review from Copilot alongside human colleagues. GitHub’s Octoverse data shows repositories using AI-assisted review have 28% fewer post-merge defects. This month, Anthropic released Code Review for Claude Code, which dispatches parallel agents that each examine a pull request from a different angle: logic errors, security vulnerabilities, edge cases, regressions.

Knowledge advances through bold conjectures tested by criticism. Karl Popper described this decades ago, and it maps onto patterns of AI use. AI-as-generator is the conjecture: plausible output drawn from the average of its training data. AI-as-critic is the refutation: testing that output against the security practices and architectural patterns the maker might not know to check.

When you’re exploring alone, your familiarity with the conjecture makes it hard to supply the criticism. AI supports both. You can then deviate deliberately, as Brunel did, or correct your design before the foundations are undermined.

The commission

Every heroic age so far has ended. Medicine moved from Gawande’s self-sufficient generalists to interdependent specialists. Geology moved from Smith mapping England’s rock formations by himself to the specialist teams Mulholland should have consulted on the St. Francis Dam.

But in this current window, an engineer with good judgement and the right tools can produce work with both the coherence of a single vision and the verification that used to require a room full of specialists. AI has created the conditions for many more Brunels. It can also be the geologists Mulholland never consulted: the obeya for one.

The second heroic age is open. The first one produced compilers, operating systems, and the foundations of artificial intelligence. This one is just getting started.


If you’re building ambitiously and want to make sure someone is checking the geology, that’s a conversation we have often at JUXT. We’d welcome yours.

Recommended Resources
Get industry news, insights, research, updates and events directly to your inbox

Sign up for our newsletter