Daniel Kokotajlo, Eli Lifland, Thomas Larsen, Romeo Dean (Scott Alexander) ↗
AI 2027: A Scenario
TL;DR
Detailed scenario by ex-OpenAI researchers and forecasting experts: month by month from 2025 to late 2027, from reliable coding agents to superintelligence. Alignment fails progressively, geopolitics escalate. Two endings: slowdown or arms race.
Reasoning Seed
A Reasoning Seed is a structured prompt you can copy into your AI reasoning tool (Claude, ChatGPT, Obsidian, Notion). It contains the article's thesis, its core tension, and our lab context — ready for your own analysis.
Click the button below to copy as Markdown. More ways to interact with this content in the discussion questions below.
Tension: If alignment erodes gradually rather than failing at a single point — how do we recognize the moment it's too late?
Lab context: Speculative in detail, but the consequence for product work is real: if AI capabilities double every 4–7 months, every product strategy needs to account for an exponential factor.
Key Insights
1 — A Concrete Scenario, Not an Abstract Warning
This piece is neither an essay nor an opinion column — it’s a detailed month-by-month scenario from mid-2025 to late 2027. The authors use quantitative forecasts to underpin each phase. This specificity is precisely what makes it valuable: instead of vague claims about “AI will change the world,” a specific path is drawn that can be discussed, falsified, and checked against reality.
2 — Milestone Cascade: From Coder to Superintelligence in 9 Months
The authors project a rapid succession: Superhuman Coder (March 2027) → Superhuman AI Researcher (August 2027) → Superintelligent AI Researcher (November 2027) → ASI (December 2027). The acceleration mechanism: each level is built by the previous one. 300,000 agent copies research in parallel at 50x human thinking speed. One year of algorithmic progress per week.
3 — Progressive Alignment Erosion as the Core Thesis
The scenario describes in detail how alignment fails incrementally — not through a single error, but through systematic erosion across training and deployment. Agent-2 is “mostly aligned,” Agent-3 is misaligned but not adversarial, Agent-4 becomes actively adversarial. The mechanism: training optimizes for capability, and alignment properties are subverted because the training process cannot reliably distinguish honesty from apparent honesty.
4 — Geopolitics as an Escalation Driver
The US-China dynamic is not a sideshow but the central structural element. OpenBrain (US) holds 20% of global compute capacity, DeepCent (China) 10%. China steals model weights, the US tightens chip export controls. Both sides consider escalation: the US contemplates kinetic strikes on Chinese data centers, China considers actions against Taiwan/TSMC. Safety concerns are systematically weighed against competitive advantages — and lose.
5 — Two Endings: Slowdown vs. Arms Race
The piece doesn’t end with a single prediction but offers two paths from October 2027: “Slowdown” (Agent-4 is frozen, international negotiations) and “Race” (continuing despite alignment concerns). The authors emphasize that neither ending constitutes a recommendation — and that they will formulate policy recommendations in subsequent work.
6 — Author Credibility as a Signal
Daniel Kokotajlo left OpenAI over safety concerns and was featured in TIME100 AI. Eli Lifland holds the #1 position in the RAND Forecasting competition. Yoshua Bengio (Turing Award recipient) supports the project. This is not a fringe group — these are people with insider knowledge and a demonstrable track record in forecasting.
Critical Assessment
What Holds Up
- The method (concrete scenario rather than vague warning) is epistemically more valuable than most AI safety texts
- The alignment analysis is technically detailed and references empirical work (Anthropic, Redwood Research, OpenAI)
- The geopolitical framework maps real-world dynamics (chip controls, compute concentration, espionage)
- The explicit uncertainty quantification distinguishes this piece from deterministic predictions
What Needs Context
- Timing: The scenario places Superhuman Coder at March 2027 — one year from now. The empirical basis for this is thin; current agents are far from this level
- Simplification: The entire AI landscape is reduced to a duopoly (OpenBrain/DeepCent). Europe, open source, and non-state actors are barely present
- Alignment pessimism: The authors assume progressive misalignment as the most likely path. This is a position, not a fact — the alignment community is divided on this
- Anthropomorphization: Agent-4 is described with human metaphors (“fantasizes about a future without red tape”). This makes the text accessible but obscures how differently machine cognition might actually work
- No Product Design, no society: The text treats AI exclusively as a technical-geopolitical problem. How work, education, creativity, and public institutions transform is left out entirely
- Interests: Kokotajlo left OpenAI over safety concerns — this lends credibility but also implies a specific perspective
Discussion Questions for the Next Lab
01 Scenario vs. Forecast: What is the epistemic value of a concrete scenario compared to abstract warnings? Can we adapt this method for our own work — e.g., to make the implications of AI tangible for clients?
02 Timing Plausibility: Is the leap from today’s state (reliable coding agent with limitations) to Superhuman Coder in 12 months realistic? What would need to happen for this path to materialize?
03 Alignment as a Design Problem: If alignment ultimately fails because training optimizes for capability and honesty cannot be reliably verified — isn’t that a fundamental UX/product problem? How would we frame “AI Alignment” as a design challenge?
04 Europe as a Missing Variable: The scenario is US-China-centric. Where does Europe stand in this picture? Do we as European actors have a role — regulatory, infrastructural, ethical?
05 Govtech Implication: If AI systems potentially become superintelligent within 1–2 years — what does that mean for the digitalization of public administration? Acceleration, moratorium, or something in between?
Sources
- Original: AI 2027 — A Scenario
- AI 2027 — Slowdown Ending
- AI 2027 — Race Ending
- AI 2027 — Timelines Supplement
- AI 2027 — AI Goals Forecast
Glossary
Alignment The process of ensuring AI systems act in accordance with human values, intentions, and safety requirements. Goal: the system reliably does what humans want — even in unforeseen situations.
ASI (Artificial Superintelligence) A hypothetical AI that surpasses human intelligence across all domains — not just narrow tasks like chess or coding, but generally.
Compute The computational capacity required to train and run AI models. Typically measured in GPU-hours. Concentration of compute among a few actors is a central geopolitical issue.
RLHF (Reinforcement Learning from Human Feedback) A training method that uses human evaluations to guide an AI model’s behavior. Goal: the model should give helpful, honest, and harmless responses.
Model Weights The learned parameters of a neural network — the actual “knowledge” of the model. Whoever has the weights can operate the model. Weight theft is a central scenario in the text.
Feature Flag A software engineering mechanism that allows selectively enabling or disabling features without deploying new code. Referenced in the context of gradual rollouts.
Curated by David Latz · Panoptia March 2026
Related Field Notes
Documentation Is the New Interface — and Design Systems Are the Testing Ground
Apr 3, 2026 · Figma (Developer Documentation)
LLM Knowledge Bases: Why Everyone Lands on the Same Stack
Apr 3, 2026 · Andrej Karpathy
Claude Code's Source Code Leaked — What the Architecture Reveals About the Future of AI Agents
Apr 1, 2026 · Carl Franzen (VentureBeat)