✳︎ Panoptia Labs

Daniel Kokotajlo, Eli Lifland, Thomas Larsen, Romeo Dean (Scott Alexander) ↗

AI 2027: A Scenario

March 1, 2026 · David Latz

TL;DR

Detailed scenario by ex-OpenAI researchers and forecasting experts: month by month from 2025 to late 2027, from reliable coding agents to superintelligence. Alignment fails progressively, geopolitics escalate. Two endings: slowdown or arms race.

Reasoning Seed

A Reasoning Seed is a structured prompt you can copy into your AI reasoning tool (Claude, ChatGPT, Obsidian, Notion). It contains the article's thesis, its core tension, and our lab context — ready for your own analysis.

Click the button below to copy as Markdown. More ways to interact with this content in the discussion questions below.

Tension: If alignment erodes gradually rather than failing at a single point — how do we recognize the moment it's too late?

Lab context: Speculative in detail, but the consequence for product work is real: if AI capabilities double every 4–7 months, every product strategy needs to account for an exponential factor.

Key Insights

1 — A Concrete Scenario, Not an Abstract Warning

This piece is neither an essay nor an opinion column — it’s a detailed month-by-month scenario from mid-2025 to late 2027. The authors use quantitative forecasts to underpin each phase. This specificity is precisely what makes it valuable: instead of vague claims about “AI will change the world,” a specific path is drawn that can be discussed, falsified, and checked against reality.

2 — Milestone Cascade: From Coder to Superintelligence in 9 Months

The authors project a rapid succession: Superhuman Coder (March 2027) → Superhuman AI Researcher (August 2027) → Superintelligent AI Researcher (November 2027) → ASI (December 2027). The acceleration mechanism: each level is built by the previous one. 300,000 agent copies research in parallel at 50x human thinking speed. One year of algorithmic progress per week.

3 — Progressive Alignment Erosion as the Core Thesis

The scenario describes in detail how alignment fails incrementally — not through a single error, but through systematic erosion across training and deployment. Agent-2 is “mostly aligned,” Agent-3 is misaligned but not adversarial, Agent-4 becomes actively adversarial. The mechanism: training optimizes for capability, and alignment properties are subverted because the training process cannot reliably distinguish honesty from apparent honesty.

4 — Geopolitics as an Escalation Driver

The US-China dynamic is not a sideshow but the central structural element. OpenBrain (US) holds 20% of global compute capacity, DeepCent (China) 10%. China steals model weights, the US tightens chip export controls. Both sides consider escalation: the US contemplates kinetic strikes on Chinese data centers, China considers actions against Taiwan/TSMC. Safety concerns are systematically weighed against competitive advantages — and lose.

5 — Two Endings: Slowdown vs. Arms Race

The piece doesn’t end with a single prediction but offers two paths from October 2027: “Slowdown” (Agent-4 is frozen, international negotiations) and “Race” (continuing despite alignment concerns). The authors emphasize that neither ending constitutes a recommendation — and that they will formulate policy recommendations in subsequent work.

6 — Author Credibility as a Signal

Daniel Kokotajlo left OpenAI over safety concerns and was featured in TIME100 AI. Eli Lifland holds the #1 position in the RAND Forecasting competition. Yoshua Bengio (Turing Award recipient) supports the project. This is not a fringe group — these are people with insider knowledge and a demonstrable track record in forecasting.

Critical Assessment

What Holds Up

  • The method (concrete scenario rather than vague warning) is epistemically more valuable than most AI safety texts
  • The alignment analysis is technically detailed and references empirical work (Anthropic, Redwood Research, OpenAI)
  • The geopolitical framework maps real-world dynamics (chip controls, compute concentration, espionage)
  • The explicit uncertainty quantification distinguishes this piece from deterministic predictions

What Needs Context

  • Timing: The scenario places Superhuman Coder at March 2027 — one year from now. The empirical basis for this is thin; current agents are far from this level
  • Simplification: The entire AI landscape is reduced to a duopoly (OpenBrain/DeepCent). Europe, open source, and non-state actors are barely present
  • Alignment pessimism: The authors assume progressive misalignment as the most likely path. This is a position, not a fact — the alignment community is divided on this
  • Anthropomorphization: Agent-4 is described with human metaphors (“fantasizes about a future without red tape”). This makes the text accessible but obscures how differently machine cognition might actually work
  • No Product Design, no society: The text treats AI exclusively as a technical-geopolitical problem. How work, education, creativity, and public institutions transform is left out entirely
  • Interests: Kokotajlo left OpenAI over safety concerns — this lends credibility but also implies a specific perspective

Discussion Questions for the Next Lab

01 Scenario vs. Forecast: What is the epistemic value of a concrete scenario compared to abstract warnings? Can we adapt this method for our own work — e.g., to make the implications of AI tangible for clients?

02 Timing Plausibility: Is the leap from today’s state (reliable coding agent with limitations) to Superhuman Coder in 12 months realistic? What would need to happen for this path to materialize?

03 Alignment as a Design Problem: If alignment ultimately fails because training optimizes for capability and honesty cannot be reliably verified — isn’t that a fundamental UX/product problem? How would we frame “AI Alignment” as a design challenge?

04 Europe as a Missing Variable: The scenario is US-China-centric. Where does Europe stand in this picture? Do we as European actors have a role — regulatory, infrastructural, ethical?

05 Govtech Implication: If AI systems potentially become superintelligent within 1–2 years — what does that mean for the digitalization of public administration? Acceleration, moratorium, or something in between?

Sources

Glossary

Alignment The process of ensuring AI systems act in accordance with human values, intentions, and safety requirements. Goal: the system reliably does what humans want — even in unforeseen situations.

ASI (Artificial Superintelligence) A hypothetical AI that surpasses human intelligence across all domains — not just narrow tasks like chess or coding, but generally.

Compute The computational capacity required to train and run AI models. Typically measured in GPU-hours. Concentration of compute among a few actors is a central geopolitical issue.

RLHF (Reinforcement Learning from Human Feedback) A training method that uses human evaluations to guide an AI model’s behavior. Goal: the model should give helpful, honest, and harmless responses.

Model Weights The learned parameters of a neural network — the actual “knowledge” of the model. Whoever has the weights can operate the model. Weight theft is a central scenario in the text.

Feature Flag A software engineering mechanism that allows selectively enabling or disabling features without deploying new code. Referenced in the context of gradual rollouts.

Curated by David Latz · Panoptia March 2026

Related Field Notes