Esc
Original: Cat Wu (Anthropic) · ·

Product Management on the AI Exponential

TLDR

Anthropic's Head of Product for Claude Code describes how exponentially improving models break the traditional PM playbook — and the four shifts teams need to stay on the curve instead of behind it.

Reasoning Seed

Ein Reasoning Seed ist ein strukturierter Prompt, den du in dein KI-Reasoning-Tool kopieren kannst (Claude, ChatGPT, Obsidian, Notion). Er enthält die These des Artikels und die zentrale Spannung — bereit für deine eigene Analyse.

A Reasoning Seed is a structured prompt you can copy into your AI reasoning tool (Claude, ChatGPT, Obsidian, Notion). It contains the article's thesis and central tension — ready for your own analysis.

Spannung

If technical feasibility changes faster than the planning cycle — is product management still planning, or just reacting?

These

Anthropic's Head of Product for Claude Code describes how exponentially improving models break the traditional PM playbook — and the four shifts teams need to stay on the curve instead of behind it.

· Auf LinkedIn diskutieren Discuss on LinkedIn

Key Insights

1 — The PM Playbook Rests on an Assumption That No Longer Holds

Traditional product management assumes that what’s technologically possible at the start of a project is roughly what’s possible at the end. With exponentially improving models, this assumption breaks down: features that were impossible at sprint start become feasible mid-sprint. Wu illustrates this with her own experience — Claude Code failed at simple tasks with Sonnet 3.5, worked occasionally with Opus 4, and delivered reliable live demos with Opus 4.6.

2 — Four Operational Shifts for Teams on the Exponential

Wu distills four adaptations: Short sprints over long-term roadmaps — the team uses “side quests” (self-directed experiments outside the official roadmap) that produced features like Claude Code on Desktop. Demos over documentation — a rough prototype changes the conversation more than a polished spec. Re-evaluate features with every model release — the Chrome integration emerged when the team noticed users manually switching between Claude Code and their browser. Embrace simplicity — complex workarounds become obsolete when the next model solves the task natively. The team reduced system prompting by 20% with Opus 4.6.

3 — Three-Tool Workflow as Division of Labor

Wu describes a clear split: Claude.ai for strategic thinking and ideation, Claude Code for prototypes, evaluations, and scripts, Cowork for knowledge work, planning, and administration. Peers at Decagon and Datadog confirm similar hybrid workflows that dramatically shorten development cycles.

4 — 41x Improvement in 16 Months

Wu cites METR research: Opus 4.6 can complete software tasks that take a human roughly 12 hours — compared to 21 minutes with Sonnet 3.5. That’s an approximately 41x improvement in 16 months. This isn’t a linear trend but an exponential curve that fundamentally compresses planning horizons.

5 — From Engineering Tool to Organization-Wide Acceleration

The effect doesn’t stop at product and engineering. Data science, finance, legal, marketing, and design adopt AI-native workflows. The shift: instead of sequential handoffs between departments, parallel AI-powered work processes emerge across the entire organization.

Critical Assessment

What Holds Up

What Needs Context

Discussion Questions for the Next Lab

01 Roadmap Revision: If traditional roadmaps fail under exponential model improvement — how do we plan client projects where scope and feasibility can shift fundamentally during implementation?

02 Side Quests as Method: Wu describes self-directed experiments outside the roadmap as an innovation source. What would such a format look like in a consulting context — with fixed budgets, deadlines, and client expectations?

03 Prototype Culture: “Even a rough prototype changes the conversation” — how do we shift our own balance from documentation to demos without sacrificing architecture quality and maintainability?

04 Re-evaluation as Discipline: Every new model should trigger re-evaluation of existing features. How do we systematize this without ending up in permanent re-planning? What’s the right interval?

05 Organization-Wide Shift: Wu describes how not only engineering but also legal, finance, and marketing go AI-native. Which of our clients are ready for this shift — and where is the governance gap?

Sources

Glossary

Side Quest A self-directed experiment outside the official product roadmap. Serves exploratory innovation in environments with high uncertainty about future feasibility.

METR (Model Evaluation & Threat Research) An independent research organization that evaluates AI models for capabilities and risks. Provides standardized benchmarks for task complexity and agent performance.

System Prompting Instructions given to a language model before the actual user query to steer behavior, tone, and capabilities. Less system prompting with better models suggests the model infers more context independently.

Exponential Curve A growth pattern where capability doubles at regular intervals rather than increasing linearly. In the AI context: model capabilities grow faster than human planning typically anticipates.

Weiterführende Diskussionsfragen auf ✳︎ Panoptia Labs Further discussion questions on ✳︎ Panoptia Labs