The era of the augmented architect

The product management landscape of this year is defined by convergence: pressure to build faster (Vibe Coding), generate cash (Revenue Accountability), and prove value (Founder Mode).

The "Product Management is dead" narrative is hyperbolic, but the role is changing. The administrative, process-heavy layer is dying: automated by agents or eliminated by Founder Mode flattening. The traffic cop model doesn't survive when AI generates working prototypes in minutes.

What survives is the core function: value creation. The strategic architecture that determines what should exist and why it matters.

What's changing

The job description from 2020 is obsolete. Writing tickets, managing backlogs, coordinating sprints. These were never the valuable parts. They were scaffolding around the valuable parts.

Agentic AI eliminates the scaffolding. When you can describe a feature and watch an agent generate code, the "product manager as translator" role evaporates. What remains is strategic judgment: what should exist and why it matters commercially.

The augmented architect

The Product Manager of 2026 operates as an Augmented Architect:

Augmented by agentic AI. Agents handle execution: code, design, analysis, experiments. The PM orchestrates agents, judges outputs, and iterates rapidly.

Architecting strategy. The "what" and "why" remain human territory. What should exist? Why does it matter? What's the business model? These require taste and strategic intuition that AI can't replicate.

Accountable for commercial success. The shift from vanity metrics (DAU, NPS) to business metrics (ARR, NRR, ROI) is permanent. Product managers own outcomes, not activities.

The role converges toward what founders do: direct creation, strategic judgment, and commercial accountability.

The evolution

DimensionClassic PM (2020)Modern PM (2025)
Primary Metaphor"CEO of the Product""Full-Stack Architect"
Core ActivityManaging BacklogOrchestrating AI Agents
Key MetricDAU, NPS, VelocityARR, NRR, ROI
Technical SkillConversational LiteracyVibe Coding / Prototyping
Relationship to CodeHandoff to EngineeringDirect AI Generation
Relationship to DesignHandoff to UX/UIGenAI Prototyping
Organizational ModeManager ModeFounder Mode

The divergence

Two groups are emerging. The gap will be dramatic by end of 2026.

One group embraces full-stack capabilities, orchestrates AI agents, builds prototypes, owns commercial outcomes. They operate in Founder Mode: execution-focused, directly accountable.

The other clings to the traffic cop model. They wait for engineers, delegate to designers, treat AI as optional. The scaffolding they relied on is gone. The coordination work is automated. The handoff model is irrelevant.

The first group thrives. The second becomes obsolete.

What this means

The test isn't whether you understand this shift. It's whether you're shipping differently this quarter than last. The capability gap between these groups won't be bridgeable through training or hiring six months from now.

The administrative layer of product management is dying. The strategic layer is more vital than ever. The Product Manager of 2026 is an Augmented Architect: AI-powered, strategy-focused, commercially accountable.

The era of the augmented architect isn't coming. It's here.

Your platform is either a tax or a multiplier

Internal product managers often believe a dangerous myth. They think they don't have to worry about churn.

Since the company mandates the use of their API gateway, design system, or data warehouse, they assume their user base is guaranteed. They have a captive audience.

But in platform product management, users don't churn. They rot.

When users are forced to use a tool they hate, they engage in malicious compliance. They do the bare minimum. They build "shadow IT" workarounds.

They complain to leadership until your budget gets cut. Every internal platform falls into one of two buckets: it is either a Tax or a Multiplier.

If you don't know which one you are, you are probably a tax.

The platform tax

A tax is something you pay because you have to. In the context of a platform, it is a tool that slows users down in exchange for organizational compliance.

Your platform is a tax if it adds friction ("I have to file a ticket and wait three days just to get an API key"), requires distinct effort ("I have to rewrite my code to fit your rigid schema"), or delivers abstract value (the user feels the pain, the company gets the gain).

When you operate as a tax, your relationship with users is adversarial. You are the police. They are the citizens trying to avoid a ticket.

The multiplier effect

A multiplier gives users leverage. It abstracts away the boring, hard, or dangerous parts of development so product teams can skip infrastructure setup and go straight to feature development.

Your platform is a multiplier if it removes friction ("I dropped my code in, and the platform handled auth, logging, and scaling automatically"), accelerates velocity ("This saved me two weeks of integration work"), and delivers immediate value (the user feels the gain directly in their sprint velocity).

When you operate as a multiplier, your relationship with users is a partnership. You are the pit crew. They are the driver.

The mandate trap

The biggest enemy of the multiplier mindset is the corporate mandate.

When a CTO says, "Everyone must use Platform X," the platform team stops selling. They stop treating their users like customers. They start treating them like subordinates.

This is where the product rot begins.

A "tax" team leans on the mandate to explain away their bad UX. They say things like, "They have to use it because of security compliance."

A "multiplier" team ignores the mandate. They build a product so good that teams would voluntarily choose it even if they were allowed to use AWS directly.

Shifting from tax to multiplier

If you suspect you are building a tax, the shift starts with reframing your value proposition. You cannot just be a gatekeeper. The role shifts to gardener.

Sell the "boring" work. Don't sell compliance. Sell the fact that you handle the tasks nobody wants to do. "You must use our library for security compliance" is a tax pitch. "You never have to worry about patching a security vulnerability again" is a multiplier pitch.

Measure time-to-hello-world. Tax teams measure uptime. Multiplier teams measure how fast a new user can ship value. If it takes three days to get onboarded, you are a tax. If it takes five minutes, you are a multiplier.

Compete with the open market. Assume your users can leave. If they could use Vercel, Stripe, or Auth0 instead of your internal tool, would they? If the answer is no, find out why. That gap is your roadmap.

The litmus test

There is one simple question to determine where you stand.

If the executive mandate were removed tomorrow, how many teams would keep using your platform?

If the answer is "nobody," you are a tax. You are living on borrowed time.

If the answer is "most of them," you are a multiplier. You are building leverage.

The choice is clear: gates or ramps. Most teams are still building gates.

Customer satisfaction is a hierarchy, not a metric

We have all been in that strategy meeting. The dashboard is green. Uptime is 99.9%, support ticket volume is down, roadmap is on schedule.

And yet, customers are churning.

The problem isn't the data. It's the definition. We treat "customer satisfaction" as a single bucket. We dump everything into it: bug fixes, new features, polite support emails, brand colors. If the bucket is full, we assume we are winning.

But satisfaction isn't a bucket. It's a hierarchy.

To fix retention, the question isn't "Are they satisfied?" It's "Which layer of the hierarchy is leaking?"

Level 1: The removal of friction (hygiene)

This is the basement. The basic expectations: the login works, the page loads, the data saves, the bill is accurate.

The defining characteristic of this layer is asymmetry. You get zero credit for getting it right. Nobody writes a thank-you note because the "Reset Password" button worked.

But if you get it wrong? You lose everything.

When you fail at this level, the customer emotion is anger. Angry customers submit tickets. They tweet. They churn loudly.

Many product teams confuse fixing friction with "delivering value." They spend quarters squashing bugs and refactoring code, thinking they are improving the product. They aren't. They are simply stopping the bleeding.

This is the floor. You cannot build a strategy here, but you can certainly die here.

Level 2: The delivery of outcome (utility)

This is the ground floor. This is why the customer hired your product in the first place.

This layer isn't about whether the software works. It's about whether the customer works. Did they get the result?

A tax software user doesn't want a delightful interface. They want their taxes filed without an audit. A ride-share user doesn't want a chatty driver. They want to be at the airport by 5:00 PM.

When you fail at this level, the customer emotion isn't anger. It is indifference. [...]

Gemini 3 Proves the NVIDIA Tax Is Optional

Google's Gemini 3 landed last week with impressive reviews: frontier-class performance that beats OpenAI and Anthropic except for its agentic code. Conventional wisdom said Google was lagging behind OpenAI, which remains true in adoption. But on capability, they have a real chance to catch up. The tech press is focused on benchmarks and capabilities.

They're missing the real headline: Google trained these models entirely on TPUs. Zero NVIDIA dependency.

While competitors debate model quality, Google just decoupled its cost structure from the entire market. OpenAI and Microsoft pay massive margins to Jensen Huang. Google is paying cost-plus to its own hardware division.

It's the margin war

In a commodity market, the low-cost provider eventually wins. AI inference is racing toward commoditization. Google's vertical integration is the ultimate cheat code.

Consider the math. NVIDIA-dependent companies must price APIs to cover energy, cloud operations, and NVIDIA's 75% gross margins. Google prices APIs to cover energy and silicon fabrication costs.

This allows Google to trigger a race to the bottom on pricing that no NVIDIA-dependent competitor can survive. They can effectively subsidize the model indefinitely to protect Search.

Competitors, meanwhile, burn capital paying premium rents for compute. This is "compute sovereignty": scaling capacity without third-party permission.

Amazon proves it's not just Google

Google isn't alone. Amazon is running the same playbook with AWS Trainium chips.

Anthropic is training Claude 4.x on 500,000 AWS Trainium2 chips, with plans to scale to one million by year-end. AWS claims 30-40% better price-performance than NVIDIA equivalents. That's structural cost advantage, not incremental improvement.

Amazon's approach mirrors Google's vertical integration but with a different business model. Where Google optimizes for its own models, AWS sells compute sovereignty as a service.

The software moat is deeper than the hardware

Cheap chips don't matter if nobody knows how to program them. This is the "Island Problem": great infrastructure developers won't adopt. It remains the single biggest risk to Google's strategy.

NVIDIA is just not a hardware company. They are a software platform disguised as a chip manufacturer. The CUDA ecosystem is the deepest moat in tech.

Every researcher coming out of Stanford or MIT learns PyTorch on CUDA. Every major open-source library is optimized for CUDA first. Moving a production workload from NVIDIA to TPU isn't just a "recompile." It often requires rewriting code in JAX or dealing with XLA compiler friction.

If Google wins on efficiency but loses on developer mindshare, they end up with the best internal infrastructure that nobody else wants to use.

The model is the only leverage left

On the flip side, if Google cannot break the CUDA stranglehold, its TPUs remain a private island. In this scenario, their only path to victory is abstracting the hardware away entirely.

They must force the market to consume Inference APIs, not raw compute.

If Gemini 3 is sufficiently powerful (and sufficiently cheap) developers won't care what silicon it runs on. They will never touch the metal. They will just hit the API endpoint. Google's strategy relies on turning the TPU from a developer hurdle into a hidden margin engine.

The industry is splitting into two distinct camps.

OpenAI, Meta (training Llama 4 on H100s), and most startups rent NVIDIA capacity. They optimize for speed-to-market and developer familiarity but remain locked into NVIDIA's margin structure and the CUDA ecosystem.

Google with TPUs and Amazon with Trainium optimize for unit economics. They're betting that price-performance eventually trumps developer familiarity, and that vertical integration becomes the only sustainable path in a commoditizing market.

The question isn't which group is right. It's which advantage compounds faster: ecosystem lock-in or cost structure.

Dead Time Is Story Time

I stumbled on David Maister's 1985 research on waiting psychology while thinking about loading screens. His foundational principle: occupied time feels shorter than unoccupied time. But he uncovered something deeper. Anxiety, uncertainty, and unexplained delays make waits feel exponentially longer.

That got me thinking. Most product teams treat waiting as friction to eliminate. Load faster. Reduce checkout steps. Skip the queue.

But you can't eliminate all dead time. Users will wait during loading, onboarding, checkout, processing. The question isn't whether they wait, it's what you do with that captive attention.

Look at how the best products handle waiting

As I explored this more, I started noticing patterns everywhere.

Look at how Disney handles theme park queues. They don't just move lines faster. They build interactive games into waiting areas. Characters parade through crowds. Space Mountain's twisting paths hide the full queue length, reducing anxiety about what's ahead.

Consider how Asana onboards new users. They don't overwhelm people with Gantt charts and dependencies upfront. They start with basic project creation, then gradually reveal advanced features as users gain confidence. The onboarding wait becomes capability building.

The pattern: people tolerate waits when they understand why they're waiting, see progress toward completion, and have something meaningful to occupy the time.

What separates high performers from the rest

The best SaaS products get dramatically higher onboarding completion than industry averages. They use progressive disclosure: revealing one valuable feature at a time, teaching as users engage, turning configuration into discovery.

E-commerce brands fighting cart abandonment don't just optimize checkout speed. They explain shipping timelines transparently, show order progress in real-time, and use recovery emails that provide value (product recommendations, size guidance) alongside the nudge to complete purchase.

These companies understand that dead time isn't just infrastructure to optimize. It's experience to design.

Why this matters more than you think

Page speed matters. Even small delays tank conversions. Amazon, Google, and others have documented how milliseconds of slowdown directly impact revenue.

But speed alone doesn't solve this. Pinterest discovered something interesting: they reduced perceived wait times dramatically without dramatically changing actual load speeds. Better skeleton screens and progress indicators made the experience feel faster. The result was measurable increases in both search traffic and sign-ups.

The insight: perception matters as much as performance.

Where is your dead time?

Look for moments where users have no choice but to wait:

  • Onboarding flows with account setup and data imports
  • Checkout processes with payment verification
  • Loading states during search or complex calculations
  • Processing delays for reports, exports, or deployments
  • Queue systems for high-demand access

Each one is captive attention. Each one is an opportunity to educate, build confidence, reduce anxiety, or deepen engagement.

The question isn't whether your product has dead time. It's whether you're wasting it on spinners, or turning it into story time.

The Feature Factory Problem AI Amplifies

Your team just shipped three features this week. Last quarter, that would have taken a month. AI tools turned your engineers into feature factories. Your designers generate variants in minutes. Your PMs prototype without waiting for engineering resources.

Everyone's celebrating velocity. Who's checking if you're solving the right problems?

Creation velocity isn't validation velocity

Recent research shows contradictory results on AI's impact. Some studies report significant productivity gains, others find developers actually slow down when using AI tools. The pattern emerging is clear: AI accelerates code generation, but delivery stability often decreases and quality concerns rise.

You can ship faster. But are you learning faster?

The real gap isn't speed. It's the mismatch between how quickly you can build and how quickly you can validate what to build. As I explored previously, AI amplifies senior strategic judgment while commoditizing tactical execution. Without that senior judgment directing what to build, teams become feature factories.

Most teams obsess over who builds features (PMs? Engineers? Designers?) and how fast they ship. Almost nobody asks whether those features matter to customers.

The warning signs aren't dramatic. In B2B, users don't immediately churn. Instead, adoption stays low. Features exist but go unused. Customer success teams field more "why did you build this?" questions. Renewal conversations get harder six months later.

Traditional product-market fit frameworks assume you're testing hypotheses at a measured pace. You build, you measure, you learn, you iterate. AI tools break that rhythm. You can generate and ship features faster than your instrumentation can tell you if the previous batch worked.

The path-of-least-resistance problem

Teams pick AI use cases based on what's readily available, not what aligns with strategic goals or delivers meaningful customer impact.

You have a new AI tool that generates UI components. So you build features that need UI components, whether or not those features solve real user problems. The tool shapes the roadmap instead of customer needs shaping tool selection.

Amazon's principle of working backwards matters more now, not less. Start with customer needs, then figure out what to build and how to build it. Most teams do the opposite: start with what AI tools can do easily, then justify it with assumptions about customer value.

Research on AI project failures consistently points to the same issue: too much attention on technical capabilities, not enough on actual market needs. Most AI projects fail not because the technology doesn't work but because teams built the wrong thing quickly.

Speed as tactic, not strategy

The companies navigating this well treat speed as a tactic, not a strategy. They use AI to accelerate after they've validated direction, not to spray features and hope something sticks.

This requires different metrics. Not just "features shipped per sprint" but "validated learning per sprint." Not just "time to market" but "time to meaningful user engagement." Not just "productivity gains" but "impact per unit of effort."

Product managers still balance the same three dimensions: user needs, business viability, and technical feasibility. What changes is the first pillar. Understanding user needs now means distinguishing between what users actually need and what AI makes easy to build.

AI can help with validation too. Synthesize user interviews at scale. Analyze support tickets for patterns. Enable rapid prototype testing. But the craft of knowing what to validate and interpreting results still requires human judgment. That's where the gap lives.

The test that matters

Ask what you learned this week, not what you shipped.

If the answer is "we shipped five features," you're measuring activity. If the answer is "we validated that users struggle with X and confirmed that approach Y doesn't solve it," you're measuring learning. Fast teams don't ship more, they learn faster.

AI tools are incredible for accelerating learning. Generate multiple variations, test them quickly, and see what works. But only if you're testing the right problem.

Speed is a compounding advantage when paired with direction. Speed without direction just gets you lost faster.

Even when you're building the right features at the right pace, there's another bottleneck most teams ignore. AI can generate code quickly. Someone still has to ensure it's production-ready.

AI Commoditizes Entry-Level Work While Amplifying Senior Value

Everyone's asking the wrong question about AI and product teams.

The debate splits into two camps: one believes product managers will code their way to replacing engineers, the other thinks engineers will own strategy and eliminate PMs. Both narratives miss what's actually happening. AI isn't replacing entire functions. It's splitting each function into three tiers, and only one of them is shrinking.

The same pattern across three functions

Look at what's happening to engineers first. Employment for software developers aged 22-25 has declined nearly 20% from its peak in late 2022, according to Stanford's Digital Economy Lab. Computer engineering graduates now face a 7.5% unemployment rate, one of the highest across majors.

What's getting automated? Boilerplate code. Unit tests. API maintenance. The tasks companies used to assign to junior developers to help them learn.

Now look at designers. Many studios used to hire junior designers specifically for wireframing. Those roles are shrinking. The State of AI in Design Report found that 89% of designers improved their workflow with AI this year, but the improvements concentrate on production work: variant generation, copy filling, visual polish.

Product managers see the same split. McKinsey found that generative AI improved PM productivity by 40%, but the gains came from automating tactical work: user stories, performance reports, backlog maintenance, PRD drafting. The strategic work, the judgment calls, the stakeholder dynamics? Still human.

Three tiers emerging

The pattern is clear across all three functions. Entry-level tactical work gets commoditized. Senior strategic work gets amplified. Mid-level roles transform into something new.

Tier 1: Commoditized

Junior engineers who write boilerplate code. Junior designers who create wireframes. Junior PMs who draft user stories. AI tools handle these tasks faster and often as well as someone in their first year on the job.

Marc Benioff announced Salesforce will hire "no new engineers" in 2025, citing AI-driven productivity gains. Anthropic CEO Dario Amodei warned that AI could wipe out half of entry-level white-collar jobs. But AWS CEO Matt Garman pushed back, calling the idea of replacing junior developers with AI "one of the dumbest things I've ever heard."

The debate reveals the tension: AI can do the tasks, but those tasks serve a purpose beyond just getting work done. They're how people learn.

Tier 2: Transformed

Mid-level roles aren't disappearing. They're morphing into hybrid positions that require both AI literacy and deep domain expertise. These are the people who know enough to guide AI tools effectively and spot when AI outputs miss the mark.

The new job descriptions reflect this: "AI Orchestrators," "Product Engineers," "Full-Stack PMs." People who move fluidly between strategy and execution, using AI as a force multiplier rather than a replacement.

Tier 3: Amplified

Senior roles gain value. System architecture. Product vision. Strategic design decisions. User empathy. These capabilities matter more, not less, in an AI-enabled world.

Here's why: the more skilled you are at your craft, the better results you get with AI. A senior engineer can spot architectural flaws in AI-generated code that a junior might miss. A seasoned designer knows when AI-generated variants sacrifice usability for aesthetics. An experienced PM can tell when AI-drafted requirements miss the strategic context.

The hollowed-out career ladder

This creates a systemic problem nobody wants to talk about: if entry-level roles commoditize but senior expertise remains valuable, how does anyone become senior?

You need plenty of seniors at the top. AI tools handle grunt work at the bottom. But there are very few juniors in the middle learning the craft.

This threatens the long-term talent pipeline across product, engineering, and design. Companies benefit from AI productivity gains today while quietly eroding their ability to develop senior talent tomorrow.

The skills that make someone valuable at Tier 3 develop through years of doing Tier 1 work. You learn system architecture by first building components. You develop product sense by first writing user stories and seeing what ships versus what sits in the backlog. You understand good design by first creating wireframes and getting feedback.

Remove the learning ground, and you cut off the path to expertise. The question of whether curiosity beats tenure in this environment isn't academic: it determines whether companies can develop the senior talent they'll need.

What this means in practice

The conversation shouldn't be "will AI replace PMs or engineers or designers?" It should be: "How do we structure learning when tactical work gets automated?"

Some companies respond by raising the bar for entry-level roles. Junior positions now require AI literacy plus traditional fundamentals. That solves the immediate hiring problem but makes it harder for people to break into the field.

Other companies may be experimenting with new mentorship models: pairing early-career people directly with senior staff and using AI tools to accelerate their learning rather than replace their contributions.

The companies getting this right recognize that AI doesn't just change how work gets done. It changes how expertise develops. They're designing career paths that acknowledge both the productivity gains from AI and the human judgment that AI amplifies but cannot replace.

But commoditization is only part of the story. The bigger question is what teams are actually building with all this newfound speed. Shipping 3x faster sounds great. Whether that speed translates to customer value is a different question, one I'll tackle next.

Read next: The Feature Factory Problem AI Amplifies

Why AI Agents Fail Today Despite the Hype

The agentic AI hype promises autonomous decision-makers that replace employees or dramatically boost efficiency. The reality, according to Maria Sukhareva (Principal AI Expert at Siemens) in "Why AI Agents Disappoint," is that general-purpose AI agents don't work for most real-world business use cases.

The WebArena benchmark proves the gap quantitatively. Researchers created realistic web environments (e-commerce sites, forums, development platforms) and asked GPT-4-based agents to complete end-to-end tasks like "Find the cheapest phone case and email me the link." Success rate: 14.41%.

The architecture is brittle

Current agents follow a ReAct cycle (Plan, Act, Observe, Repeat) as I explored previously. This looks like reasoning, but it's sequential token prediction choosing from pre-defined tools. When one step fails, the entire chain collapses.

The problem isn't the architecture alone. It's what happens when reality deviates from the plan.

If an agent struggles through a task and eventually succeeds, it retains zero memory of that success. Next session, it repeats the same mistakes. There's no mechanism to update its knowledge base or behavior from experience. (Research like Meta's Early Experience approach explores ways agents could learn from their own rollouts, but these methods aren't yet production-ready.)

Error propagation compounds this. A single wrong click or a misread file in step two ruins the entire workflow. Attempted fixes like self-reflection or multi-agent debate act as band-aids. They sometimes amplify false reasoning rather than correct it.

Why coding works

AI-assisted coding (GitHub Copilot, Claude Code) is the exception. Agents genuinely deliver value here.

The environment is constrained. The IDE provides clear boundaries. The data is primarily text and code. Feedback loops are immediate. You run the code, see if it works, and adjust.

This reveals the requirements for agents to succeed elsewhere: constrained environments, homogeneous data types, and fast feedback loops.

Real-world business tasks fail all three tests. A financial audit requires juggling emails, database logs, PDF invoices, and regulatory texts simultaneously. When a document is missing or unreadable, agents hallucinate results rather than problem-solve like humans do.

Business workflows assume human common sense. Nobody emails a contact ten times in one minute. Nobody needs explicit instructions not to delete the production database. Agents require rigid IF...THEN rules for everything. They can't handle dynamic obstacles (a file moved to a new database, a contact out of office) without being explicitly programmed for each scenario.

We're a decade away from autonomous?

The market hype suggests we're approaching the "Observer" level, where machines work fully autonomously. Sukhareva argues we're actually at the "Collaborator" level where humans guide machines. Citing Andrej Karpathy, she estimates it will take at least ten years to fix these fundamental cognitive issues.

The gap isn't just technical. It's architectural. Current agents lack the cognitive structure for proactive learning, robust error recovery, and multimodal reasoning.

Companies investing in "agentic AI" based on hype videos and demos should understand what they're actually buying: brittle sequential machines that work in constrained environments with immediate feedback. Not autonomous decision-makers.

The coding success proves agents can work when the environment matches their capabilities. The question for product teams isn't whether to use agents. It's whether your use case looks more like an IDE or like a financial audit.