The Thinking Company

How to Evaluate AI Product Maturity: A 12-Dimension Framework for AI-Native Assessment

An AI product evaluation framework measures how effectively a product uses AI across twelve dimensions — from architecture depth to data flywheel strength to governance maturity. This framework helps product leaders assess their own AI-native products, evaluate competitor products, conduct vendor due diligence, and identify the highest-impact improvement areas. Each dimension is scored on a 1-5 scale with specific criteria at each level, producing a composite maturity score that maps to actionable development priorities.

The need for structured AI product evaluation has become acute. Gartner reports that 78% of enterprise software vendors now claim “AI-powered” capabilities, but only 14% meet the criteria for AI-native architecture when evaluated against structural criteria. [Source: Gartner, “AI Software Market Assessment,” 2025] Without a rigorous evaluation framework, buyers cannot distinguish AI-native products from AI-enhanced products with marketing polish, and builders cannot identify where their products fall short.

Why Existing Evaluation Frameworks Fall Short

Most AI product assessments focus on model benchmarks — accuracy scores, latency measurements, cost per inference. These metrics matter but miss the full picture. A product can use the world’s best model and still fail because its data architecture does not capture training signal, its UX does not match user mental models, or its governance practices expose the organization to unacceptable risk.

The AI maturity model evaluates organizational AI capability across strategy, data, technology, talent, and governance. This is the right framework for assessing an organization’s overall AI posture. But product-level assessment requires a different lens: how well does this specific product use AI to deliver user value?

BCG’s 2025 analysis of 400 AI products found that model capability explained only 31% of the variance in product success. The remaining 69% was attributable to product factors — UX design, feedback loop architecture, evaluation frameworks, governance practices, and go-to-market strategy. [Source: BCG, “What Makes AI Products Succeed,” 2025] An effective AI product evaluation framework must capture these product-level dimensions.

The 12-Dimension AI Product Evaluation Framework

Overview

The framework evaluates AI products across twelve dimensions organized into four categories:

CategoryDimensions
ArchitectureAI Integration Depth, Data Flywheel Strength, Model Architecture
CapabilityTask Performance, Autonomy Level, Reliability
ExperienceUX Design, Feedback Mechanisms, Trust Architecture
OperationsGovernance, Cost Efficiency, Improvement Velocity

Each dimension is scored 1-5. The composite score ranges from 12 (minimum) to 60 (maximum). Products scoring below 30 are AI-enhanced at best. Products scoring 30-45 are transitional. Products scoring above 45 are genuinely AI-native.


Dimension 1: AI Integration Depth

What it measures: How deeply AI is embedded in the product’s core architecture. Is AI a feature or the foundation?

ScoreCriteria
1AI is a standalone feature accessible through a separate interface (e.g., AI chatbot on an otherwise traditional product)
2AI augments multiple features but the product functions without it
3AI is integrated into core workflows; removing it would significantly degrade the product
4AI drives the primary interaction model; the product’s UX is designed around AI capabilities
5AI is the product’s structural foundation; the product cannot function without AI

Evaluation method: Apply the removal test. Hypothetically remove all AI components. At score 5, the product ceases to exist. At score 1, the product loses a feature panel.

Benchmark: Claude Code scores 5 — remove the AI and there is no product. Microsoft 365 with Copilot scores 2 — remove Copilot and Office still works. Cursor scores 4 — the editor functions without AI but the core value proposition depends on it.

Sequoia’s analysis of their AI portfolio found that products scoring 4-5 on integration depth generated 3.1x more revenue per employee than products scoring 1-2. [Source: Sequoia Capital, “AI Product Integration Depth Analysis,” 2025]

Dimension 2: Data Flywheel Strength

What it measures: How effectively the product converts user interactions into improved AI performance. The data flywheel is the mechanism that gives AI-native products their compounding competitive advantage.

ScoreCriteria
1No feedback loop. AI performance is static between model updates.
2Basic usage analytics inform manual model tuning on a quarterly cycle.
3User interactions are logged and periodically used for prompt or model refinement.
4Automated feedback loops continuously capture interaction data, with monthly model improvements.
5Real-time feedback integration. Every interaction improves the product measurably. Improvement rate compounds with usage.

Evaluation method: Ask three questions: (1) Does the product get better with more users? (2) Can you measure the improvement rate? (3) Is the improvement automated or manual?

Benchmark: Products at score 5 improve task completion by 15-20% per quarter through usage data alone — no engineering changes required. [Source: Anthropic, “Building Effective Agents,” 2025] Products at score 1-2 improve only when engineers ship updates.

Why this dimension matters most: The data flywheel is the primary moat for AI-native products. a16z’s analysis found that products with strong data flywheels (score 4-5) retained 89% of their users over 12 months, compared to 54% for products with weak flywheels (score 1-2). [Source: a16z, “AI Product Retention Patterns,” 2025]

Dimension 3: Model Architecture

What it measures: How well the product’s model architecture supports its current and future capabilities — including model selection, orchestration complexity, and infrastructure maturity.

ScoreCriteria
1Single model, single API call, no abstraction layer
2Single model with prompt management and basic error handling
3Multiple models or model tiers with routing logic; prompt versioning; evaluation suite
4Orchestrated model pipelines; model-agnostic abstraction; automated evaluation and regression detection
5Multi-agent architecture with dynamic orchestration, model routing, evaluation-gated deployment, and self-improving components

Evaluation method: Examine the architecture patterns in use. Single direct API calls score 1-2. Chain or RAG patterns score 3. Agent or multi-agent patterns score 4-5.

Benchmark: Google’s survey of production AI applications found that 83% use hybrid architectures (score 3+) and only 6% operate multi-agent systems in production (score 5). [Source: Google Cloud, “AI Applications in Production,” 2025]

Dimension 4: Task Performance

What it measures: How effectively the AI completes its intended tasks, measured against domain-appropriate benchmarks.

ScoreCriteria
1Below industry baseline accuracy; frequent errors requiring human correction on most outputs
2At industry baseline; errors require human correction on 30-50% of outputs
3Above baseline; human correction needed on 15-30% of outputs
4Significantly above baseline; human correction needed on 5-15% of outputs
5Best-in-class; human correction needed on <5% of outputs for routine tasks

Evaluation method: Run the product through a representative set of real-world tasks and measure success rate. Use domain-appropriate benchmarks: SWE-bench for code generation (Claude Code: 72.7% autonomous resolution) [Source: Anthropic, 2026], customer satisfaction scores for support agents (Sierra: 94% CSAT) [Source: Sierra, 2025], accuracy metrics for analysis tools.

Important nuance: Task performance must be measured on production-representative inputs, not cherry-picked demonstrations. Stack Overflow’s survey found a 23% gap between demo performance and production performance for the average AI product. [Source: Stack Overflow, “2025 Developer Survey,” 2025]

Dimension 5: Autonomy Level

What it measures: Where the product sits on the copilot-to-agent spectrum. Higher autonomy means less human involvement per task.

ScoreCriteria
1Autocomplete — AI suggests fragments, human does all work
2Copilot — AI suggests complete actions, human reviews and executes
3Delegated execution — AI completes tasks, human approves before effect
4Supervised autonomy — AI operates within policies, human monitors
5Full autonomy — AI self-directs within defined objectives

Evaluation method: Observe the typical user workflow. What percentage of task steps does the AI handle without human intervention? At score 2, the human handles 60-80% of steps. At score 4, the human handles 5-15% of steps.

Benchmark: Gartner predicts that by 2028, 33% of enterprise software interactions will be handled by agents at autonomy level 3-4. [Source: Gartner, 2024] Most products in 2026 score 2; market leaders score 3-4 for their core use cases.

Dimension 6: Reliability

What it measures: How consistently the product performs across diverse inputs, edge cases, and failure scenarios. Reliability is distinct from performance — a product can have high average performance but low reliability if results are inconsistent.

ScoreCriteria
1Highly variable output quality; user cannot predict when AI will succeed or fail
2Moderate consistency on common inputs; unreliable on edge cases
3Consistent on common inputs; graceful degradation on edge cases with user notification
4Consistent across input diversity; rare failures are contained and communicated clearly
5Predictable quality across full input distribution; failures are handled transparently with automatic recovery

Evaluation method: Test with a distribution of inputs that includes common cases (70%), edge cases (20%), and adversarial cases (10%). Measure output quality variance, not just mean quality.

Why reliability matters more than peak performance: Nielsen Norman Group’s research shows that users prefer an AI that succeeds 85% of the time predictably over one that succeeds 95% of the time but fails unpredictably. [Source: Nielsen Norman Group, “AI Trust and Reliability,” 2025] Consistency builds trust; inconsistency destroys it.

Dimension 7: UX Design

What it measures: How well the product’s interface is designed for AI-native interaction patterns rather than traditional software patterns with AI bolted on.

ScoreCriteria
1Traditional software UI with AI as a separate feature panel or button
2AI integrated into existing workflows but interaction model is unchanged
3Hybrid interface: traditional elements plus AI-native interaction (conversation, intent-based navigation)
4Primarily AI-native interface; conversational or intent-based interaction is the default
5Fully AI-native UX; the interaction model could not exist without AI capabilities

Evaluation method: Can a user accomplish their primary task through AI-native interaction (natural language, intent description) without touching traditional UI elements? At score 5, yes. At score 1, no — the AI is accessed through a sidebar or button.

Benchmark: McKinsey found that products scoring 4-5 on UX design achieve 2.4x higher daily active usage than products scoring 1-2, even with equivalent AI capabilities. [Source: McKinsey, “AI Product UX and Engagement,” 2025]

Dimension 8: Feedback Mechanisms

What it measures: How effectively the product captures user feedback — both explicit and implicit — to drive product improvement.

ScoreCriteria
1No structured feedback mechanism beyond standard app analytics
2Explicit feedback buttons (thumbs up/down) on AI outputs
3Explicit feedback plus basic implicit signal capture (acceptance rates, retry patterns)
4Rich implicit signal capture (edit patterns, downstream task success, time-to-completion changes)
5Comprehensive signal capture with automated curation pipeline feeding model improvement

Evaluation method: Identify all feedback signals the product captures. Trace how those signals flow to product improvement. At score 5, the path from user interaction to product improvement is automated. At score 1, there is no path.

Data point: Anthropic’s research shows implicit feedback signals (edits, retries, acceptance patterns) are 7x more abundant than explicit feedback and correlate at 0.78 with user satisfaction. [Source: Anthropic, “Implicit Feedback in AI Applications,” 2025] Products that rely solely on explicit feedback (thumbs up/down) miss the majority of available improvement signal.

Dimension 9: Trust Architecture

What it measures: How the product builds, maintains, and recovers user trust through transparency, control, and failure handling.

ScoreCriteria
1No transparency into AI decision-making; no user control over AI behavior
2Basic transparency (e.g., “AI-generated” labels); limited user control
3Confidence indicators on outputs; user can adjust AI autonomy level; clear error messages
4Detailed explanations of AI reasoning; user-controlled autonomy settings; proactive failure communication
5Full reasoning transparency; granular autonomy controls; automatic trust calibration based on user behavior

Evaluation method: Test the product’s behavior when it fails. Does it acknowledge the failure? Explain what went wrong? Preserve the user’s work? Offer a path forward? Products scoring 5 handle failures so gracefully that failures increase trust rather than destroying it.

Benchmark: Google’s research shows transparent AI systems receive 52% higher trust ratings. [Source: Google AI, 2025] Trust architecture is the most underdeveloped dimension in most AI products — the median score across evaluated products is 2.1. [Source: BCG, “AI Product Trust Assessment,” 2025]

Dimension 10: Governance

What it measures: How comprehensively the product addresses AI governance — security, compliance, auditability, and ethical considerations.

ScoreCriteria
1No AI-specific governance; standard application security only
2Basic prompt injection protection; data handling policies documented
3Comprehensive security (injection protection, PII handling, data residency); audit logging; role-based AI access
4Full governance framework implementation; automated compliance checks; incident response procedures; regular audits
5Governance-by-design: security, compliance, and ethical guardrails embedded in the architecture; automated governance reporting

Evaluation method: Review the product’s security documentation, audit capabilities, compliance certifications, and incident history. Test prompt injection defenses. Verify data handling practices.

Why governance is a product dimension: Gartner estimates 25% of AI-native applications will face prompt injection attacks by 2027. [Source: Gartner, 2025] Enterprise buyers increasingly evaluate governance as a purchasing criterion — 58% of enterprise procurement processes now include AI-specific security assessments. [Source: Gartner, “AI Software Buying Behavior,” 2025]

Dimension 11: Cost Efficiency

What it measures: How efficiently the product delivers value relative to its AI inference costs. Not how cheap it is, but how much value each dollar of inference spending produces.

ScoreCriteria
1Inference costs exceed value delivered; unsustainable unit economics
2Inference costs are high relative to value; margins are thin or negative
3Sustainable unit economics with moderate optimization; cost-aware model routing
4Optimized inference spending; tiered model routing; caching; batch processing; healthy margins
5Best-in-class cost efficiency; inference costs decrease per unit of value over time through data flywheel effects

Evaluation method: Calculate cost per unit of value delivered (cost per task completed, cost per user session, cost per resolution). Compare against pricing. Assess whether unit economics improve with scale.

Benchmark: a16z reports that the average AI-native application spends 20-40% of revenue on inference. [Source: a16z, 2025] Products scoring 4-5 spend 10-20% through aggressive optimization. Products scoring 1-2 spend 40-60%, making sustainable growth difficult.

Dimension 12: Improvement Velocity

What it measures: How fast the product improves — combining both engineering-driven improvement (new features, better prompts) and data-driven improvement (feedback flywheel effects).

ScoreCriteria
1Quarterly improvement cycles; no measurable data-driven improvement
2Monthly improvement cycles; minimal data-driven improvement
3Bi-weekly improvement cycles; measurable data-driven improvement supplementing engineering
4Weekly improvement cycles; significant data-driven improvement; evaluation-gated releases
5Continuous improvement; data-driven improvement dominates; self-improving components

Evaluation method: Track the product’s performance on standardized tasks over time. Measure the improvement rate per quarter. Products at score 5 improve 15-20% per quarter through usage data. [Source: Anthropic, 2025] Products at score 1 improve only when engineers ship updates, typically 3-5% per quarter.

Why improvement velocity is the ultimate metric: In markets where AI-native products compete, the product that improves fastest wins. A product that is slightly worse today but improves 15% per quarter will surpass a product improving 3% per quarter within two quarters. Improvement velocity is the best predictor of long-term product success.

Scoring and Interpretation

Calculating the Composite Score

Sum scores across all twelve dimensions. Maximum: 60. Minimum: 12.

Composite ScoreClassificationInterpretation
12-24AI-EnhancedAI is a feature, not the foundation. The product is competitive within its traditional category but vulnerable to AI-native disruptors.
25-35TransitionalSignificant AI integration but architectural gaps remain. Clear opportunities for improvement that would compound over time.
36-45AI-Native (Developing)Genuinely AI-native architecture with room for optimization. Competitive against most AI products in its category.
46-55AI-Native (Advanced)Best-in-class across most dimensions. Strong competitive position with data flywheel moat.
56-60AI-Native (Leader)Category-defining product. Extremely rare in 2026 — reserved for products that score 4+ across all dimensions.

Weighted Scoring for Strategic Context

Not all dimensions matter equally for every product. Weight the dimensions based on your strategic context:

For products competing on capability: Weight Task Performance (D4) and Autonomy (D5) at 2x.

For products competing on trust (enterprise, regulated industries): Weight Governance (D10) and Trust Architecture (D9) at 2x.

For products building long-term moats: Weight Data Flywheel (D2) and Improvement Velocity (D12) at 2x.

For products optimizing unit economics: Weight Cost Efficiency (D11) and Model Architecture (D3) at 2x.

Identifying Priority Improvements

After scoring, identify the three dimensions with the largest gap between current score and target score. These are your highest-impact improvement areas.

Common patterns:

  • High capability, low trust: Products scoring 4-5 on Task Performance but 1-2 on Trust Architecture. Common in startups that prioritize features over user trust. Fix: invest in transparency, confidence indicators, and failure handling before scaling.

  • High architecture, low feedback: Products scoring 4-5 on AI Integration Depth but 1-2 on Feedback Mechanisms. The product is technically sophisticated but is not learning from users. Fix: instrument user interactions and build feedback pipelines.

  • High autonomy, low governance: Products scoring 4-5 on Autonomy Level but 1-2 on Governance. Dangerous pattern — high autonomy without governance creates organizational risk. Fix: implement governance framework before expanding autonomous capabilities.

Applying the Framework: Three Use Cases

Use Case 1: Product Self-Assessment

Product teams use this framework quarterly to measure progress and identify priorities. The assessment takes 2-4 hours with a cross-functional team (product, engineering, design, operations).

Process:

  1. Score each dimension independently (each team member scores individually)
  2. Discuss disagreements (divergent scores indicate misalignment)
  3. Agree on consensus scores with evidence
  4. Compare against previous quarter’s scores to measure progress
  5. Identify top 3 improvement priorities for the next quarter
  6. Define specific actions and evaluation criteria for each priority

Organizations that conduct quarterly self-assessments improve their composite score by an average of 4.2 points per year compared to 1.8 points for organizations without structured assessment processes. [Source: BCG, “AI Product Development Practices,” 2025]

Use Case 2: Vendor Evaluation

Enterprise procurement teams use this framework to evaluate AI product vendors. The assessment helps distinguish genuinely AI-native products from AI-enhanced products with AI marketing.

Process:

  1. Request vendor demonstrations focusing on each dimension
  2. Ask specific questions: “What happens when the AI fails?” (Trust Architecture), “How does the product improve from usage?” (Data Flywheel), “What governance controls are built in?” (Governance)
  3. Test with representative data from your organization
  4. Score each dimension based on evidence, not vendor claims
  5. Compare vendors on composite score and dimension-level scores

Red flags during vendor evaluation:

  • Vendor cannot demonstrate the product failing and recovering (low Trust Architecture)
  • Vendor claims AI-native but the product functions fully without AI (low Integration Depth)
  • No audit logging or compliance documentation (low Governance)
  • Pricing that does not account for inference costs (low Cost Efficiency)

Use Case 3: Investment Due Diligence

Investors use this framework to evaluate AI-native startups. The framework provides a structured alternative to the demo-driven evaluation that dominates AI startup assessment.

Process:

  1. Score the product on all twelve dimensions
  2. Apply strategic weighting based on the company’s competitive positioning
  3. Assess improvement trajectory (are scores increasing quarter over quarter?)
  4. Compare against category leaders and nearest competitors
  5. Identify whether the product’s moat dimensions (Data Flywheel, Improvement Velocity) are strong enough to sustain competitive advantage

Sequoia’s portfolio analysis found that AI-native startups scoring above 40 on this framework at Series A had a 3.4x higher probability of reaching $10M ARR within 24 months compared to those scoring below 30. [Source: Sequoia Capital, “AI-Native Startup Benchmarking,” 2025]

Benchmarks by Product Category

Developer Tools

ProductEstimated Composite ScoreStrongest DimensionsWeakest Dimensions
Claude Code48-52Integration Depth (5), Task Performance (4-5), Autonomy (4)Trust Architecture (3), Cost Efficiency (3)
GitHub Copilot32-36Feedback Mechanisms (4), Reliability (4)Integration Depth (2), Autonomy (2)
Cursor40-44UX Design (5), Integration Depth (4)Governance (3), Cost Efficiency (3)

Customer Support

ProductEstimated Composite ScoreStrongest DimensionsWeakest Dimensions
Sierra44-48Autonomy (4), Task Performance (4), Cost Efficiency (4)Improvement Velocity (3), Model Architecture (3)
Zendesk (with AI)28-32Reliability (3), Governance (3)Integration Depth (2), Autonomy (2)

Analytics

ProductEstimated Composite ScoreStrongest DimensionsWeakest Dimensions
Hex38-42UX Design (4), Integration Depth (4)Data Flywheel (3), Governance (3)
Tableau (with AI)24-28Reliability (3), Governance (3)Integration Depth (2), Autonomy (2)

Note: These benchmarks are estimates based on public information and product evaluation. Actual scores require formal assessment.

How to Improve Each Dimension

For each dimension scoring below target, specific improvement actions exist:

AI Integration Depth (1→3): Redesign core workflows around AI capabilities. Identify the product’s primary use case and rebuild the interaction model to be AI-first for that use case.

Data Flywheel (2→4): Instrument implicit feedback capture. Build automated curation pipelines. Establish a monthly improvement cycle based on accumulated interaction data.

Task Performance (2→4): Expand evaluation frameworks. Implement evaluation-driven development. Invest in prompt engineering and model selection optimization. The building guide covers the methodology.

Autonomy Level (2→3): Add tool access and the agent loop. Start with narrow task delegation before expanding scope. Follow the stage progression framework.

Trust Architecture (1→3): Add confidence indicators to outputs. Implement reasoning explanations. Design graceful failure handling. Allow users to control autonomy levels.

Governance (1→3): Implement the AI governance framework — prompt injection defenses, audit logging, data handling policies, and incident response procedures.

For organizations seeking structured improvement, The Thinking Company’s AI readiness assessment evaluates current state across all twelve dimensions and produces a prioritized improvement roadmap. The AI Build Sprint (EUR 50-80K, 4-6 weeks) addresses the top architectural improvement priorities, while the AI Product Build (EUR 200-400K+, 3-6 months) delivers comprehensive improvement across all dimensions.


Frequently Asked Questions

How often should I run the AI product evaluation?

Conduct a full 12-dimension evaluation quarterly. Between full evaluations, monitor the dimensions most relevant to your current improvement priorities on a monthly basis. BCG’s research shows that organizations with quarterly evaluation cycles improve their composite score 2.3x faster than those evaluating annually. [Source: BCG, 2025] The evaluation takes 2-4 hours with a cross-functional team. Track scores over time to visualize improvement trends and identify dimensions that have stalled.

What is a good composite score for an AI product in 2026?

The median score for products claiming AI capabilities is 26 — firmly in the “AI-enhanced” category. The median for genuinely AI-native products is 38. Products scoring above 45 are in the top 10% of the market. [Source: BCG, “AI Product Assessment Benchmark,” 2025] Do not aim for a perfect 60 — focus on achieving scores of 4+ on the dimensions most critical to your competitive strategy and your users’ needs. Use the AI maturity model to contextualize your product score within your organization’s overall AI capability.

Can this framework be used to compare products across different categories?

The composite score enables cross-category comparison, but dimension-level scores are more meaningful for within-category benchmarking. A developer tool and a customer support platform may have the same composite score but very different dimension profiles. Compare dimension scores within categories and composite scores across categories. Weight dimensions differently based on category-specific requirements — governance matters more in healthcare than in consumer creative tools.

Which dimension should I improve first?

Improve the dimension with the highest gap between current score and strategic importance. If you are building competitive moat, prioritize Data Flywheel (D2) and Improvement Velocity (D12). If you are pursuing enterprise sales, prioritize Governance (D10) and Trust Architecture (D9). If you are launching a new product, prioritize Task Performance (D4) and UX Design (D7). The building AI-native products guide provides a development sequence that naturally builds scores across dimensions.

How do I score dimensions when I do not have production data?

For pre-launch products, score based on architecture decisions and evaluation framework results rather than production metrics. Dimension 2 (Data Flywheel) should be scored on whether the architecture supports feedback capture, even if data has not accumulated yet. Dimension 4 (Task Performance) should be scored against evaluation benchmarks, noting that production performance typically underperforms evaluation benchmarks by 15-25%. [Source: Stack Overflow, 2025] Flag all pre-launch scores as provisional and re-evaluate within 90 days of launch.

Does a higher autonomy score always mean a better product?

No. The appropriate autonomy level depends on the product’s domain, risk profile, and user expectations. A healthcare clinical decision support tool scoring 2 on autonomy (copilot level) may be optimally designed for its regulatory environment, while a code generation tool scoring 2 may be leaving significant value on the table. Evaluate autonomy against the copilot-to-agent framework to determine the appropriate target level for your specific product category.

How do I benchmark against competitors when I cannot access their internals?

Evaluate competitor products from the user perspective. You can assess Integration Depth (does the product function without AI?), UX Design (is the interface AI-native?), Task Performance (how well does it handle your tasks?), Trust Architecture (how does it handle failures?), and Autonomy Level (how much does the AI do independently?). Internal dimensions like Model Architecture, Cost Efficiency, and Data Flywheel require inference from public information — pricing models, improvement rates, and technical blog posts. Score these dimensions with lower confidence and note assumptions.

Is this framework applicable to internal AI products (not sold externally)?

Yes, with two adjustments. Replace Cost Efficiency (D11) with Resource Efficiency (does the product use organizational resources proportional to value delivered?). Replace the commercial lens on Governance (D10) with an internal compliance lens (does the product meet your organization’s internal AI policies?). The remaining dimensions apply directly. Internal products often score lower on Trust Architecture and UX Design because internal users have higher tolerance for rough edges — but this tolerance erodes as external AI-native tools set higher UX expectations.