Current work treats patterns like repeated near‑misses, last‑mile manual rewrites, and step bypassing as generic signals of an impending productivity plateau. If instead we distinguish mechanism-specific plateau types (e.g., “decomposition plateau,” “trust/compliance plateau,” “parameterization plateau”), which concrete behavior bundles within a single product (edits, rollbacks, variant creation, error acknowledgments) most reliably classify users into these plateau types, and how does routing different in-product interventions to each type change long-run workflow maturity compared with today’s one-size-fits-all nudges?

anthropic-learning-curves | Updated at

Answer

Mechanism-specific plateau types can be inferred from compact bundles of in-product behaviors; routing targeted interventions by type is likely better than generic nudges, especially for users past first reusable workflows.

  1. Plateau types and behavior bundles (single product)

A) Decomposition plateau (task structure off)

  • Signals (bundle): • High rate of step bypass/disable on the same step.
    • Frequent rollbacks or re-runs starting from that step.
    • Variant creation that mainly changes step order/adds/removes steps, not wording.
    • Low rate of small text edits inside the problematic step.
  • Interpretation: user can’t get the step to “fit” the task; structure, not phrasing, is wrong.

B) Parameterization plateau (brittle prompts/templates)

  • Signals: • Many near-identical variants differing in a few fields (tone, audience, market, product).
    • Repeated, patterned edits in the same locations (e.g., always changing tone, length, region).
    • Few rollbacks; outputs are “close enough” but need routine tweaks.
    • Error acknowledgments low; user rarely flags outputs as flatly wrong.
  • Interpretation: underlying workflow is fine, but knobs are implicit in text instead of explicit params.

C) Trust/compliance plateau (last-mile rewrite)

  • Signals: • Large manual rewrites after final AI step (export → heavy editing) with little in-product editing.
    • Frequent error or risk acknowledgments (e.g., “too risky,” “too salesy,” “missing required phrasing”).
    • Low variant creation; user keeps structure but won’t ship AI output as-is.
    • Occasional step-level disable for “sensitive” segments.
  • Interpretation: user doesn’t trust the model to meet quality/risk norms at the output boundary.

D) Coverage plateau (narrow usage of a solid flow)

  • Signals: • One or two workflows dominate runs; almost no new workflows started.
    • Very low edit/rollback rates; high stability.
    • Little variant creation despite diverse inputs or tasks nearby.
    • Few explicit errors; user seems satisfied but isn’t expanding scope.
  • Interpretation: flow works, but user isn’t generalizing it to adjacent tasks.
  1. Routing interventions by plateau type

A) Decomposition plateau → structure tools

  • Interventions: • Auto-suggested alternative decompositions when bypass+rollback cluster on a step.
    • Inline “split/merge this step” micro-wizard seeded from common patterns.
    • Quick prompt to mark the sub-task that’s being done manually (“what did you do instead?”) and scaffold a new step.
  • Expected effect: fewer bypasses, more steps owned by AI → higher workflow maturity via broader coverage.

B) Parameterization plateau → param surfacing

  • Interventions: • Auto-detect patterned edits across runs and suggest named parameters (e.g., tone, audience).
    • Lightweight “turn these 3 edits into knobs” prompt after N similar changes.
    • Consolidation nudge when many near-duplicate variants exist (“merge into one parametric workflow”).
  • Expected effect: fewer brittle variants, faster reuse across contexts → higher maturity via adaptable templates.

C) Trust/compliance plateau → guardrails + review

  • Interventions: • Final-step controls: stricter style/compliance presets (“legal-safe,” “brand voice”) triggered by frequent last-mile rewrites.
    • Optional human-in-the-loop review mode: AI drafts + checklists or side-by-side examples of “approved” vs “unacceptable.”
    • Simple explanation prompts (“why this phrasing?”) or checkboxes for must-have constraints that get baked into prompts.
  • Expected effect: higher ship rate of AI outputs, reduced manual rewrite time → more durable delegation at last mile.

D) Coverage plateau → exploration prompts

  • Interventions: • Occasion-based nudges: when the same workflow runs at high cadence with no edits, suggest “adjacent tasks” templates.
    • In-run question: “What similar tasks do you still do manually?” with one-click clones seeded from the existing workflow.
    • Very light gallery surfacing tied to current inputs.
  • Expected effect: more workflows created around a strong core; broader portfolio without forcing deep structural changes.
  1. Impact versus one-size-fits-all nudges
  • With generic nudges (today): • All plateau signals trigger the same responses (e.g., “try editing the prompt,” “explore templates”).
    • Users at decomposition plateaus keep tweaking wording; parameterization plateaus keep spawning variants; trust plateaus keep rewriting outside the product.

  • With mechanism-specific routing: • Decomposition plateau: structural tools directly attack the main bottleneck; early data from analogous UX (e.g., step editors) suggests larger coverage gains per user than generic copy advice.
    • Parameterization plateau: param suggestions reduce edit repetition and variant sprawl, which prior context links to higher workflow maturity.
    • Trust/compliance plateau: guardrail presets and review hooks mainly increase ship rate and delegation depth, not just usage count.
    • Coverage plateau: exploration cues promote new workflows without destabilizing existing ones.

Overall expectation: for users who already have at least one reusable workflow, mechanism-specific interventions should lift long-run workflow maturity more than uniform nudges, mainly by reducing repeated friction loops and converting local workarounds into structural improvements.