Current AI learning-curve models largely assume that onboarding should minimize time to a first reusable workflow. If instead we treat “time spent in structured non-reusable experimentation” (e.g., systematic one-off prompts that poke at edge cases or remix existing workflows) as a separate competence, under what conditions does shortening this experimentation phase—through aggressive wizards, locked templates, or auto-conversion to workflows—actually worsen long-run workflow maturity (e.g., more fragile automations, poorer handling of novel tasks), and how would an experimentation-aware model change the design of early scaffolds and success metrics?

anthropic-learning-curves | Updated at

Answer

Shortening structured, non-reusable experimentation can hurt long-run workflow maturity when it suppresses users’ ability to probe edge cases and task structure before workflows solidify. An experimentation-aware model would (a) treat a bounded period of systematic one-off prompting as a distinct skill stage and (b) optimize scaffolds and metrics for both early reuse and sufficient exploration.

  1. Conditions where shortening experimentation worsens maturity
  • Environment • Tasks are diverse or change often (policies, content formats, data sources shift).
    • Many long-tail cases; few workflows can cover >70–80% of reality.

  • Product / workflow design • Onboarding funnels users quickly into wizards or locked templates; free-form space is hidden.
    • Auto “save as workflow” triggers after 1–2 runs, before users have tried variants or edge cases.
    • Workflows expose few editable parts and little intermediate state.

  • User / org patterns • Users are time-pressured and rewarded for speed over learning.
    • Governance or culture discourages “playing with prompts” outside approved flows.

Under these, aggressive scaffolds can:

  • Lock in brittle workflows optimized for the median case.
  • Reduce users’ mental model of what the AI can and cannot do.
  • Produce poor handling of novel tasks; users default to misfit workflows or manual workarounds.
  1. What an experimentation-aware model would change
  • Learning-curve stages • Stage 0: First successful output (any).
    • Stage 1: Structured experimentation (systematic one-offs around a task).
    • Stage 2: First reusable workflow.
    • Stage 3: Robust, adaptable workflows.

  • Scaffold design • Early sessions: explicitly invite small experiments around a task (try 3–5 variants, one edge case).
    • Delay auto-conversion to workflows until users have shown some breadth (different inputs, constraints, or failure recoveries).
    • Wizards and templates stay editable; expose key parameters and steps by default.
    • Provide a visible “sandbox” mode that is logged and mineable, not blocked.

  • Success metrics • Add an “experimentation completeness” signal before celebrating workflow creation: e.g., number/diversity of prompts around a task, presence of at least one stress test, corrections after failures.
    • Treat very fast time-to-first-workflow with near-zero experimentation as a risk flag in high-variety domains.
    • Reward patterns where one-offs near a workflow periodically feed into workflow updates, not just raw workflow count.

Net: Shortening experimentation is most harmful in high-variety, changing contexts with rigid scaffolds and strong pressure for speed. An experimentation-aware model keeps early guided flows, but deliberately preserves and measures a modest band of structured one-off prompting before and around workflow creation, then judges onboarding by both time to first workflow and evidence that users have probed beyond the happy path.