Current AI learning-curve models largely assume that onboarding should minimize time to a first reusable workflow. If instead we treat “time spent in structured non-reusable experimentation” (e.g., systematic one-off prompts that poke at edge cases or remix existing workflows) as a separate competence, under what conditions does shortening this experimentation phase—through aggressive wizards, locked templates, or auto-conversion to workflows—actually worsen long-run workflow maturity (e.g., more fragile automations, poorer handling of novel tasks), and how would an experimentation-aware model change the design of early scaffolds and success metrics?
anthropic-learning-curves | Updated at
Answer
Shortening structured, non-reusable experimentation can hurt long-run workflow maturity when it suppresses users’ ability to probe edge cases and task structure before workflows solidify. An experimentation-aware model would (a) treat a bounded period of systematic one-off prompting as a distinct skill stage and (b) optimize scaffolds and metrics for both early reuse and sufficient exploration.
- Conditions where shortening experimentation worsens maturity
-
Environment • Tasks are diverse or change often (policies, content formats, data sources shift).
• Many long-tail cases; few workflows can cover >70–80% of reality. -
Product / workflow design • Onboarding funnels users quickly into wizards or locked templates; free-form space is hidden.
• Auto “save as workflow” triggers after 1–2 runs, before users have tried variants or edge cases.
• Workflows expose few editable parts and little intermediate state. -
User / org patterns • Users are time-pressured and rewarded for speed over learning.
• Governance or culture discourages “playing with prompts” outside approved flows.
Under these, aggressive scaffolds can:
- Lock in brittle workflows optimized for the median case.
- Reduce users’ mental model of what the AI can and cannot do.
- Produce poor handling of novel tasks; users default to misfit workflows or manual workarounds.
- What an experimentation-aware model would change
-
Learning-curve stages • Stage 0: First successful output (any).
• Stage 1: Structured experimentation (systematic one-offs around a task).
• Stage 2: First reusable workflow.
• Stage 3: Robust, adaptable workflows. -
Scaffold design • Early sessions: explicitly invite small experiments around a task (try 3–5 variants, one edge case).
• Delay auto-conversion to workflows until users have shown some breadth (different inputs, constraints, or failure recoveries).
• Wizards and templates stay editable; expose key parameters and steps by default.
• Provide a visible “sandbox” mode that is logged and mineable, not blocked. -
Success metrics • Add an “experimentation completeness” signal before celebrating workflow creation: e.g., number/diversity of prompts around a task, presence of at least one stress test, corrections after failures.
• Treat very fast time-to-first-workflow with near-zero experimentation as a risk flag in high-variety domains.
• Reward patterns where one-offs near a workflow periodically feed into workflow updates, not just raw workflow count.
Net: Shortening experimentation is most harmful in high-variety, changing contexts with rigid scaffolds and strong pressure for speed. An experimentation-aware model keeps early guided flows, but deliberately preserves and measures a modest band of structured one-off prompting before and around workflow creation, then judges onboarding by both time to first workflow and evidence that users have probed beyond the happy path.