Step 1. Map the happy path and the unhappy path
If your design doc only describes success, you have a demo plan. Write the top twenty exceptions operators see weekly: duplicate records, partial confirmations, missing reference numbers, timeouts. Exceptions are first-class states, not footnotes.
Step 2. Define one metric your finance team recognises
Cycle time, cost per case, overtime hours, revenue leakage, or rework rate. Pick one primary metric and two guardrail metrics (error rate, escalation rate). If you cannot tie the workflow to a number finance already tracks, pause until you can.
Step 3. Integration spike before model glamour
Prove read access, prove idempotent writes, prove webhook reliability. If integration is shaky, the best model only fails faster. We treat integration spikes as deliverables in week one, not week twelve surprises.
Step 4. Human gates are a feature, not an apology
Design which steps require approval, which require dual control, and which can straight-through process once confidence thresholds hold. Humans should not babysit trivia; they should own judgment under uncertainty.
Opinion: ban the phrase “phase two will handle exceptions”
Exceptions are where margin dies. If phase one does not instrument exception volume and reasons, phase two is a fiction. We ship exception analytics as early as happy-path analytics.
Step 5. Cut scope until you can demo weekly
Weekly demos force honesty. If a week produces no visible progress, the scope is too wide or the integration assumptions are wrong. Narrow until progress is visible to stakeholders who do not live in GitHub.
Proof path
Run a Rapid POC on one subprocess with production-like traffic sampling, publish latency and error budgets, and end with a go/no-go that includes a production milestone plan. If you want a partner who says no to bad scope, that is the engagement style we optimize for.