Why 90 Days Matters for Production AI
Most organisations take 18–24 months to move an AI pilot into production. The gap isn't technical incompetence. It's process bloat, unclear ownership, and a false belief that governance slows you down. At Brightlume, we've shipped 85%+ of pilots to production within 90 days because we've inverted that logic: tight governance accelerates delivery.
The 90-day sprint isn't a marketing timeline. It's a deliberate compression of decision cycles, rapid validation loops, and ruthless scope management. You still run evals. You still build monitoring. You still document architecture. You just don't wait for perfect. You ship defensible.
This breakdown assumes you've already identified a pilot use case—a workflow, decision, or interaction that costs time or money today. If you haven't, start with 7 Signs Your Business Is Ready for AI Automation to validate your target first.
Weeks 1–3: Problem Framing and Feasibility Sprint
Week 1: Define the Outcome, Not the Solution
Day one: stop talking about models. Start talking about economics.
Your first week is ruthlessly focused on answering three questions:
-
What decision or action does this workflow enable today? Be specific. Not "improve customer service." Try "reduce average handling time on refund requests from 12 minutes to 4 minutes" or "reduce clinical triage time from 15 minutes to 3 minutes for low-risk cases."
-
What's the cost of the current state? Labour hours, error rate, latency, customer churn. Quantify it. If you can't measure it, you can't prove ROI, and you won't get budget for production.
-
What's the failure mode? Not all errors are equal. A hallucinated diagnosis in a health system is a patient safety event. A hallucinated email subject line is an annoyance. Define your risk threshold now, because it drives your architecture later.
This is where A Practical 90-Day AI Readiness Roadmap Built on Lean, Clarity and Intentional Transformation becomes essential—clarity on problem framing directly accelerates execution. Spend three days mapping the current workflow end-to-end. Interview the humans doing it. Record latency, error rates, decision points. This becomes your baseline for measuring pilot success.
By end of week one, you should have: a one-page problem statement, baseline metrics, and a signed-off definition of success. If you can't fit it on one page, you don't understand the problem yet.
Week 2: Technical Feasibility and Model Selection
Now you bring in the engineers. Week two is about validating that AI can actually solve this problem at the cost and latency you need.
Start with a rapid feasibility assessment. Can this workflow be decomposed into steps an LLM can execute? Does it require real-time data retrieval, or is batch processing sufficient? What's your latency budget? A hospital triage agent needs sub-second responses. A batch claims processor can tolerate 5-minute latency.
Model selection happens here, but it's not about picking the "best" model. It's about picking the right model for your constraints. Claude Opus 4 has superior reasoning for complex decision-making but costs more per token than GPT-4o. Gemini 2.0 excels at multimodal tasks. Smaller models like Llama 3.1 run on-premises for compliance-heavy industries. Your choice depends on:
- Latency requirements: Sub-100ms? You might need a smaller model or edge deployment.
- Cost per inference: Running 10,000 inferences daily changes your model economics entirely.
- Reasoning complexity: Simple classification? Smaller model. Multi-step reasoning with tool use? Larger model.
- Data residency: Financial services and health systems often can't send data to US-hosted APIs. On-premises or regional deployment becomes mandatory.
Run a proof-of-concept on 50–100 real examples from your workflow. Don't use synthetic data. Real data reveals edge cases, format inconsistencies, and domain-specific language that benchmarks miss. Test Claude Opus 4 and GPT-4o side-by-side if cost allows. Measure accuracy, latency, and token consumption. This takes 3–4 days and costs under $200.
By end of week two, you should have: a selected model, validated accuracy on real data, latency measurements, and a rough cost-per-inference estimate. You should also have a clear list of failure modes and edge cases that your eval strategy needs to catch.
Week 3: Architecture and Governance Scaffolding
Week three is about designing the system that lets you move fast without breaking things.
Start with architecture. Most production AI workflows follow a pattern: ingest → validate → enrich → decide → act → monitor. Your job is to map your specific workflow onto this pattern and identify where humans stay in the loop.
For a claims processing agent, that might look like:
- Ingest: PDF claim arrives via email or API.
- Validate: Extract structured fields, flag format errors.
- Enrich: Query claims history, policy database, fraud rules.
- Decide: LLM classifies as approve, deny, or escalate. If escalate, route to human.
- Act: Update CRM, trigger payment or request documentation.
- Monitor: Log all decisions, track approval rate, measure downstream claim denials.
For a clinical triage agent in a health system:
- Ingest: Patient presents with symptoms via chat or kiosk.
- Validate: Capture vital signs, medication list, allergy status.
- Enrich: Query medical records, clinical guidelines, recent test results.
- Decide: LLM recommends triage level (self-care, urgent care, ED, admit). If uncertain, escalate to nurse.
- Act: Generate visit summary, route to appropriate care level, notify staff.
- Monitor: Track triage accuracy, compare to gold-standard clinician assessments, measure downstream ED utilisation.
Governance isn't bureaucracy at this stage. It's the infrastructure that lets you move fast without creating compliance debt. Define:
-
Eval strategy: How will you measure accuracy in production? For a claims agent, that might be a sample of 50 decisions per week reviewed by a claims specialist. For a triage agent, compare to nurse assessment. Build this into your architecture from day one—don't bolt it on later.
-
Escalation logic: When does the agent defer to a human? If confidence below 85%? If the case matches a specific pattern? Make this explicit. Ambiguous escalation creates bottlenecks.
-
Audit trail: Every decision the agent makes must be logged with the input, reasoning, and output. This is non-negotiable for regulated industries. Build it into the prompt or application layer.
-
Rollback plan: If accuracy drops, can you revert to the previous agent version or fall back to human processing? Design for graceful degradation.
This is also where you understand AI Agents as Digital Coworkers: The New Operating Model for Lean Teams—your agent isn't replacing humans; it's augmenting them. Design the handoff points explicitly.
By end of week three, you should have: a data flow diagram, a list of integration points (APIs, databases, systems), a defined eval strategy, and a first draft of your monitoring dashboard. You should also have a security review checklist—data residency, access controls, encryption, audit logging.
Weeks 4–6: Prototype Build and Rapid Iteration
Week 4: MVP Agent Development
Week four is where code ships. You're building a minimal viable agent that handles 80% of cases and escalates the rest.
Start with the happy path. What's the simplest version of your workflow that delivers measurable value? For a claims agent, that might be: ingest claim, extract fields, check against simple rules (claim amount < threshold, no duplicate in last 30 days), approve or escalate. Don't build the 20% of edge cases yet.
Choose your framework. If you're using Claude Opus 4, Anthropic's Agentic SDK or LangChain give you tool use and extended thinking out of the box. If you're building on GPT-4o, LangChain or LlamaIndex work well. If you're deploying on-premises, consider vLLM for model serving and LangChain for orchestration. The choice matters for latency and cost, but don't over-engineer. A simple Python FastAPI app calling an LLM API can be production-ready if you build monitoring correctly.
Implement tool use from day one. Your agent should call APIs, query databases, and execute actions—not just generate text. This is where AI Agents That Write and Execute Code: When to Use Them becomes relevant. If your agent needs to query complex data or run calculations, code execution (sandboxed, with timeout limits) is safer than trying to do it with prompts alone.
Build your eval harness in parallel. Create a test set of 100–200 real examples from your workflow. For each, record the expected output. Run your agent against the test set daily. Track accuracy, latency, and cost. If accuracy drops below your threshold, investigate immediately. This is your safety net.
By end of week four, you should have: a working agent that handles happy-path cases, a test harness with daily evals, and latency/cost metrics. You should also have a clear list of failure modes from your test set—these become your priorities for week five.
Week 5: Edge Case Handling and Accuracy Tuning
Week five is about closing the gap between your test set accuracy and production accuracy.
Take the 20% of cases your agent failed on in week four. Categorise them: ambiguous input format, missing context, domain-specific logic, hallucination, tool call error. For each category, decide: fix in the prompt, add a tool, add a validation step, or escalate to human?
Prompt engineering is real engineering. Test systematically. If your agent is misclassifying claims with missing line items, try:
- Adding an explicit validation step: "Before deciding, confirm all line items are present."
- Adding context: "If line items are missing, escalate to human review."
- Adding examples: Include 3–5 examples of correctly handled claims with missing data in your prompt.
Measure the impact of each change on your test set. A 2% accuracy improvement that adds 500ms latency might not be worth it. A 5% improvement that adds 50ms might be.
This is where The 90-Day AI Launch Sprint — AI Product Development framework becomes practical—your iteration cycles need to be tight and measurable. Run daily evals. Deploy changes to staging, test, then promote. Don't wait for weekly releases.
Also, start building your monitoring dashboard. What metrics matter? For a claims agent: approval rate, escalation rate, average processing time, cost per claim, downstream denials (claims approved by agent but denied at payment stage). For a triage agent: triage accuracy vs. nurse assessment, escalation rate, patient satisfaction, ED utilisation. These metrics tell you if your agent is working in production, not just in testing.
By end of week five, you should have: >90% accuracy on your test set, latency and cost within budget, and a monitoring dashboard ready for launch. You should also have clear documentation of when the agent escalates and why.
Week 6: Security, Compliance, and Staging Validation
Week six is about making sure your agent doesn't break anything when it goes live.
Run a security review. If you're handling customer data, PII, or health information, you need:
- Data residency: Is data leaving your region? If you're in Australia and using a US-hosted API, that might violate data residency requirements. Consider regional endpoints or on-premises deployment.
- Access controls: Who can trigger the agent? What data can it access? Use role-based access and least-privilege principles.
- Encryption: Data in transit (TLS) and at rest (database encryption). Non-negotiable.
- Audit logging: Every decision logged with timestamp, user, input, output. Immutable logs for compliance.
- Secrets management: API keys, database credentials stored securely, rotated regularly. Use a secrets vault, not environment variables.
Run a compliance check. If you're in financial services, check against AML/KYC requirements. If you're in health, check against privacy regulations (Australian Privacy Act, HIPAA if applicable). If you're in hospitality, check data protection and guest privacy. Document your compliance approach—this becomes your audit trail.
Deploy to staging with production-like data volumes and latency. Run your agent against a week's worth of real data in a non-production environment. Does it handle the load? Does latency stay within budget? Do any edge cases emerge? Fix them now, not after launch.
Also, brief your ops team. Who monitors the agent? Who escalates if accuracy drops? What's the incident response process? If the agent starts making bad decisions, can you kill it and fall back to manual processing in under 5 minutes? Design for operational resilience.
By end of week six, you should have: a security review completed, compliance documented, staging validation passed, and ops procedures documented. You're ready for a controlled production launch.
Weeks 7–9: Production Launch and Optimisation
Week 7: Controlled Rollout and Live Monitoring
Week seven is launch week, but not a big bang. You're rolling out to a subset of your workflow, monitoring closely, and ready to rollback.
Start with a canary deployment. If you process 1,000 claims per day, run your agent on 100 claims for day one. Measure accuracy, latency, cost. If all metrics are green, increase to 250 claims on day two, 500 on day three, 1,000 on day four. If anything goes wrong, rollback to 100% human processing in under 5 minutes.
Monitoring is critical. You need real-time dashboards showing:
- Accuracy: Daily sample review. If you're processing 1,000 claims, sample 50 per day and have a domain expert review them. Track approval rate, escalation rate, error rate.
- Latency: P50, P95, P99 latency. If P99 latency spikes, investigate immediately.
- Cost: Cost per inference, total daily cost. If it's trending above budget, investigate model choice or prompt efficiency.
- Business metrics: For claims, track downstream denials (are approved claims being denied at payment?). For triage, track ED utilisation (is the agent triaging appropriately?).
Set up alerting. If accuracy drops below 85%, page someone. If latency P99 exceeds 2 seconds, page someone. If daily cost exceeds budget, page someone. Don't wait for weekly reviews.
This is where understanding AI Agent Orchestration: Managing Multiple Agents in Production matters—if you're running multiple agents (claims + fraud detection, for example), you need orchestration logic to route work correctly and handle failures gracefully.
By end of week seven, you should have: 100% of your workflow running through the agent, green metrics across the board, and zero production incidents. You should also have a clear process for daily eval sampling and rapid iteration.
Week 8: Accuracy Optimisation and Cost Reduction
Week eight is about squeezing every percentage point of accuracy and every dollar of cost.
Review your daily evals from week seven. Find the 10–15 cases your agent got wrong. Categorise them again. Are they solvable with prompt tweaks, or do they require architectural changes?
For prompt optimisation, try:
- Chain-of-thought: Ask the agent to reason step-by-step before deciding. This often improves accuracy on complex cases.
- Few-shot examples: Add 5–10 examples of correctly handled edge cases to your prompt.
- Explicit constraints: "You must always check for duplicate claims before approving" or "If confidence < 80%, escalate to human."
Measure the impact. A 1% accuracy improvement that costs 10% more in tokens isn't worth it. A 5% improvement that costs 2% more is.
For cost reduction, consider:
- Model downgrade: If your test set shows that GPT-4o achieves 92% accuracy and Claude 3 Haiku achieves 90%, and Haiku costs 80% less, switch to Haiku. The 2% accuracy loss might be acceptable.
- Prompt compression: Remove unnecessary context. Shorter prompts = fewer tokens = lower cost.
- Batch processing: If your workflow allows, batch 10 claims into a single API call instead of 10 separate calls. Reduces overhead.
- Caching: If you're querying the same policy database repeatedly, cache the results. Fewer API calls = lower cost.
By end of week eight, you should have: accuracy at or above your target (90%+), cost optimised without sacrificing accuracy, and a clear understanding of your cost-per-inference economics.
Week 9: Handoff and Scaling Plan
Week nine is about handing off to your ops team and planning for scale.
Document everything. Your ops team needs to understand:
- How the agent works: A one-page architecture diagram, not a 50-page technical specification.
- How to monitor it: What dashboards to watch, what alerts matter, what to do if accuracy drops.
- How to escalate: Who do they call if something breaks? What's the incident response process?
- How to iterate: If you want to improve the agent, what's the process? How do you test changes safely?
Create a runbook. Step-by-step instructions for common scenarios: accuracy drops, latency spikes, cost overrun, agent starts making a specific type of error, rollback to previous version, emergency fallback to 100% human processing.
Plan for scale. If this agent is working, you probably want to expand it. Do you apply it to other claim types? Other workflows? Other business units? Map out the roadmap. What changes do you need to make to scale? Do you need to handle higher volume? More complex decision-making? Multiple languages? Plan for it now.
Also, measure ROI. You set a baseline in week one: "reduce average handling time from 12 minutes to 4 minutes." Are you hitting that? What's the financial impact? If you're processing 1,000 claims per day and saving 8 minutes per claim, that's 133 hours per day. At $50/hour labour cost, that's $6,650/day in savings. Compare that to your infrastructure and model costs. Is it positive? By how much?
By end of week nine, you should have: ops team trained, runbook documented, scaling roadmap defined, and ROI measured and communicated to leadership.
Critical Success Factors: What Actually Matters
Governance Accelerates, Not Slows
The biggest misconception is that governance is a brake on speed. At Brightlume, we've found the opposite: clear governance accelerates delivery because it eliminates ambiguity and rework.
When you define your eval strategy in week three, you eliminate the week-eight panic of "wait, how do we know if this is working?" When you define escalation logic upfront, you eliminate the week-six conversation about "what does the agent do when it's uncertain?" When you document compliance requirements early, you avoid the week-eight discovery that you need to redesign your architecture.
Governance that works is lightweight and automated. A daily eval harness that automatically flags accuracy drops is governance. A monitoring dashboard that alerts when latency spikes is governance. A runbook that tells ops exactly what to do when something breaks is governance. These things make you faster, not slower.
Metrics Drive Decisions
Every decision in your sprint should be driven by metrics, not opinions. Your model choice should be driven by accuracy, latency, and cost on real data—not marketing claims. Your prompt tweaks should be driven by eval results—not gut feel. Your rollout strategy should be driven by monitoring data—not confidence.
This is why you build your eval harness in week four, not week eight. This is why you run daily evals, not weekly. This is why you have clear success metrics from day one.
Escalation Is Not Failure
If your agent escalates 20% of cases to humans, that's not a failure. That's the agent doing its job. It's identifying cases where it's uncertain and deferring to human judgment. That's safer than the agent forcing a decision it's not confident about.
Design your escalation logic to be explicit and measurable. "If confidence < 85%, escalate" is clear. "Escalate if it feels uncertain" is not. Track your escalation rate. If it's trending up, investigate why. If it's trending down, that's good—your agent is getting more confident over time.
Speed Requires Ruthless Scope Management
You can't build the perfect agent in 90 days. You can build a good agent that handles 80% of cases and escalates the rest. That's the trade-off.
Every feature request, every edge case, every "nice to have" is a tax on your timeline. Track them in a backlog. After launch, prioritise based on impact and effort. Some of them won't be worth building—the 2% of cases they'd handle aren't worth the engineering effort.
Production Readiness Isn't Optional
Your agent needs to be production-ready from day one. That means:
- Monitoring: You can see what it's doing in real-time.
- Logging: Every decision is auditable.
- Rollback: You can revert to the previous version in minutes.
- Scaling: It can handle 10x the volume without breaking.
- Security: Data is encrypted, access is controlled, compliance is documented.
These aren't nice-to-haves. They're prerequisites. Build them into your architecture in week three. Don't bolt them on in week eight.
Putting It Together: A Real Example
Let's walk through a health system deploying a clinical triage agent.
Week 1: Define outcome. "Reduce average triage time from 15 minutes to 3 minutes for low-risk presentations. Improve triage accuracy to match nurse assessment 95% of the time. Escalate any case where confidence < 90% to nurse review."
Week 2: Feasibility test. Test Claude Opus 4 on 100 real patient presentations. Accuracy: 88%. Latency: 1.2 seconds. Cost: $0.03 per triage. Failure modes: missing vital signs (10 cases), ambiguous symptom descriptions (7 cases).
Week 3: Architecture. Design: ingest presentation → validate vital signs and medical history → query clinical guidelines → Claude Opus 4 triage → if confidence < 90%, escalate to nurse → log decision with reasoning for audit.
Week 4: MVP build. Implement agent with basic triage logic. Test set accuracy: 87%.
Week 5: Improve accuracy. Add validation for missing vitals. Add few-shot examples of ambiguous presentations. Add explicit constraint: "If any vital sign is missing, escalate." Test set accuracy: 92%.
Week 6: Security and compliance. Review against Privacy Act. Implement audit logging. Deploy to staging. Run against a week of real presentations. All metrics green.
Week 7: Canary rollout. Run on 10% of presentations day one, 25% day two, 50% day three, 100% day four. Monitor accuracy, escalation rate, patient satisfaction. All green.
Week 8: Optimisation. Review 50 escalated cases. 30% are escalations where confidence was < 90% (correct behaviour). 15% are cases where the agent was uncertain but accuracy was high (lower the threshold to 80%). 5% are genuine failures (add more examples to prompt). Retrain. Accuracy now 94%.
Week 9: Handoff. Document for nursing team. Create runbook. Measure impact: triage time down from 15 to 3 minutes. Accuracy 94% vs. 95% target (acceptable). Escalation rate 12% (appropriate for safety-critical domain). ROI: 12 nurses saved 12 minutes per presentation × 200 presentations/day = 2,400 nurse-minutes saved per day = 40 hours/day = 2 FTE saved. At $80k/year per FTE, that's $160k annual savings.
That's the 90-day sprint in action.
Beyond 90 Days: Continuous Improvement
Your agent doesn't stop improving after launch. It evolves based on production data.
Set up a continuous improvement process. Every week, sample 50 decisions and have a domain expert review them. Categorise errors. If you see a pattern (e.g., the agent consistently mishandles a specific type of case), add it to your prompt or architecture. Retrain, test, deploy.
Also, monitor for model updates. When Anthropic releases Claude Opus 5, test it against your current model. Does it improve accuracy? Does it cost less? If yes, upgrade. This isn't a big project—it's a day of testing and a deployment.
For deeper insights into scaling agentic workflows across your organisation, explore AI Agent Orchestration: Managing Multiple Agents in Production. As your agent portfolio grows, orchestration becomes critical.
Also, revisit your success metrics quarterly. Are they still relevant? If your triage agent has been running for 6 months and accuracy is consistently 95%, maybe your target should be 97%. If your claims agent has reduced processing time by 80%, what's next? Can you apply it to other claim types?
Why Brightlume's Approach Works
Brightlume ships production AI in 90 days because we've inverted the typical approach. Most consultancies spend months on strategy and planning, then months on building, then months on deployment. We compress all three into a tight, iterative cycle.
We do this by:
-
Starting with outcomes, not technology: We ask "what decision or action are we automating?" before we ask "what model should we use?"
-
Building governance into the architecture, not bolting it on later: Monitoring, logging, escalation, evals—these are part of the MVP, not post-launch additions.
-
Measuring relentlessly: Every decision is driven by metrics on real data, not opinions or benchmarks.
-
Treating the first 90 days as a sprint, not a project: We iterate weekly, not monthly. We deploy daily, not quarterly. We escalate problems in hours, not days.
-
Handing off to your team, not staying on as advisors: We document everything, train your ops team, and leave you with a clear roadmap for scale. You own the agent after day 90.
If you're ready to move a pilot to production, explore Brightlume's capabilities. If you want to understand how AI agents differ from copilots and what your organisation actually needs, read Agentic AI vs Copilots: What's the Difference and Which Do You Need?
The 90-day sprint is achievable. You just need clear outcomes, ruthless scope management, and tight governance. Everything else follows.
Key Takeaways
- Weeks 1–3: Define outcomes, validate feasibility, design architecture with governance built in.
- Weeks 4–6: Build MVP, iterate on accuracy, validate in staging.
- Weeks 7–9: Controlled rollout, optimise cost and accuracy, hand off to ops.
- Governance accelerates: Clear eval strategy, monitoring, logging, and escalation logic make you faster, not slower.
- Metrics drive decisions: Every choice (model, prompt, architecture) should be validated on real data.
- Escalation is not failure: An agent that confidently handles 80% and escalates 20% is better than one that forces decisions on 100%.
- Production readiness is mandatory: Build monitoring, logging, rollback, and security from day one.
- Scale comes after launch: Prove the model works on one workflow, then expand to others.
For more on building AI-native organisations and moving from pilots to production at scale, check out AI-Native Companies Don't Have IT Departments — They Have AI Departments and AI Automation Maturity Model: Where Is Your Organisation?
The path from pilot to production is well-trodden. You just need to walk it fast, with eyes open, and with clear metrics at every step. Ninety days is achievable. Let's ship.