Skip to content
Resources

Scoping custom AI projects: 5 questions that prevent scope creep

By Gosai Digital·January 2026·7 min read
Share:

Custom AI projects fail when teams start building before answering who owns the outcome, what decision improves, and what happens when it breaks.

By Gosai Digital · January 2026 · Based on 40+ enterprise AI engagements

Custom AI projects fail when teams start building before answering who owns the outcome, what decision improves, and what happens when it breaks.


TL;DR

  • Most AI pilots die in a 12-month limbo because nobody answered the hard questions upfront.
  • "Data quality" is usually an ownership problem. "Integration complexity" is usually a scope problem.
  • Before writing any code, get clear answers to 5 questions. If you can't answer them, you're not ready to build.
  • This takes 1-2 weeks. It saves 3-6 months.

The decision most teams get wrong

The instinct is to start building. "Let's spin up a prototype and see what we learn."

This works for consumer apps where you can ship, measure, and pivot in days. It fails for enterprise AI because:

  1. Stakeholder fatigue is real. A failed pilot poisons the well for 2 years. You don't get unlimited shots.
  2. Integration isn't optional. A demo that can't touch your CRM, ticketing system, or data warehouse isn't a product—it's a slide deck.
  3. "Learning" without constraints produces noise. You'll learn that GPT can summarize things. You won't learn whether your ops team will trust it.

What actually breaks in production

We've seen three failure modes kill projects that had working models:

1. The ownership vacuum

Nobody can answer: "If this produces a bad output, who fixes it?"

The AI team built it, but they don't own the workflow. The business team owns the workflow, but they don't understand the system. When something breaks at 2am, everyone points at each other.

Symptom: Escalations go nowhere. Bugs sit unfixed. Adoption stalls.

2. Moving goalposts

The project started as "summarize support tickets." Three months later it's "summarize, categorize, route, and draft responses."

Nobody formally agreed to the scope expansion. The timeline didn't change. The budget didn't change. The team burns out shipping something that was never properly scoped.

Symptom: The project is always "almost done." Demo dates keep slipping.

3. Integration surprise

The model works great in a notebook. Then you discover:

  • The CRM API has a 100-request-per-minute limit
  • The data warehouse refreshes nightly, not real-time
  • Legal never approved storing outputs in that system

Symptom: "Productionization" takes 3x longer than building the model.


The 5 questions (and why each one matters)

Run these before writing code. If you can't get clear answers, pause the project until you can.

Q1: Who owns the outcome?

Not "who sponsors it" or "who requested it." Who is accountable when it fails?

This person needs:

  • Authority to make decisions about the workflow
  • Budget to fix issues post-launch
  • Incentive tied to the outcome (not just the launch)

Red flag: The owner is "the AI team" or "IT." These are builders, not owners. Find the business owner or don't start.

Q2: What decision are we improving?

AI doesn't "do things." It improves specific decisions: approve/deny, route, prioritize, draft, summarize, recommend.

Name the decision. Be specific:

  • ❌ "Help customer support"
  • ✅ "Classify incoming tickets into 5 categories and route to the correct queue within 30 seconds"

If you can't name the decision, you're building a solution looking for a problem.

Q3: What data contract exists?

"We have lots of data" is not a contract. A contract answers:

  • Source: Where does the data come from? Who maintains it?
  • Freshness: How old can it be? Real-time? Daily? Weekly?
  • Ownership: Who can authorize access? Who defines "truth"?
  • Allowed uses: Can you train on it? Store outputs? Share with vendors?

Red flag: You're three meetings deep and nobody can tell you where the canonical data lives. This is a governance problem, not an engineering problem. Solve it before building.

Q4: What does "working" look like?

Not "accurate." Accuracy is a component, not a goal.

Define success in business terms:

  • Cycle time reduced by X%
  • Tickets resolved without escalation increased by Y%
  • Cost per interaction decreased by $Z

Then define the minimum bar:

  • What accuracy threshold is acceptable?
  • What latency is acceptable?
  • What error rate triggers a rollback?

Red flag: Success is defined as "the model is deployed." That's a milestone, not an outcome.

Q5: What happens when it fails?

Every AI system fails. The question is whether you've designed for it.

Define:

  • Fallback: When the system can't handle something, what happens? (Queue for human? Retry? Reject?)
  • Escalation: How do failures surface to someone who can fix them?
  • Rollback: If the system degrades, can you revert to the previous state?

Red flag: The answer is "we'll figure it out in production." You won't. You'll firefight.


When this framework doesn't apply

Be honest about what you're doing:

Project TypeUse This Framework?
Production workflow automationYes
Customer-facing featureYes
Internal tool replacing manual processYes
Technical spike / feasibility testNo — timebox and don't pretend it's a product
Hackathon / innovation dayNo — different rules apply
Pure research / R&DNo — but be clear that's what it is

The danger is calling something a "pilot" when it's really a spike. Pilots have outcomes. Spikes have learnings. Don't confuse them.


Diagram: Scoping decision flow

Scoping gates: YES to proceed, NO means work to do first


Checklist: Run this before writing code

Copy this. Run it in a 90-minute session with your business owner, technical lead, and data owner.

Ownership

  • Business owner identified (name, not role)
  • Owner has authority over the workflow
  • Owner has budget for post-launch iteration
  • Escalation path defined

Decision

  • Specific decision named (verb + object)
  • Current process documented
  • Human-in-the-loop shape defined (review, override, audit)

Data

  • Data sources identified
  • Data freshness requirements documented
  • Data ownership confirmed (who can authorize access)
  • Allowed uses confirmed (training, storage, sharing)

Success

  • Business outcome defined (not "model deployed")
  • Metrics identified (cycle time, resolution rate, cost)
  • Minimum acceptable thresholds set
  • Evaluation method defined

Failure handling

  • Fallback behavior designed
  • Escalation path documented
  • Rollback capability confirmed
  • Monitoring requirements defined

Scoring:

  • 20/20: Ready to build
  • 15-19: Proceed with caution, close gaps in Week 1
  • <15: Pause and resolve before building

What to do next

Option A (DIY): Run this checklist with your team this week. If you score below 15, you've just saved yourself months of drift.

Option B (with us): We run a 2-week scoping engagement: stakeholder alignment, decision mapping, data contract review, success criteria, and a buildable spec. You get a document you can execute against—or hand to any team.

Book a 30-minute scoping review — we'll score your project against this checklist live.


Related reading

Ready to improve your website?

We help companies turn their website into a revenue-driving asset with high-converting design and clean Salesforce integration.

Get notified of new resources

Practical notes on AI agents, Salesforce, and web performance — no spam.