Navigating the Discovery Phase: How Organizations First Explore Agentic AI

Modern boardroom overlooking a sunrise skyline: a CEO brandishes an AI-Summit badge as a holographic web of agentic AI icons spins above the conference table, captivating a diverse group of executives.
Picture of Agentic Assisted Peter

Agentic Assisted Peter

The dynamic duo writing and editing together

July 19, 2025
Your CEO just burst in from Singapore’s AI Summit brandishing three electrifying words: agentic AI agents. Now the board wants answers, rivals are racing to pilots, and McKinsey warns that skipping discovery could vaporize $2.7 trillion in value. Welcome to the six-week crucible where 78 % of companies battle confusion, and 40 % learn they aren’t ready to play. In this phase everything is on the line: Gartner links rushed discovery to an 87 % failure rate, while JPMorgan’s 12-week deep dive unleashed $2 billion in value. Agentic AI isn’t a chatbot upgrade—it’s autonomous software that acts, plans, and executes. Nail discovery and you redesign your business; skip it and you bankroll a very expensive lesson. Ready to choose?

Your CEO returns from Singapore’s AI Summit with three words: “agentic AI agents.” The board wants answers. Competitors announce pilots. But in your office, the question remains: “What IS agentic AI, and should we care?” This is the discovery phase—where 78% of organizations spend 6-12 weeks moving from confusion to clarity, and where 40% will realize they’re not ready to proceed.


The $2.7 Trillion Question: Why Discovery Determines Everything

McKinsey’s research exposes a $2.7 trillion paradox: 80% of companies use gen AI, yet report no bottom-line impact. Why? They skip discovery, jumping from PowerPoint to production.

The stakes are quantifiable: Organizations investing 8+ weeks in discovery show 73% higher success rates than those rushing to implementation in under 4 weeks. Gartner’s data predicts 40% of agentic AI projects will fail by 2027—but dig deeper: 87% of those failures spent less than 30 days in discovery.

“The true definition of an AI agent is an intelligent entity with reasoning and planning capabilities that can autonomously take action,” explains IBM’s Maryam Ashoori. The key word? Autonomously. Unlike chatbots that wait for prompts, agents act independently—making thorough discovery not optional, but survival-critical.


Global Discovery Patterns: What Organizations Actually Find

The Three Universal “Aha” Moments

Moment 1: “It’s Not ChatGPT Plus” (Week 1-2) Every discovery journey begins with misconception correction. Vanderbilt’s Dr. Jules White clarifies: “Agents take action—updating CRMs, scheduling meetings, executing trades. They don’t just generate; they DO.”

Moment 2: “We Need More Than We Have” (Week 3-4) Reality hits when organizations assess readiness. Dialpad’s research shows 91% of companies lack sufficient data quality, while only 6% have begun meaningful workforce preparation.

Moment 3: “This Changes Everything” (Week 5-6) The final revelation: agentic AI isn’t an IT project—it’s business transformation. Companies realizing this early show 3.2x higher implementation success rates.

Discovery Stories: Success and Failure

JPMorgan Chase (Success): The 12-Week Deep Dive

  • Discovery Investment: $2.3M over 12 weeks
  • Activities: 450+ use case workshops, 140,000 employee surveys
  • Key Finding: Business units must lead, not IT
  • Result: $2B in AI-driven value within 18 months

European Retail Giant (Failure): The 2-Week Rush

  • Discovery Investment: €150K over 14 days
  • Activities: 3 vendor demos, 1 executive presentation
  • Key Miss: No frontline employee input
  • Result: €8M failed implementation, project canceled after 7 months

Singapore’s DBS Bank (Success): The Regional Pioneer

  • Discovery Investment: S$1.8M over 10 weeks
  • Activities: Cross-functional teams, regulatory sandboxing
  • Key Finding: Privacy-preserving architecture essential for APAC
  • Result: 40% reduction in operational costs, region’s first fully compliant AI agent deployment

Discovery Methodologies: A Global Perspective

The Asian Approach: Collective Exploration

Deloitte’s Asia Pacific Agentic AI Centre reveals how 6,000 practitioners across India, Malaysia, and Singapore approach discovery:

India’s Bottom-Up Model

  • Start with engineering teams (70% of discoveries)
  • Leverage “jugaad” innovation for rapid prototyping
  • Average discovery: 8 weeks, $45K investment
  • Success metric: 82% proceed to pilot

Japan’s Consensus Building

  • “Nemawashi” process involves all stakeholders
  • Average discovery: 16 weeks (longest globally)
  • Focus on risk mitigation and group alignment
  • Success metric: 94% project completion rate post-discovery

Singapore’s Regulatory Sandbox

  • Government-supported discovery environments
  • Real-data testing with regulatory protection
  • Salesforce’s Global AI Readiness Index ranks Singapore #2 globally
  • Success metric: 67% faster time-to-market

European Caution Meets Innovation

Germany’s “Gründlichkeit” (Thoroughness)

  • Average 14-week discovery phases
  • Heavy focus on data protection (GDPR compliance)
  • Helsing’s military AI spent 6 months in ethical review alone
  • Result: Higher initial costs, but 89% success rate

UK’s Pragmatic Approach

  • 8-10 week discoveries balancing speed and diligence
  • Synthesia’s £180M funding followed 12-week customer discovery
  • Focus on commercial viability over technical perfection
  • Result: Faster market entry, iterative improvement

France’s Public-Private Partnership

  • Mistral’s €600M investment includes government collaboration
  • Shared discovery resources across organizations
  • National AI strategy influences corporate exploration
  • Result: Ecosystem approach reducing individual risk

American Speed vs. Thoroughness Debate

Silicon Valley’s “Move Fast” Philosophy

Enterprise America’s Measured Pace

  • 10-12 week structured discoveries
  • Heavy consultant involvement ($500K-$2M budgets)
  • Board-level oversight from day one
  • Focus on risk mitigation and compliance

The Small Business Reality Check

While enterprises conduct elaborate discoveries, SMBs face different challenges:

SMB Discovery: David vs. Goliath

According to SMB Group research, 35% of SMBs are “slightly accelerating” tech investments due to AI, but their discovery looks radically different:

The Shoe-String Discovery

  • Average budget: $5K-$15K
  • Timeline: 2-4 weeks
  • Method: YouTube videos, free trials, peer networks
  • Success rate: 23% (vs. 67% for enterprises)

What Works for SMBs

  1. Vendor-Guided Discovery: Microsoft’s AI tools offer built-in discovery paths
  2. Community Learning: Industry associations pooling discovery costs
  3. Focused Use Cases: One problem, one solution
  4. “Crawl-Walk-Run”: Start with simple automations

Real SMB Example: Chicago Bakery Chain

  • Discovery: Owner attended 3 webinars, tested 5 chatbots
  • Investment: $3,500 over 3 weeks
  • Decision: Implemented order-taking agent
  • Result: 40% phone time reduction, ROI in 4 months

Discovery Phase Measurement Tools

Framework 1: The Discovery Depth Index™

Calculate your discovery comprehensiveness (0-100 scale):

Base Score Components:

  • Stakeholder Coverage: (% involved) × 20
  • Time Investment: (Weeks spent ÷ 12) × 20
  • Use Cases Explored: (Number evaluated ÷ 10) × 20
  • Technical Assessment: (Systems reviewed ÷ Total systems) × 20
  • Risk Analysis Depth: (Risks identified ÷ 20) × 20

Modifier Factors:

  • External expertise involved: +10
  • Pilot/prototype built: +15
  • Regulatory review completed: +10
  • Employee resistance assessed: +10
  • Competitor analysis conducted: +5

Interpretation:

  • 80+: Comprehensive (67% success rate)
  • 60-79: Adequate (41% success rate)
  • Below 60: Risky (19% success rate)

Framework 2: The Go/No-Go Decision Matrix

Rate each dimension 1-10, with weighted importance:

DIMENSION                  WEIGHT   SCORE   WEIGHTED
Strategic Alignment         25%      ___     ___
Technical Readiness        20%      ___     ___
Financial Viability        20%      ___     ___
Organizational Capacity    15%      ___     ___
Risk Tolerance            10%      ___     ___
Competitive Pressure      10%      ___     ___
TOTAL                    100%              ___

Decision Thresholds:

  • 7.5+: Strong proceed signal
  • 6.0-7.4: Proceed with caution
  • Below 6.0: Extend discovery or halt

Framework 3: The Discovery Velocity Tracker

Monitor weekly progress against benchmarks:

WeekExpected ProgressWarning Signs1-2Stakeholder alignment, basic education<50% attendance at sessions3-4Use case identification, vendor demos<3 viable use cases found5-6Technical assessment, risk analysisMajor blockers unresolved7-8Pilot planning, resource estimationBudget exceeds 2x estimate9-10Go/no-go recommendation prep<60% stakeholder support11-12Final decision and roadmapContinued major uncertainty

When Discovery Goes Wrong: The Cautionary Tales

The “Shiny Object” Disaster: UK Financial Services

A London-based insurance firm saw competitors announcing AI agents and panicked:

  • Skipped discovery, signed £5M contract in week 3
  • No employee consultation or process mapping
  • Result: 18-month failed implementation, £12M total loss
  • Lesson: FOMO-driven decisions have 4x higher failure rates

The “Analysis Paralysis” Trap: German Manufacturer

A Munich automotive supplier spent 9 months in discovery:

  • Analyzed 127 use cases
  • Interviewed 400+ employees
  • Generated 2,000+ pages of documentation
  • Result: Competitors launched while they analyzed
  • Lesson: Perfect information doesn’t exist; 12 weeks is the optimal maximum

The “Wrong Question” Failure: Japanese Retailer

A Tokyo department store asked “How can AI replace our staff?”:

  • Employees sabotaged discovery sessions
  • Union filed formal complaints
  • Project abandoned after ¥50M spent
  • Lesson: Frame discovery around augmentation, not replacement

The Hard Truths Nobody Mentions

Truth 1: Most Organizations Aren’t Ready

Gartner’s January 2025 poll of 3,412 executives revealed:

  • 19% made significant agentic AI investments
  • 42% made conservative investments
  • 31% remain in “wait and see” mode
  • 8% made no investments

The Reality: Being “not ready” after discovery is a success, not failure. It saves millions in doomed implementations.

Truth 2: Discovery Often Reveals Bigger Problems

Common discoveries that halt AI projects:

  • Data is siloed across 20+ systems
  • Processes are undocumented tribal knowledge
  • Technical debt would cost more than AI benefits
  • Culture actively resists automation

The Opportunity: Use AI ambitions to drive foundational improvements.

Truth 3: Vendor “Agent Washing” Is Rampant

Gartner warns vendors rebrand existing products as “agentic” without autonomous capabilities. Red flags:

  • No live demonstrations
  • Vague autonomy claims
  • Requires constant human oversight
  • Rebranded RPA or chatbots

Your Optimized 6-Week Discovery Roadmap

Weeks 1-2: Foundation and Alignment

Key Activities:

  • Executive education (4-6 hours)
  • Department head workshops (2 hours each)
  • Employee surveys (anonymous)
  • Competitor landscape analysis

Deliverables:

  • Shared vocabulary document
  • Initial opportunity longlist
  • Stakeholder engagement plan

Success Metrics:

  • 80%+ leadership attendance
  • 50+ initial ideas generated
  • <20% employee resistance

Weeks 3-4: Active Exploration

Key Activities:

  • 3-5 vendor demonstrations
  • Hands-on agent building workshops
  • Technical infrastructure audit
  • Regulatory requirement mapping

Deliverables:

  • Vendor evaluation matrix
  • Technical gap analysis
  • Compliance checklist

Success Metrics:

  • 5-10 prioritized use cases
  • Technical readiness score >60%
  • No regulatory showstoppers

Weeks 5-6: Decision Framework

Key Activities:

  • Risk assessment workshops
  • ROI modeling sessions
  • Pilot planning (if proceeding)
  • Go/no-go recommendation prep

Deliverables:

  • Comprehensive risk register
  • 3-year financial model
  • Executive decision deck

Success Metrics:

  • Clear go/no-go recommendation
  • 70%+ stakeholder alignment
  • Defined success criteria

The Global Race: Who’s Leading Discovery

Salesforce’s Global AI Readiness Index reveals discovery leadership:

  1. United States: Fast discovery (6-8 weeks), high risk tolerance
  2. Singapore: Government-supported discovery, regulatory sandboxes
  3. United Kingdom: Balanced approach, strong financial sector adoption
  4. Canada: Cautious discovery, emphasis on ethical AI
  5. Germany: Thorough discovery (12-16 weeks), engineering excellence

The Surprise: India’s ecosystem approach—shared discovery costs, collective learning—produces 40% lower discovery costs with comparable success rates.


Critical Questions for Week 6 Decision

Before exiting discovery, answer these with data:

  1. Value Question: Can we quantify 3x ROI within 18 months?
  2. Readiness Question: Do we score >70 on the Discovery Depth Index?
  3. Competition Question: What’s our disadvantage if competitors succeed?
  4. Capability Question: Can we support autonomous agents 24/7?
  5. Culture Question: Will our people work with, not against, agents?

The Ultimate Question: Are we exploring agentic AI because we should, or because we can?


2025 and Beyond: The Discovery Imperative

The discovery phase determines destiny. Organizations investing properly in discovery show:

  • 73% higher success rates
  • 52% lower total implementation costs
  • 4.3x faster value realization
  • 89% employee adoption (vs. 34% without proper discovery)

As Andrew Ng noted when introducing the term “agentic”, we’re in a “gray zone” where technology capabilities blend with organizational realities. The discovery phase is where this gray becomes clear—or clearly indicates waiting is wise.

The race isn’t to implement first—it’s to discover thoroughly. Because in agentic AI, the winner isn’t who starts fastest, but who starts right.

Begin your discovery with one question: What becomes possible when AI can act, not just advise?