Olito Labs Geological strata diverging, warm gold and cream tones

The Widening AI Gap

I scored 50 consultants across top strategy firms (McKinsey, BCG, Bain), Big 4, venture capital and private equity funds, and boutique practices. All use AI daily. A small group builds systems they say multiply output several times over. Everyone else has plateaued at basic chat. The gap appears to be widening.

Tanush Jagdish, Founder & CEO, Olito Labs February 2026 8 min read Primary Research
At a Glance

I scored 50 consultants across top strategy firms, Big 4, venture capital / private equity, and boutique practices on their AI capability across three capability tiers: everyday chat, advanced prompting (giving AI detailed instructions for complex tasks), and automation and AI-agent tools (software that runs multi-step tasks with minimal human input). All use AI daily, but nearly all usage concentrates in chat, with almost no activity beyond prompting. The gap is not between AI users and non-users. It is within the users themselves.

  • Capability clusters around everyday chat, with a sharp dropoff beyond it. Most participants are confident AI users, but the tools that actually multiply productivity are barely touched.
  • The few who push past chat grow rapidly because each tier appears to build on the last. Those who stop at chat stand still. The split widens at every level.
  • The gap is closable. Structured practice over weeks produced in one case, 5x reported deal throughput; in another, an estimated 80% time saving on a specific deliverable. But almost nobody is investing in closing it.
01

The AI Capability Score

Everyone started early. Here’s where they actually are, and what the distance between them means.

I interviewed 50 knowledge workers across McKinsey, BCG, Bain, Deloitte, Kearney, boutique practices, and venture capital / private equity firms. These are not laggards. They use AI consistently. They are driven, high-performing members of the workforce.

I assessed their capabilities across three capability tiers, ranging from everyday chat to AI agents that run tasks independently. The result is not what I expected.

Exhibit 1 AI Capability Score distribution (n=50)
AI Capability Score 0 10 20 30 40 50 60 70 80 90 CHAT BASICS CHAT + PROMPTING MULTI-TOOL AUTOMATION MEDIAN: 35 GAP: 24 PTS 7 of 50 above 50 →
Source: Olito Labs Primary Research, n=50. In-depth interviews with consultants and knowledge workers across top strategy firms, Big 4, boutique and independent practices, and venture capital / private equity. Each participant scored 0–100 across three capability tiers: everyday chat, advanced prompting, and automation and AI agents. Full methodology →

Nearly everyone lands in the lower half. Not because they are behind, but because they do one thing: everyday chat. They ask questions, draft emails, summarize documents. Then it stops. The right side of the chart is almost empty. Many of these are people their organizations would consider AI-forward.

The historian Paul David documented the same pattern with electricity: factories had electric motors for decades before productivity moved, because gains required redesigning how work was done around the new technology. The same thing is happening now. The tools are here; the reorganization of work around them has barely begun.

Key Findings
  • Most of the cohort scores in the lower half, despite daily AI use and strong self-assessments.
  • Only a handful break above 50; the highest score is 62.
  • Scores bunch in a narrow band, but the distance between 35 (median) and 62 (top) produces outsized differences in output.
  • The gap lives inside the user base, not at its edges.
02

The Separation

Break the composite score into three tiers, and a pattern emerges at every level.

The composite score tells us most people land in the lower half, but not why. Break it into the three capability tiers and a hidden pattern emerges: two distinct clusters, one of people still learning and one of people who have pushed through.

Exhibit 2 The skill cliff: three tiers, one cohort
Everyday Chat Comfort zone GROUP A GROUP B Advanced Prompting Two distinct groups GROUP A GROUP B Automation & Agentic The untouched frontier 66% below 20 0 10 20 30 40 50 60 70 80 90 Score (0–100) · Each dot = 1 participant · n=50
Source: Olito Labs Primary Research, n=50. Automation & AI Agents tier uses the higher of the two per participant. Scoring methodology →
What each tier means
Everyday Chat: ChatGPT, Claude.ai, Gemini. Asking questions, getting summaries, basic drafting. Advanced Prompting: Giving AI a defined role and detailed instructions, requesting output in specific formats, and maintaining context across multiple conversations. Automation & AI Agents: Workflow automation platforms (Zapier, Make), connecting software systems programmatically, AI-powered coding tools (Claude Code, Cursor), and autonomous agents that run multi-step tasks with minimal intervention. This tier will subdivide as the field matures. For this cohort, the frontier is unified: almost no one has crossed it. See full tier definitions →

Start with Everyday Chat, the tier where confidence lives. Nearly everyone scores high here. For most, “using AI” begins and ends at this level. But even within chat the distribution splits: Group A uses it intermittently or reluctantly. Group B has made it a daily reflex. The gap between them is already visible, and it is the first sign of the pattern that repeats at every level.

Advanced Prompting is where the cohort splits more sharply. One group has mastered detailed AI instructions, structured output formats, and keeping context across sessions. The other has not touched any of it. Two distinct clusters with a gap in between. This is where real separation starts, because prompting well takes deliberate practice: learning what works, iterating, building a routine. Those who invested that time are pulling away. Those who did not remain at chat, and the distance appears to compound over time.

Each tier depends on the one below it. You cannot automate what you have not learned to prompt well. You cannot run AI agents without the scaffolding that automation provides. The tiers are a ladder, not a menu. Each one multiplies the last, and stalling at any rung means you lose everything above it.

Beyond prompting, the frontier is empty. Automation, connecting tools together, command-line AI tools, autonomous agents. The layer where productivity should compound sits near zero for almost everyone. Two-thirds of the cohort score below 20 on Automation & Agentic. One outlier built an automated system connecting deal sourcing, CRM, and due diligence through software integrations, while everyone else remains at the chat level.

Yet this is where capability grew fastest in my data. One participant went from zero command-line experience to advanced usage in weeks of structured practice. The tools improve so quickly that the gap between someone who practices and someone who waits opens wider every month.

Key Findings
  • Everyday Chat clusters high, making it table stakes rather than a differentiating capability.
  • Advanced Prompting splits the cohort into two distinct groups with a valley between them.
  • Automation & Agentic collapses: 66% score below 20, with virtually no capability beyond prompting.
  • Fastest capability gains came from structured practice at this tier, measured in weeks rather than months.
  • Midpoint scores decline from 70 to 45 to 10. The separation is sequential: miss one tier, stall on all above it.
03

What the Gap Costs

The separation is not abstract. It shows up in deals closed, hours burned, and projects that implode.

What does this separation produce? Three patterns emerged from the interviews, each a version of the same question: did you restructure your work around the tools, or not?

Exhibit 3 Three archetypes: what the gap produces
Auto+Agent
Prompt
Chat
The System Builder Partner, Growth Equity Fund

“We can probably see five times as many deals.”

5x deal throughput
via automated pipeline

Built an automated system connecting deal sourcing, CRM, and due diligence through software integrations. First-day due diligence quality matches what used to take a full manual process. Yet the system is only half connected, and his team cannot replicate it.

Auto+Agent
Prompt
Chat
The Skill Builder Incoming Associate, MBB Firm

“I am 99% sure no one at the firm will be using it that way.”

80% time saved
on client deliverable

Started with zero command-line experience. Over weeks of structured practice, built reusable AI instructions, memory that persists between conversations, and workflows that span multiple sessions. A deliverable that typically takes an experienced associate a full week, he produced in a day. Senior colleagues with years of domain expertise could not match his throughput. Not because they lacked knowledge, but because they lacked the tools.

Auto+Agent
Prompt
Chat
The Skipper Consultant, Healthcare Strategy

“All the projects sold relying on AI in the proposal… they imploded.”

0 projects delivered
on AI-scoped timeline

She was comfortable with chat and never invested in advanced prompting. Like most, she was waiting for someone else to build the path. When her firm sold projects scoped to AI-accelerated timelines, the gap fell on her team. They had automation-tier deadlines with chat-tier skills. She absorbed the manual work and went on medical leave from burnout.

Source: Olito Labs Primary Research, n=50. These patterns recurred across firm types in the sample.

The archetypes follow the tier structure exactly. The System Builder operates across all three; his output multiplied because each tier builds on the last. The Skill Builder invested weeks of deliberate practice climbing the same ladder. The Skipper never built prompting-tier competence, so when the firm priced automation-tier productivity, the ladder broke beneath her.

The result is role compression. When one person handles 5x the deal flow and another saves 80% of the time on real deliverables, the distance from everyone else doesn’t grow by addition. It grows by multiplication. The rest are doing the same job without the same leverage.

Between these extremes sits the majority. As one participant put it: “I am the silent middle. If it is there, I will use it. If it is not there, you know.” Comfortable with chat, aware that more exists, waiting for someone else to build the path.

If you recognize yourself in the silent middle, our data suggests the gap is real and likely growing. Your peers who climbed appear to be pulling further away each month. The barrier is investment, not access or intelligence.

But the Skill Builder’s trajectory shows this is not a permanent position. He invested on his own. No firm in the sample had a structured program to move people up the tiers. The distance from chat to command-line tools closed in just a few weeks of deliberate practice, which is what makes this worth paying attention to: the gap is closable, but closing it requires investment that almost no one is making.

The gap lives inside the user base, widening because each tier appears to build on the last and almost nobody climbs. For every individual, the question is the same: invest in closing the gap, or watch it widen.

The pattern repeated across the sample. One consultant built a pipeline that reads 50 earnings transcripts, extracts specific financial metrics, and produces a comparative analysis. One participant reported that what took a team two days now runs in 15 minutes. Another set up an agent that monitors regulatory publications daily, surfacing only what matters to their client. A third automated the path from raw interview notes to structured deliverable drafts with citations, cutting first-draft time by 80%. These are not theoretical workflows. They are running today, built by people who were using only basic chat six months ago.

Key Findings
  • Climbing the tier ladder compounds output; skipping a rung breaks it.
  • Role compression is already here. A few people produce multiples more than everyone else in the same role.
  • The “silent middle,” comfortable with chat and waiting for structure, is the majority position and the most precarious.
  • The gap closed fastest through structured practice over weeks, not months. The barrier is investment, not access.

What this study does not capture

Most firms in the sample restrict sending proprietary data to external AI tools. That constraint is rational. Client data, deal terms, and internal analyses belong behind access controls, and no productivity gain justifies compromising that. The question is not whether to have governance, but how to build governance that enables adoption rather than blocking it. In practice, this means three things: using firm-approved tools where they exist, ensuring any external AI provider meets enterprise compliance standards (SOC 2 Type II, data processing agreements, zero-retention policies), and establishing clear protocols for what data can go where. The major AI providers, Anthropic and OpenAI included, now offer enterprise tiers with these controls. The gap between “we use AI” and “we use AI responsibly” is a governance design problem, and most firms have not begun to solve it. If your organization is working through AI governance and compliance patterns, I would welcome that conversation at olitolabs.com.

AI agents are confident when they are wrong. Every participant who moved beyond basic chat encountered this. The higher you climb the capability ladder, the more you depend on AI outputs, and the harder it becomes to verify them. Several participants described spending almost as long checking AI-generated work as they would have spent doing it manually. The validation burden does not disappear with better tools. It shifts shape.

This sample is 50 professionals across consulting, finance, and adjacent fields. They skew motivated: people who said yes to a research interview about AI are not a random draw from the workforce. The scores, distributions, and patterns reported here describe this cohort, not the broader population. The framework is designed to be replicable (the full pipeline is open-source), but the findings should be read as directional, not representative.

Tool builders ship improvements weekly. Features that required command-line access a year ago now have graphical interfaces. The tier boundaries in this taxonomy will shift as tools mature. What this study captures is a snapshot: where the gap stands today, among this cohort, with today’s tools. The structural observation, that each tier compounds on the last, is likely durable. The specific distances may not be.

Close the gap
The Skill Builder went from zero command-line experience to 80% time savings in weeks. No firm in this sample had a structured program to get people there. I built one: a coaching program designed around this tier progression, with structured practice, real deliverables, and accountability. If you are interested in structured practice along these tiers, I would welcome that conversation at olitolabs.com.

Sources & References

1
Olito Labs Primary Research
50 in-depth interviews · Feb 2026 · Primary data: 50 scored participants, 212 pain points
Primary
2
McKinsey Global Survey: The State of AI in 2025
72% organizational adoption
Context
3
Harvard / BCG: Navigating the Jagged Technological Frontier
12% more tasks, 25% faster with AI
Context
4
Paul A. David, “The Dynamo and the Computer: An Historical Perspective on the Modern Productivity Paradox,” American Economic Review, 1990
Workflow restructuring lag · electricity/productivity parallel
Context