Skip to main content
Background Image

AI's Dual Edge: When to Disrupt and When to Compound

·3545 words·17 mins·
Pini Shvartsman
Author
Pini Shvartsman
Architecting the future of software, cloud, and DevOps. I turn tech chaos into breakthrough innovation, leading teams to extraordinary results in our AI-powered world. Follow for game-changing insights on modern architecture and leadership.
Table of Contents

The CEO just announced an “AI transformation” in the all-hands.

Your board wants to know your AI strategy. Product is pitching AI features for every roadmap. And you’re the one who has to turn vague executive enthusiasm into actual engineering work that creates value.

Here’s the decision you’re actually making: AI has two fundamentally different plays, and they require different resource allocation, different timelines, and different organizational commitment.

You can disrupt: fundamentally rewrite the economics of something, change what’s possible. Or you can augment: make existing systems measurably better without rebuilding them.

Disruption sounds impressive in board decks. Augmentation sounds boring. But picking wrong costs you six months of engineering time, burns team morale, and kills your credibility when you have nothing to show for it.

The question isn’t “should we do AI?” It’s “which play can we actually execute with the team, timeline, and organizational support we have right now?”

Most engineering leaders default to disruption because it’s what executives want to hear. The reality is that augmentation is usually the better play: faster to value, lower risk, and it builds organizational muscle for bigger bets later.

Disruption: When You’re Changing the Game (And What It Actually Costs)
#

Disruption isn’t about being radical for board slides. It’s about fundamentally changing what’s economically viable: making something possible that wasn’t before, or making something so much cheaper that it changes behavior.

What real disruption looks like:

Tesla’s FSD learns from every mile driven by every car. They ship updates weekly because the fleet is the training ground. Hardware companies used to iterate in 3-year cycles. Their AI stack iterates in 3-week cycles. That’s not an improvement. It’s a different game.

Retail demand forecasting used to mean: forecast six months out, order inventory, pray you got it right, discount what you got wrong. Short-horizon AI forecasting turns that into a control system. Inventory, labor, and pricing adjust daily based on what’s actually happening. Companies doing this aren’t just reducing stockouts. They’re changing their cost of capital and margin structure.

Drug discovery used to mean brute-forcing millions of combinations. AI narrows the search space dramatically, eliminating 95% of dead ends before anyone wastes time and money on them.

Here’s what nobody tells engineering leaders about disruption:

It’s expensive, slow, and organizationally risky. You need:

12–18 month runway. Not “we’ll pilot it for a quarter.” Real disruption takes multiple iterations to get right. Your exec team needs to understand this is a long bet.

Dedicated team capacity. You can’t do this with 20% of someone’s time or as a side project. You need engineers who can focus without getting pulled into production fires every week.

Robust instrumentation from day one. You need to measure what’s actually happening in production, not what you hope is happening. Shadow mode, A/B testing infrastructure, automated rollback.

Executive air cover. When this takes longer than expected (it will), or when early results are mixed (they will be), someone senior needs to protect the team from getting cancelled.

Risk tolerance. Data will be brittle. Regulators might have opinions. Users might not trust it initially. These aren’t edge cases. They’re the entire problem space.

The question you need to answer honestly: Do we actually have these things, or are we pretending we do because the CEO is excited about AI?

If you don’t have this organizational support, you’re not doing disruption. You’re setting your team up for a science project that gets cancelled in Q3 when it hasn’t shipped yet.

Augmentation: Where Most Engineering Leaders Should Start
#

This is the play most teams should run first. Not because it’s less ambitious, but because it compounds faster, fails cheaper, and builds organizational credibility for bigger bets.

Augmentation means: take what you’re already doing, make it measurably better with AI, repeat. You’re not rebuilding the system. You’re making existing operations perform at a higher level.

What this looks like in practice:

Your warehouse operations team guesses where to put high-velocity items. You add AI slotting optimization. Labor costs drop 15%, on-time delivery improves 10%. Same warehouse, same people, better math. Engineering investment: 2-3 engineers for 8 weeks.

Your support team bounces tickets between departments until someone knows the answer. You add AI triage that routes correctly the first time. First-contact resolution goes up. Handle time goes down. Same team handles more volume with less frustration. Engineering investment: 1 team lead + 2 engineers for a quarter.

Your factory maintenance team discovers equipment failures when production stops. You add predictive maintenance that gives 48 hours warning. Unplanned downtime craters. You schedule repairs during planned maintenance windows. OEE improves without adding headcount. Engineering investment: 1 senior engineer + 1 ML engineer for 12 weeks.

Your fraud detection flags 1,000 transactions for manual review (970 are false positives). You improve risk scoring with AI. Manual review team focuses on actual problems. You catch more fraud with less work. Engineering investment: 2 engineers for 6 weeks.

None of this is revolutionary. All of it creates measurable value.

The business case is straightforward: improve a process by 10–15%, replicate across 20 facilities or 50 teams, create millions in value without changing your fundamental business model.

The leadership advantage: if it doesn’t work, you turn it off. Your rollback plan is “go back to how we did it last month.” You haven’t burned 18 months rewriting core systems. Your team learned something. Your organizational credibility is intact. You’re ready to try the next thing.

The Engineering Leader’s Playbook
#

Here’s what separates AI projects that succeed from ones that become expensive lessons. This isn’t theory. It’s what you need to set up and enforce to ship value.

Force Clarity on Metrics Before You Allocate Headcount
#

Before you assign engineers, demand this: What are we measuring, and what’s the current baseline?

Pick three use cases maximum. Each gets exactly two metrics:

Demand forecasting → Mean Absolute Percentage Error (MAPE) ↓, stockouts ↓
Fulfillment → cost per order ↓, on-time delivery ↑
Support → first-contact resolution ↑, handle time ↓

If your team can’t define success this precisely, don’t start. You’ll burn engineering capacity building something technically impressive that nobody can prove is working.

This is also how you protect your team from scope creep. When product comes back with “let’s add AI to five more things,” you point at the three use cases you committed to. Nail those first. Prove they work. Then, and only then, expand.

Teams that try to do ten AI initiatives simultaneously ship zero things that create value. Your job is to say no until the first three are in production and measured.

Set Default Architecture Standards (And Enforce Them)
#

Your team will want to overcomplicate this. They read papers, get excited about fine-tuning and agentic systems, and skip the boring foundations that actually ship.

Set this as the default path for 90% of use cases:

Start with RAG. Retrieval-Augmented Generation gets good results fast. The model pulls relevant context, then generates answers based on that context. Tell your team: make retrieval great and evals solid before touching anything fancier.

Fine-tune only when proven necessary. RAG solves most problems. Only let teams fine-tune when they’ve proven RAG can’t work and identified specific, consistent gaps. Fine-tuning is expensive, brittle, and requires maintaining training pipelines. Make them write a decision doc explaining why simpler approaches won’t work.

Agents require approval. Tool use and autonomous behavior are powerful, but they need rock-solid evals, guardrails, and failure handling. Don’t let teams build agents until they’ve proven they can ship and maintain production RAG systems.

Why this matters as a leader: Teams that skip straight to fine-tuning and agents because it sounds impressive waste six months debugging before admitting they should have started simpler. Meanwhile, teams that follow the standard path are in production after 8 weeks, collecting user feedback, and iterating based on real usage.

Your job is to protect your team from their own over-enthusiasm. Set the standard. Make exceptions require written justification.

Make Evals Non-Negotiable Infrastructure
#

Here’s what you need to enforce: No AI system goes to production without automated evaluation. Period.

Without evals, your team is flying blind. They don’t know if prompt changes improve things or break them. They don’t know if performance is degrading. They’re operating on vibes and anecdotes, and that’s how you end up with production incidents at 3am.

Mandate these measurements for every AI system:

Task success rate. Can it actually do the job? Your team defines what “success” means for each use case and measures it automatically. No handwaving.

Harmful/false output rate. How often does it hallucinate? How often does it generate something actively wrong or dangerous? This number needs to go in your operational dashboard.

Latency budget. Set it based on user expectations, not engineering wishful thinking. A perfect answer that takes 30 seconds is useless if users expect 2 seconds.

Drift detection. Is performance degrading over time as data or user behavior changes? Automated alerts when things slide.

Adversarial testing. Prompt injection, jailbreaks, data exfiltration attempts. These aren’t one-time tests. Make them part of CI/CD.

Enforce a deployment process that assumes failure:

Shadow mode → compare AI output to current system without user exposure
Canary → 5–10% of traffic
Staged rollout → gradual expansion with metric monitoring
Automated rollback → one command to revert

If your team can’t roll back in minutes, don’t let them ship. “Hope nothing breaks” isn’t an operational strategy.

Your role: Make evals part of definition-of-done. No PR merged, no deployment approved, until automated evaluation exists and passes.

Budget for Data Quality Like You Budget for Security
#

The engineering leaders winning with AI aren’t the ones with the fanciest models. They’re the ones who allocated engineering time to data infrastructure.

Your AI is only as good as your data. If your critical tables are stale or wrong, your AI will be confidently incorrect. Unlike traditional software where bad data causes visible errors, AI with bad data generates plausible-sounding nonsense. Users trust it, act on it, and then you discover the problem three weeks later when decisions were based on garbage.

What you need to mandate and fund:

Automated freshness and accuracy checks. If inventory data should update hourly and hasn’t updated in six hours, automated alerts fire before your AI starts making predictions based on stale state. This requires ongoing engineering time.

Feature stores and lineage. When AI goes wrong (it will), your team needs to trace it back. Where did this feature come from? How was it computed? When was it last updated? Without lineage, debugging takes days instead of hours. Budget for building this.

Privacy boundaries as architecture. PII redaction, consent management, access controls. These need to be architectural decisions from day one, not patches you add when legal asks questions or customers complain.

The mistake most leaders make: treating data quality as a one-time project. “We’ll clean it up in Q1, then focus on AI in Q2.”

That’s not how this works. Data quality is continuous infrastructure work like security or performance monitoring. If you don’t budget ongoing engineering time for it, your AI systems degrade slowly until they’re generating nonsense and nobody knows why.

Allocate 20-30% of your AI engineering capacity to data infrastructure. Yes, that feels like a lot. No, you can’t skip it and succeed.

Instrument Cost Tracking from Day One
#

Set up cost instrumentation before your team ships anything. You need to see problems before they show up on the bill.

Track unit cost per task, not cost per token. Tokens are implementation details. What matters to your P&L: how much does it cost to process one customer inquiry? Generate one forecast? Triage one ticket? Make your team instrument this.

Set budget caps per service with automated alerts. If your support bot suddenly makes 10x more API calls because someone changed a prompt, you want alerts firing immediately, not a surprise $50K bill at month-end.

Default to “good enough” models with justification required for upgrades. Most tasks don’t need GPT-5. They need consistent, fast, correct answers at reasonable cost. Smaller models deliver that for 10% of the cost. Make your team write a doc explaining why they need expensive models before approving it.

Why this matters: AI costs scale with usage in ways traditional infrastructure doesn’t. A prompt change can 10x your API costs overnight. Without instrumentation, you discover this when finance asks why cloud spend jumped 300% last month.

Set Security Policies Early
#

Organizations that treat AI security like web security from 2005 learn through expensive incidents. Don’t be one of them.

Mandate isolation for untrusted tools. If your AI can call APIs or access systems, require sandboxing and signed function calls. Don’t let teams assume models will only do what they want. Make them plan for unexpected behavior.

Require output filtering for sensitive data. If AI works with PII, PHI, or confidential information, mandate automated checks that verify sensitive data doesn’t leak through responses. Trust but verify.

Include models in post-incident reviews. When things break, your team needs to trace through code, data, and model behavior. “The AI did something weird” isn’t a root cause. Make them explain why it behaved that way.

Assume hostile users from day one. Users will try to jailbreak your system. They’ll attempt prompt injection. They’ll try to extract training data. Make adversarial testing part of your standard release process, not something you add after an incident.

What to Demand from Your Executive Leadership
#

If you’re an engineering leader trying to get organizational support for doing AI right, here’s what you need from the C-suite. Don’t assume they understand this—educate them.

They need to ask for metrics, not demos. Train your CEO to say “show me the before/after chart” instead of “show me the demo.” Demos prove nothing. Metrics prove value.

They need to enforce constraints. When the CEO says “add AI to everything,” your job is to push back: “We’re committing to three use cases. We’ll nail those, prove they work, then expand.” Get executive support for saying no to scope creep.

They need to protect measurement windows. AI projects need time to collect data and iterate. When the board wants to see progress every week, your CEO needs to explain that AI isn’t like shipping features. It requires measurement cycles. Get them to buy you that time.

They need to understand build vs. buy. Most AI infrastructure is undifferentiated. Default to buying foundation models and tooling. Build only where you control the workflow and the data improves by being used. Make sure your CFO understands why you’re spending $50K/month on API calls instead of building custom models.

They need to tie incentives to adoption and impact, not shipped features. Shipping AI features is easy. Making them create measurable value is hard. Make sure compensation and promotions reward outcomes, not output.

If you can’t get this from executive leadership: Your job is harder but not impossible. Set these expectations yourself through data. Track baselines religiously. Publish metrics that show real impact. Kill things that don’t work publicly. Build your credibility through measured results, then use that credibility to demand better organizational support.

A 90-Day Plan for Engineering Leaders
#

Here’s a realistic timeline that assumes you have normal organizational constraints: technical debt, competing priorities, and a team that’s already fully loaded. Adjust based on your capacity.

Week 0–2: Define Success and Get Organizational Alignment
#

Pick 3 use cases maximum. Document the success metrics and measure current baseline. Get exec buy-in on these metrics (they become your definition of success). If you can’t measure it today, you can’t prove AI improved it later.

Assign one engineer to stand up a basic evaluation harness. Start simple: a script that runs AI on test cases and validates outputs.

Have your data engineering team add quality checks to tables that feed these use cases. You need automated alerts when input data goes stale or wrong.

Organizational work: Get your CEO/CFO to agree that these three use cases are the commitment for the quarter. Push back on new requests until you deliver these.

Week 3–6: Ship v1 in Shadow Mode
#

Allocate 2-3 engineers to build v1. Put it behind a feature flag. Run in shadow mode (processes real traffic, users don’t see output). Compare AI decisions to what your current system does.

Have one engineer instrument cost tracking per task. Set budget caps with automated alerts.

Run red-team exercises. Assign someone to try breaking it. Fix the top five issues.

Organizational work: Weekly metrics review with exec team. Show shadow mode results. Manage expectations: this is data collection, not feature launches.

Week 7–10: Canary to Real Users (Finally)
#

Route 5–10% of traffic to the AI system. Monitor metrics obsessively. Is it actually better than baseline from week 1?

Run table-top incident exercises with your ops team. Practice rollback procedures. Make sure everyone knows how to revert quickly if needed.

Make a hard decision: Look at your three use cases. Kill the weakest one. Reallocate that team capacity to double down on the strongest performer.

Organizational work: Present early results to exec team. Explain why you killed one project. Frame it as disciplined resource allocation, not failure.

Week 11–13: Scale What Works, Stop What Doesn’t
#

Increase traffic to 25–50%. Publish before/after charts showing real business impact: cost reduction, quality improvement, time savings. Whatever metrics you committed to in week 0.

If you have appetite for risk and spare capacity, move one agentic capability (tool use, function calling) into a low-risk workflow with human approval required for every action.

Refresh your backlog. Add one new use case only if you’ve proven the others work and have team capacity. Don’t accumulate half-finished AI projects that drain morale.

Organizational work: Deliver a quarterly retrospective to leadership. What worked, what didn’t, what you learned. Set expectations for next quarter based on demonstrated capacity, not aspirations.

Anti-Patterns Engineering Leaders Need to Kill
#

If you see these in your organization, stop the work and fix them. These are the warning signs of projects that will fail expensively.

“Our AI is 90% accurate!” Ask: 90% of what? Measured how? Against what baseline? Compared to what existing system? If your team can’t answer precisely, they’re not measuring. They’re guessing. Don’t let them continue without proper evaluation.

Prompts managed in Notion, Slack, or tribal knowledge. If prompts aren’t in version control with regression tests, they will drift. Someone will make a “small change” that breaks production, and your team won’t know what changed or how to roll back. Mandate version control for prompts like you mandate it for code.

“We’ll clean the data after we ship the feature.” This never happens. Your team will ship with dirty data, get weird results, spend weeks debugging, and trace it back to data quality issues they identified in week 1 but deprioritized. Make data quality a prerequisite, not a nice-to-have.

Building agents before mastering basic RAG. If your team can’t reliably retrieve the right document and generate a good answer with basic RAG, don’t let them add autonomy and tool use. It doesn’t make failures better. It makes them more expensive and harder to debug.

Quarterly demos with unchanged metrics. If your teams demo AI features every quarter but unit costs, cycle times, and error rates haven’t moved, they’re building demos, not products. Metrics are reality. Demos are theater. Shut down projects that can’t show measurable business impact.

What Success Looks Like for Engineering Leaders
#

The gap between “we’re doing AI” and “we’re getting measurable value from AI” isn’t technology or budget. It’s leadership discipline.

The organizations winning aren’t the ones with the biggest AI teams or the fanciest models. They’re the ones whose engineering leaders:

Force outcome clarity before allocating resources. They know exactly what they’re optimizing for before assigning engineers. No vague mandates, no “we’ll figure it out as we go.”

Build boring infrastructure first. Data quality checks, evaluation harnesses, cost tracking, rollback mechanisms. The unglamorous work that doesn’t make good board slides but determines whether you succeed in production.

Measure and publish honestly. Before/after charts with real baselines. When something doesn’t work, they say so publicly. When something works, they have numbers to prove it.

Kill things decisively. They’re as comfortable shutting down failed experiments as launching new ones. They frame it as disciplined resource allocation, not failure.

Protect their teams from organizational chaos. They push back on scope creep. They demand measurement windows. They buffer their engineers from executive enthusiasm that would otherwise destroy focus.

This isn’t science fiction or research. It’s practical systems thinking applied to a new capability.

Warehouses that stop guessing where to put inventory. Support teams that route correctly the first time. Maintenance teams that fix things before they break. All of it measurable. All of it replicable. All of it built by engineering leaders who understood the difference between disruption and augmentation, picked the right play for their organization, and executed with discipline.

Your CEO wants AI transformation. Your board wants competitive advantage. Your job is to deliver measurable business impact while protecting your team’s capacity for the work that actually matters.

Pick your play. Set your constraints. Allocate deliberately. Measure obsessively. Kill ruthlessly. Scale what works.

That’s how you turn executive enthusiasm for AI into lasting organizational value.

Related

Ship Faster Without Breaking Things: DORA 2025 in Real Life
·1643 words·8 mins
The Context Problem: Why AI Can't Remember You Across Apps (And Why That's Not an Accident)
·1933 words·10 mins
Google Chrome's AI Leap: Timely, Strategic, and Dominant
·190 words·1 min