Why 95% of AI Pilots Fail (MIT Study) – And How to Beat the Odds

Last week, a MIT study sent shockwaves through the AI and business community: 95% of AI pilots fail to deliver measurable business returns. Headlines spread fast, with investors and executives questioning whether enterprise AI is a bubble.

But behind the headlines lies a more nuanced story. The study doesn’t show that AI lacks potential—it shows that most organizations are not yet equipped to turn AI experiments into real business impact.


Myth vs. Reality: What Other Research Tells Us

While the MIT report highlights execution gaps, other studies paint a more balanced picture:

  • McKinsey (2025): AI adoption is rising fast, with value emerging where firms rewire processes and governance.
  • Stanford AI Index (2025): Investment and adoption continue to accelerate, signaling confidence in the long-term upside.
  • Field studies: Copilots in customer service and software engineering deliver double-digit productivity gains—but only when properly integrated.
  • MIT SMR–BCG: Companies that give individuals tangible benefits from AI—and track the right KPIs—are 6x more likely to see financial impact.

The picture is clear: AI works, but only under the right conditions.


Why AI Projects Fail (The 10 Traps)

1. No learning loop
Many AI pilots are clever demos that never improve once deployed. Without feedback mechanisms and continuous learning, the system remains static—and users quickly revert to old ways of working.

2. Integration gaps
AI may deliver great results in a sandbox, but in production it often fails to connect with core systems like CRM or ERP. Issues with identity management, permissions, and latency kill adoption.

3. Vanity pilots
Executives often prioritize flashy use cases—like marketing campaigns or customer-facing chatbots—while ignoring back-office automations. The result: excitement without measurable cash impact.

4. Build-first reflex
Organizations rush to build their own AI tools, underestimating the complexity of User eXperience (UX), guardrails, data pipelines, and monitoring. Specialist partners often outperform in speed and quality.

5. Six-month ROI traps
Leadership expects visible returns within half a year. But AI adoption follows a J-curve: disruption comes first, with benefits only materializing once processes and people adapt.

6. Weak KPIs
Too many pilots measure activity—such as number of prompts or usage time—rather than outcomes like error reduction, cycle time improvements, or cost savings. Without the right metrics, it’s impossible to prove value.

7. No product owner
AI projects often sit “between” IT, data, and the business, leaving no single accountable leader. Without an empowered product owner with a P&L target, projects stall in pilot mode.

8. Change ignored
Technology is deployed, but users aren’t engaged. Poor UX, lack of training, and trust concerns mean adoption lags. In response, employees turn to consumer AI tools instead of sanctioned ones.

9. Data & policy drag
Even when the AI works, poor data quality, fragmented sources, and unclear governance delay rollouts. Legal and compliance teams often block scaling because policies are not defined early enough.

10. Wrong first bets
Too many companies start with complex tasks. Early success is more likely in “thin-slice” repetitive processes—like call summarization or contract intake—that can prove value quickly.


How to Beat the Odds (10 Fixes That Work)

1. Design for learning
Build AI systems with memory, feedback capture, and regular improvement cycles. If a tool cannot learn and adapt in production, it should never progress beyond pilot stage.

2. Fix integration before inference
Prioritize robust connections into your CRM, ERP, and ticketing systems. AI without seamless workflow integration is just an isolated chatbot with no business impact.

3. Pick quick-win use cases
Target repetitive, document- and conversation-heavy flows—like claims processing, contract extraction, or helpdesk queries. These areas deliver ROI within 90–120 days and build momentum.

4. Appoint an AI Product Owner
Every use case should have a leader with budget, KPIs, and authority. This person is responsible for hitting targets and driving the project through pilot, limited production, and full scale-up.

5. Measure outcomes, not activity
Define 3–5 hard business KPIs (e.g., −25% contract cycle time, −20% cost per contact) and track adoption leading indicators. Publish a regular value scorecard to make progress visible.

6. Buy speed, build advantage
Use specialist vendors for modular, non-differentiating tasks. Save your in-house resources for proprietary applications where AI can become a true competitive edge.

7. Rebalance your portfolio
Shift investments away from glossy front-office showcases. Focus on back-office operations and service processes where AI can cut costs and generate visible ROI quickly.

8. Make change a deliverable
Adoption doesn’t happen automatically. Co-design solutions with frontline users, train them actively, and make fallback paths obvious. Manage trust as carefully as the technology itself.

9. Educate the board on the J-curve
Set realistic expectations that ROI takes more than six months. Pilot fast, but give production deployments time to stabilize, improve, and demonstrate sustained results.

10. Prove, then scale
Choose two or three use cases, set clear ROI targets up front, and scale only after success is proven. This disciplined sequencing builds credibility and prevents overreach.


The Broader Reflection

The 95% failure rate is not a verdict on AI’s future—it’s a warning about execution risk. Today’s picture is simple: adoption and investment are accelerating, productivity impacts are real, but enterprise-scale returns require a more professional approach.

We’ve seen this pattern before. Just as with earlier waves of digital transformation, leaders tend to overestimate short-term results and underestimate mid- to long-term impact.

Agents vs. Automation – How to Choose the Right Tool for the Job

As AI agents storm the market and automation technologies mature, transformation leaders face a critical question: Not just what to automate — but how.

From RPA and low-code platforms to intelligent agents and native automation tools, the choices are expanding fast.

This article offers a practical framework to help you make the right decisions — and build automation that scales with your organization.


A Layered View of the Automation Landscape

Modern automation isn’t a single tool — it’s leveraging a full stack. Here are the key layers:

🔹 1. Digital Core Platforms

Systems like SAP, Salesforce, ServiceNow and Workday host your enterprise data and business processes. They often come with native automation tools (e.g., Salesforce Flow, SAP BTP), ideal for automating workflows within the platform.

🔹 2. Integration Platforms (iPaaS)

Tools like MuleSoft, Boomi, and Microsoft Power Platform play a foundational role in enterprise automation. These Integration Platforms as a Service (iPaaS) connect applications, data sources, and services across your IT landscape — allowing automation to function seamlessly across systems rather than in silos.

🔹 3. Automation Tools

  • RPA (e.g., UiPath) automates rule-based, repetitive tasks
  • Workflow Automation manages structured, multi-step business processes
  • Low-/No-Code Platforms (e.g., Power Apps, Mendix) empower teams to build lightweight apps and automations with minimal IT support

🔹 4. AI Agents

Tools and platforms like OpenAI Agents, Microsoft Copilot Studio, Google Vertex AI Agent Builder, and LangChain enable reasoning, adaptability, and orchestration — making them well-suited for knowledge work, decision support, and dynamic task execution.


Choosing the Right Tool for the Job

No single tool is right for every use case. Here’s how to decide:

ScenarioBest Fit
Rule-based, repetitive workRPA
Structured, approval-based flowsWorkflow Automation
Inside one platform (e.g., CRM/ERP)Native Platform Automation
Cross-system data & process flowsIntegration Platforms (iPaaS)
Lightweight cross-platform appsLow-/No-Code Platforms
Knowledge-driven or dynamic tasksAI Agents

The most effective automation strategies are hybrid — combining multiple tools for end-to-end value.


Implementation Roadmaps: One Journey, Many Paths

While all automation projects follow a shared journey — identify, pilot, scale — each tool requires a slightly different approach.


1. Identify the Right Opportunities

  • Native Platform Tools: Start with what’s already built into Salesforce, SAP, etc.
  • iPaaS: Identify silos where data must flow between systems
  • RPA: Use process/task mining to find repeatable, rule-based activities
  • Workflow: Focus on bottlenecks, exceptions, and handoffs
  • Low-/No-Code: Empower teams to surface automation needs and prototype fast
  • AI Agents: Look for unstructured, knowledge-heavy processes

2. Design for Fit and Governance

Each automation type requires a different design mindset — based on scope, user ownership, and risk profile.

  • Native Platform Automation: Stay aligned with vendor architecture and update cycles
  • iPaaS: Build secure, reusable data flows
  • RPA: Design for stability, handle exceptions
  • Workflow: Focus on roles, rules, and user experience
  • Low-/No-Code Platforms: Enable speed, but embed clear guardrails
  • AI Agents: Use iterative prompt design, test for reliability

Key distinction:

  • Native platform automation is ideal for secure, internal process flows.
  • Low-/no-code platforms are better for lightweight, cross-functional solutions — but they need structure to avoid sprawl.

3. Pilot, Learn, and Iterate

  • Platform-native pilots are quick to deploy and low-risk
  • RPA pilots deliver fast ROI but require careful exception handling
  • Workflow Automation start with one process and involve users early to validate flow and adoption.
  • Low-/no-code pilots accelerate innovation, especially at the edge
  • iPaaS pilots often work quietly in the background — but are critical for scale
  • AI agent pilots demand close supervision and feedback loops

4. Scale with Structure

To scale automation, focus not just on tools, but on governance:

  • Workflow and Low-Code: Set up federated ownership or Centres of Excellence
  • RPA and iPaaS: Track usage, manage lifecycles, prevent duplication
  • AI Agents: Monitor for performance, hallucination, and compliance
  • Native Platform Tools: Coordinate with internal admins and platform owners

The most successful organizations won’t just automate tasks — they’ll design intelligent ecosystems that scale innovation, decision-making, and value creation.


Conclusion: Architect the Ecosystem

Automation isn’t just about efficiency — it’s about scaling intelligence across the enterprise.

  • Use native platform tools when speed, security, and process alignment matter most
  • Use low-/no-code platforms to empower teams and accelerate delivery
  • Use RPA and workflows for high-volume or structured tasks
  • Use AI agents to enhance decision-making and orchestrate knowledge work
  • Use integration platforms to stitch it all together

The winners will be the ones who build coherent, adaptive automation ecosystems — with the right tools, applied the right way, at the right time.

GAINing Clarity – Demystifying and Implementing GenAI

Herewith my final summer reading book review as part of my newsletter series.
GAIN – Demystifying GenAI for Office and Home by Michael Wade and Amit Joshi offers clarity in a world filled with AI hype. Written by two respected IMD professors, this book is an accessible, structured, and balanced guide to Generative AI (GenAI), designed for a broad audience—executives, professionals, and curious individuals alike.

What makes GAIN especially valuable for leaders is its practical approach. It focuses on GenAI’s real-world relevance: what it is, what it can do, where it can go wrong, and how individuals and organizations can integrate it effectively into daily workflows and long-term strategies.

What’s especially nice is that Michael and Amit have invited several other thought and business leaders to contribute their perspectives and examples to the framework provided. (I especially liked the contribution of Didier Bonnet.)

The GAIN Framework

The book is structured into eight chapters, each forming a step in a logical journey—from understanding GenAI to preparing for its future impact. Below is a summary of each chapter’s key concepts.


Chapter 1 – EXPLAIN: What Makes GenAI Different

This chapter distinguishes GenAI from earlier AI and digital innovations. It highlights GenAI’s ability to generate original content, respond to natural-language prompts, and adapt across tasks with minimal input. Key concepts include zero-shot learning, democratized content creation, and rapid adoption. The authors stress that misunderstanding GenAI’s unique characteristics can undermine effective leadership and strategy.


Chapter 2 – OBTAIN: Unlocking GenAI Value

Wade and Joshi explore how GenAI delivers value at individual, organizational, and societal levels. It’s accessible and doesn’t require deep technical expertise to drive impact. The chapter emphasizes GenAI’s role in boosting productivity, enhancing creativity, and aiding decision-making—especially in domains like marketing, HR, and education—framing it as a powerful augmentation tool.


Chapter 3 – DERAIL: Navigating GenAI’s Risks

This chapter outlines key GenAI risks: hallucinations, privacy breaches, IP misuse, and embedded bias. The authors warn that GenAI systems are inherently probabilistic, and that outputs must be questioned and validated. They introduce the concept of “failure by design,” reminding readers that creativity and unpredictability often go hand in hand.


Chapter 4 – PREVAIL: Creating a Responsible AI Environment

Here, the focus turns to managing risks through responsible use. The authors advocate for transparency, human oversight, and well-structured usage policies. By embedding ethics and review mechanisms into workflows, organizations can scale GenAI while minimizing harm. Ultimately, it’s how GenAI is used—not just the tech itself—that defines its impact.


Chapter 5 – ATTAIN: Scaling with Anchored Agility

This chapter presents “anchored agility” as a strategy to scale GenAI responsibly. It encourages experimentation, but within a framework of clear KPIs and light-touch governance. The authors promote an adaptive, cross-functional approach where teams are empowered, and successful pilots evolve into embedded capabilities.

One of the most actionable frameworks in GAIN is the Digital and AI Transformation Journey, which outlines how organizations typically mature in their use of GenAI:

  • Silo – Individual experimentation, no shared visibility or coordination.
  • Chaos – Widespread, unregulated use. High potential but rising risk.
  • Bureaucracy – Management clamps down. Risk is reduced, but innovation stalls.
  • Anchored Agility – The desired state: innovation at scale, supported by light governance, shared learning, and role clarity.

This model is especially relevant for transformation leaders. It mirrors the organizational reality many face—not only with AI, but with broader digital initiatives. It gives leaders a language to assess their current state and a vision for where to evolve.


Chapter 6 – CONTAIN: Designing for Trust and Capability

Focusing on organizational readiness, this chapter explores structures like AI boards and CoEs. It also addresses workforce trust, re-skilling, and role evolution. Rather than replacing jobs, GenAI changes how work gets done—requiring new hybrid roles and cultural adaptation. Containment is about enabling growth, not restricting it.


Chapter 7 – MAINTAIN: Ensuring Adaptability Over Time

GenAI adoption is not static. This chapter emphasizes the need for feedback loops, continuous learning, and responsive processes. Maintenance involves both technical tasks—like tuning models—and organizational updates to governance and team roles. The authors frame GenAI maturity as an ongoing journey.


Chapter 8 – AWAIT: Preparing for the Future

The book closes with a pragmatic look ahead. It touches on near-term shifts like emerging GenAI roles, evolving regulations, and tool commoditization. Rather than speculate, the authors urge leaders to stay informed and ready to adapt, fostering a mindset of proactive anticipation.posture of informed anticipation: not reactive panic, but intentional readiness. As the GenAI field evolves, so must its players.


What GAIN Teaches Us About Digital Transformation

Beyond the specifics of GenAI, GAIN offers broader lessons that are directly applicable to digital transformation initiatives:

  • Start with shared understanding. Whether you’re launching a transformation program or exploring AI pilots, alignment starts with clarity.
  • Balance risk with opportunity. The GAIN framework models a mature transformation mindset—one that embraces experimentation while putting safeguards in place.
  • Transformation is everyone’s job. GenAI success is not limited to IT or data teams. From HR to marketing to the executive suite, value creation is cross-functional.
  • Governance must be adaptive. Rather than rigid control structures, “anchored agility” provides a model for iterative scaling—one that balances speed with oversight.
  • Keep learning. Like any transformation journey, GenAI is not linear. Feedback loops, upskilling, and cultural evolution are essential to sustaining momentum.

In short, GAIN helps us navigate the now, while preparing for what’s next. For leaders navigating digital and AI transformation, it’s a practical compass in a noisy, fast-moving world.

Fusion Strategy – How real-time Data and AI will Power the Industrial Future

This book by Vijay Govindarajan and Venkat Venkatraman gives excellent insights on how industrial companies can become leaders in this Data and AI-driven age.

Rather than discarding legacy strengths, the book shows how to fuse physical assets with digital intelligence to create new value, drive outcomes, and redefine business models. It gives a compelling and well-structured roadmap for industrial companies to get ready and lead through this digital transformation


From Pipeline to Fusion: A New Strategic Paradigm

Traditional industrial firms have long operated with a pipeline mindset – designing, building, and selling physical products through linear value chains. But in a world where customer needs change in real-time, and where data flows continuously from connected devices, this model is no longer sufficient.

Fusion Strategy introduces a new playbook: combine your physical strengths with digital capabilities to compete on adaptability, outcomes, and ecosystem value. It’s about integrating the trust and scale of industrial operations with the intelligence and speed of digital platforms.


Competing in the Four Fusion Battlegrounds

At the core of the book is a powerful matrix: four battlegrounds where industrial firms must compete – and four strategic levers to win in each: Architect, Organize, Accelerate, and Monetize.

Fusion Products – Embedding intelligence into physical products

This battleground focuses on evolving the traditional product into a smart, connected version that delivers value through both physical functionality and digital enhancements. It shifts the value proposition from one-time transactions to continuous value creation.

  • Architect: Build connected products with embedded sensors and software.
  • Organize: Create cross-functional product-data-software teams.
  • Accelerate: Use real-world usage data to improve iterations and performance.
  • Monetize: Shift to usage-based pricing, subscription models, or data-informed upgrades.

Example: John Deere integrates GPS, sensors, and machine learning into its agricultural equipment, enabling precision farming and monetizing through subscription-based services.

Fusion Services – Creating new layers of customer value

This battleground addresses the transformation from product-centric to outcome-centric offerings. Services become digitally enabled and proactively delivered, increasing customer stickiness and long-term revenue potential.

  • Architect: Design service layers that improve uptime, efficiency, or experience.
  • Organize: Stand up service delivery and customer success capabilities.
  • Accelerate: Leverage AI to scale and automate service interactions.
  • Monetize: Offer predictive maintenance, remote diagnostics, or outcomes-as-a-service.

Example: Caterpillar offers remote monitoring and predictive maintenance for its heavy equipment fleet, increasing operational uptime and generating recurring service revenues.

Fusion Systems – Transforming internal operations

This battleground focuses on using data and AI to reengineer internal processes, improve agility, and reduce cost-to-serve. Real-time operational intelligence becomes a source of competitive advantage.

  • Architect: Digitize plants, supply chains, and operations with real-time visibility.
  • Organize: Break down functional silos; design around data flows.
  • Accelerate: Use AI to optimize scheduling, energy use, or resource allocation.
  • Monetize: Drive efficiency gains and free up capital for reinvestment.

Example: Schneider Electric uses digital twins and data-driven energy management to optimize operations and reduce downtime in its global manufacturing network.

Fusion Solutions – Building platforms and ecosystems

This battleground is about building broader solutions that integrate products, services, and partners. It opens new avenues for value creation through platforms, data sharing, and co-innovation.

  • Architect: Offer modular solutions with open APIs and partner integration.
  • Organize: Orchestrate partner ecosystems that create mutual value.
  • Accelerate: Foster external innovation through developer communities.
  • Monetize: Sell analytics, data products, or platform access.

Example: Tesla is reimagining mobility not just as a product (cars) but as an integrated solution combining electric vehicles, software, energy management, autonomous driving, insurance and charging/energy infrastructure.


The Role of Data Graphs in Fusion Strategy

One of the foundational concepts emphasized throughout Fusion Strategy is the importance of data graphs. These are strategic tools that connect data across silos and enable intelligent, real-time insights.

A data graph is a semantic structure that maps relationships between entities—machines, sensors, people, processes, and locations—into a flexible and navigable format. In fusion strategy, data graphs link physical and digital domains, enabling smarter operations and decisions.

How to build a data graph:

  1. Collect data from operational systems – sensors, ERP -, CRM systems, etc.
  2. Define key entities and relationships – focus on what matters most.
  3. Create semantic linkages – use metadata and business context.
  4. Ensure real-time updates – to maintain situational awareness.
  5. Enable access – for both humans and AI systems.

Why data graphs matter:

  • Provide context for AI and analytics.
  • Enable real-time visibility across assets and systems.
  • Power predictive services, digital twins, and platform innovation.

According to the authors, data graphs are essential for scaling fusion strategies. Without them, it’s difficult to unify insights, drive automation, or deliver integrated digital experiences


Why This Book Stands Out

This is book does not start from the successful digital native companies, but from the leader of the industrial age point of view, describing on how they can become leaders in the digital age.

The structure is what makes it so useful:

  • It gives executives a language to discuss digital opportunities in operational and financial terms.
  • It balances the long-term vision with near-term execution levers.
  • It connects customer value, technology, organization, and monetization in one integrated model.

It’s a strategy-led, boardroom-level guide to competing in the AI era.


My Reflections

  • Applying Fusion Strategy is a shift in how to re-architect your products and business. It requires rewiring how you create, deliver, and capture value.
  • You don’t need to become a tech company. You need to become a fusion company – one that blends operational excellence with digital innovation.
  • Winning in Fusion means rethinking strategy, governance, talent, and incentives – all at once in other words, enabling full transformation.

Fusion Strategy is essential reading for any industrial executive seeking to lead their company through this era of accelerated transformation. It’s not about jumping on the latest AI trend – it’s about designing a future-ready business, grounded in strategy.

The battlegrounds are clear. The tools are available. The time is now.

If AI Is So Smart, Why Are We Struggling to Use It?

The human-side barriers to AI adoption — and how to overcome them

In my previous newsletter, “Where AI is Already Making a Significant Impact on Business Process Execution – 15 Areas Explained,” we explored how AI is streamlining tasks from claims processing to customer segmentation. But despite these breakthroughs, one question keeps surfacing:

If AI is delivering so much value… why are so many organizations struggling to actually adopt it?

The answer isn’t technical — it’s human.

In this edition, I explore ten people-related reasons AI initiatives stall or underdeliver. Each barrier is followed by a practical example and suggestions for how to overcome it.


1. Fear of Job Loss and Role Redundancy

Employees fear AI will replace them, leading to resistance or disengagement. This is especially prevalent in operational roles and shared services.

Example: An EY survey found 75% of US workers worry about AI replacing their jobs. In several large organizations, process experts quietly slow-roll automation to protect their roles.

How to mitigate: Communicate early and often. Frame AI as augmentation, not replacement. Highlight opportunities for upskilling and create pathways for digitally enabled roles.


2. Loss of Meaning and Professional Identity

Even if employees accept AI won’t replace them, they may fear it will erode the craftsmanship and meaning of their work.

Example: In legal and editorial teams, professionals report reluctance to use generative AI tools because they feel it “cheapens” their contribution or downplays their expertise.

How to mitigate: Position AI as a creative partner, not a substitute. Focus on use cases that enhance quality and amplify human strengths.


3. Low AI Literacy and Confidence

Many knowledge workers don’t feel equipped to understand or apply AI tools. This leads to underutilization or misuse.

Example: I’ve seen this firsthand: employees hesitate to rely on AI tools and default to old ways of working out of discomfort or lack of clarity.

How to mitigate: Launch AI literacy programs tailored to roles. Give people space to experiment, and build a shared language for AI in the organization.


4. Skills Gap: Applying AI to Domain Work

Beyond literacy, many employees lack the applied skills needed to integrate AI into their actual workflows. They may know what AI can do — but not how to adapt it to their role.

Example: In a global supply chain function, team members were aware of AI’s capabilities but struggled to translate models into usable scenarios like demand sensing or inventory risk prediction.

How to mitigate: Invest in practical upskilling: scenario-based training, role-specific accelerators, and coaching. Empower cross-functional “AI translators” to bridge tech and business.


5. Trust and Explainability Concerns

Employees and managers hesitate to rely on AI if they don’t understand “how” it reached its output — especially in decision-making contexts.

Example: A global logistics firm paused the rollout of AI-based demand forecasting after regional leaders questioned unexplained fluctuations in output.

How to mitigate: Prioritize transparency for critical use cases. Use interpretable models where possible, and combine AI output with human judgment.


6. Middle Management Resistance

Mid-level managers may perceive AI as a threat to their control or relevance. They can become blockers, slowing momentum.

Example: In a consumer goods company, digital leaders struggled to scale AI pilots because local managers didn’t support or prioritize the initiatives.

How to mitigate: Involve middle managers in co-creation. Tie their success metrics to AI-enabled outcomes and make them champions of transformation.


7. Change Fatigue and Initiative Overload

Teams already dealing with hybrid work, restructurings, or system rollouts may see AI as just another corporate initiative on top of their daily work.

Example: A pharmaceutical company with multiple digital programs saw frontline disengagement with AI pilots due to burnout and lack of clear value.

How to mitigate: Embed AI within existing transformation goals. Focus on a few high-impact use cases, and consistently communicate their benefit to teams.


8. Lack of Inclusion in Design and Rollout

When AI tools are developed in technical silos, end users often feel the solutions don’t reflect their workflows or needs.

Example: A banking chatbot failed in deployment because call center staff hadn’t been involved in the design phase — leading to confusion and distrust.

How to mitigate: Involve users early and often. Use participatory design approaches and validate tools in real working environments.


9. Ethical Concerns and Mistrust

Some employees worry AI may reinforce bias, lack fairness, or be used inappropriately — especially in sensitive areas like HR, compliance, or performance assessment.

Example: An AI-based resume screener was withdrawn by a tech firm after internal concerns about gender and ethnicity bias, even before public rollout.

How to mitigate: Establish clear ethical guidelines for AI. Be transparent about data usage, and create safe channels for feedback and concerns.


10. Peer Friction: “They Let the AI Do Their Job”

Even when AI is used effectively, friction can arise when colleagues feel others are “outsourcing their thinking” or bypassing effort by relying on AI tools.

Example: In a shared services team, tension grew when some employees drafted client reports with AI in minutes — while others insisted on traditional methods, feeling their contributions were undervalued.

How to mitigate: Create shared norms around responsible AI use. Recognize outcomes, not effort alone, and encourage knowledge sharing across teams.


Final Thought: It’s Not the Tech — It’s the Trust

Successful AI adoption isn’t about algorithms or infrastructure — it’s about mindsets, motivation, and meaning.

If we want people to embrace AI, we must:

  • Empower them with knowledge, skills, and confidence
  • Engage them as co-creators in the journey
  • Ensure they see personal and professional value in change

Human-centered adoption isn’t the soft side of transformation — it’s the hard edge of success. Let’s create our transformation plans with that in mind.

Where AI is Already Making a Significant Impact on Business Process Execution – 15 Areas Explained

After exploring a wide range of expert sources—and drawing from my own experience—I collaborated with AI tools (ChatGPT, Gemini, Claude) to create a concise overview of where AI is currently having the biggest impact on business processes. The aim: to bring together the most referenced success areas across functions and reflect on why these domains are leading the way. Recognizing these patterns can help us anticipate where AI is likely to deliver its next wave of value (at the end of the article).

Below are 15 high-impact application areas where AI is already delivering significant value—each explained with clear benefits and real-world examples.


Marketing & Sales

1. Smarter Customer Service Automation
AI-powered chatbots and virtual agents are now central to handling customer inquiries. They can resolve a majority of tickets without human intervention, enabling 24/7 service while reducing costs and improving customer experience. Beyond just scripted replies, these agents learn from interactions to provide increasingly accurate and personalized support, allowing human teams to focus on complex or emotionally sensitive requests.
Example: Industry-wide AI adoption in contact centers, with 88% of firms reporting improved resolution times and reduced overhead (Statista, McKinsey).

2. Personalised Marketing at Scale
AI recommendation engines tailor content and product offerings based on individual browsing behavior, purchase history, and contextual data. This creates more relevant experiences for users and lifts conversion rates. Example, Amazon’s recommendation engine contributes over a third of its e-commerce revenue, proving the model’s commercial impact.

3. Sales Acceleration with AI
AI is transforming sales operations by taking over repetitive tasks like data entry, scheduling, and opportunity scoring. It also enables more informed decisions through predictive analytics, guiding sales teams to focus on leads with the highest conversion potential. Example: Salesforce research reveals that 83% of AI-enabled sales teams saw revenue growth versus 66% without AI. Besides this Salesforce example, I also can share from my personal working experience at Brenntag that AI solutions to guide “next best actions”  for salespeople drives significant impact.


Operations, Manufacturing & Supply Chain

4. Predictive Maintenance Efficiency
Traditional maintenance schedules often lead to unnecessary downtime or surprise equipment failures. AI flips the model by continuously analyzing sensor data to detect anomalies before breakdowns occur. This helps manufacturers schedule maintenance only when needed, extending equipment life and minimizing disruption.
Example: Mitsubishi and others use predictive maintenance tools that have led to up to 50% reduction in unplanned downtime.

5. AI-Powered Quality Control
In industries where product consistency is crucial, AI-enhanced computer vision inspects goods in real time for even the tiniest defects. These systems outperform the human eye in speed and accuracy, ensuring higher product quality and reducing waste from production errors.
Example: Automotive and electronics manufacturers now use AI to identify surface defects, alignment issues, or functional flaws instantly on the line.

6. Smarter Inventory Optimization
AI brings new precision to inventory planning by factoring in historical sales, seasonal trends, macroeconomic indicators, and real-time customer demand. This ensures businesses maintain optimal stock levels—avoiding both overstock and stockouts—while reducing working capital.
Example: Companies using AI in supply chain forecasting report inventory reductions of up to 35% (McKinsey).

7. Logistics Route Optimization
AI’s real-time route planning considers traffic, weather, delivery windows, and driver availability to suggest the most efficient routes. This leads to faster deliveries, fuel savings, and higher customer satisfaction. It also helps logistics providers scale without proportionally increasing operational complexity.
Example: DHL’s AI-driven routing platform reduces mileage per package and improves on-time delivery.


Finance, Accounting & Risk Management

8. Touchless Document Processing
Invoice entry and document reconciliation are among the most repetitive and error-prone tasks in finance. AI automates these workflows by reading, validating, and recording data with high accuracy, drastically reducing processing time and human error.
Example: Large enterprises report cutting invoice processing time by 80% and lowering cost per invoice by over 60%.

9. Smarter Fraud Detection Systems
Modern fraud schemes evolve too rapidly for traditional rules-based systems to catch. AI models can continuously learn from new data and detect suspicious behaviors in real time, flagging anomalies that might otherwise go unnoticed.
Example: A global bank using AI to process checks in real time saw a 50% drop in fraud and saved over $20M annually.

10. Automating Financial Controls
AI supports internal audit and compliance by automatically flagging unusual transactions, reconciling financial data, and generating traceable logs for auditors. This not only boosts confidence in regulatory reporting but reduces the burden on finance teams.
Example:  Deloitte finds AI-led controls improve accuracy, reduce audit costs, and streamline compliance workflows.


HR & Administration

11. Accelerated AI Recruitment
Hiring at scale is time-consuming and prone to bias. AI now supports end-to-end recruitment by screening CVs, analyzing video interviews, and predicting candidate-job fit based on past data. This enables faster, fairer hiring decisions and a better candidate experience.
Example: Unilever’s AI-powered hiring cut time-to-hire by 90%, reduced recruiter workload, and increased hiring diversity by 16%.

12. AI-Powered Admin Assistance
Whether it’s helping employees navigate HR policies or resetting passwords, AI bots respond instantly to internal requests. They resolve issues efficiently and learn from interactions to improve over time, reducing dependency on HR and IT service desks.
Example: AT&T’s HR bot answers thousands of employee questions per month, freeing up support teams and reducing internal wait times.


Software Development & IT Operations

13. AI Code Generation & Testing
AI-assisted development tools help engineers write code, suggest improvements, and run automated tests. This shortens development cycles, reduces bugs, and improves overall code quality. It also democratizes coding by assisting less-experienced developers with best practices.
Example: Enterprises report 20–30% faster feature delivery using AI-assisted development environments.

14. Intelligent IT Service Management
From incident triage to root cause analysis, AI is embedded in IT Service Management platforms to help resolve tech issues automatically. Predictive insights help prevent outages and minimize disruption across business-critical systems.
Example: Leading digitally enabled firms see average resolution time drop by 50%, with improved system reliability and user satisfaction.

15. AI-Driven DevOps Optimization
By analyzing telemetry data and past deployments, AI optimizes build pipelines, monitors production systems, and predicts future resource needs. This ensures smoother rollouts and better infrastructure planning.
Example: Cloud-native companies use AI to reduce deployment failures and improve performance-to-cost ratios in real time.


Why AI Wins in These Areas

Despite the diversity of domains, these success stories share clear commonalities:

  • High process volume: Tasks that are frequent and repetitive gain the most from automation.
  • Structured and semi-structured data: AI performs best where input data is clean or can be normalized.
  • Clear Return On Investment (ROI) levers: The efficiency gains are measurable—reduced cycle time, lowered cost, or increased accuracy.
  • Repeatable workflows: Standardized or rules-based processes allow for predictable automation.

In essence, AI is most effective where complexity meets scale. As more enterprises embed AI into their operations, it is not just making processes faster—it’s reshaping them for quality, agility, and scale in a digital-first world.

Looking ahead, the next wave of AI impact is likely to emerge in areas where unstructured data and human judgment still dominate today. Examples include:

  • Legal and contract management, where AI is starting to support contract drafting, review, and risk flagging.
  • Strategy and decision support, where generative AI can synthesize market trends, customer feedback, and financial data to help leaders shape better strategies.
  • Sustainability tracking, where AI can analyze supply chain and operational data to monitor and reduce environmental impact.

As models become more capable and context-aware, these higher-value and less-structured domains may soon follow the path of automation and augmentation already seen in the 15 areas above.

When Good Intentions Fail – Why Effective Governance Is the Fix

While many organizations focus on technology, data, and capabilities, it’s the governance structures that align strategy with execution, enable informed decision-making, and ensure accountability. Without effective governance, even the most promising digital or AI initiatives risk becoming fragmented, misaligned, or unsustainable.

This article explores how governance typically evolves during transformation, drawing on a framework presented in GAIN by Michael Wade and Amit Joshi (2025). It then outlines best practices and tools for establishing effective governance at every level of transformation—portfolio, program, and project.

The Governance Journey: From Silo to Anchored Agility
Wade and Joshi identify four phases in the evolution of transformation governance:

  • Silo: In this early phase, digital and AI initiatives are isolated within departments. There is little coordination across the organization, leading to duplicated efforts and fragmented progress.
  • Chaos: As a reaction to the issues with the siloed approach, often companies start putting governance in place—but often not very effectively. Leading to a proliferation of processes, tools and platforms.
  • Bureaucracy: In response to chaos, organizations implement formal governance structures. While this reduces risk and increases control, it can also stifle innovation through over-regulation and sluggish decision-making.
  • Anchored Agility: The desired end-state. Governance becomes a strategic enabler—embedded yet flexible. It ensures alignment and control without constraining innovation. Decision-making is delegated appropriately, while strategic oversight is maintained.

Most organisations go through this journey, understanding where your organization is helps to determine what kind of actions are needed and what to improve.

Effective Governance: Moving from Bureaucracy to Anchored Agility
Most successful digital and AI transformations mature into the Bureaucracy and Anchored Agility phases. These are the phases where effective governance must strike a balance between structure and adaptability.

Two proven approaches—PMI and Agile—offer best practices to draw from:

PMI Governance Best Practices

  • Well-defined roles and responsibilities across governance layers
  • Program and project charters to formalize scope, authority, and accountability
  • Clear stage gates, with decision points tied to strategic goals
  • Risk, issue, and change control mechanisms
  • Standard reporting templates to ensure transparency and comparability

PMI’s approach works best in large, complex transformations that require strong coordination, predictable delivery, and control of interdependencies.

Agile Governance Principles

  • Empowered teams with clear decision rights
  • Frequent review cadences (e.g., sprint reviews, retrospectives, and PI planning)
  • Lightweight governance bodies focused on alignment, not control
  • Transparent backlogs and prioritization frameworks
  • Adaptability built into the governance process itself

Agile governance is ideal for fast-evolving digital or AI initiatives where experimentation, speed, and responsiveness are critical.

Moving from Bureaucracy to Anchored Agility, is not moving away from PMI to only Agile Governance principles. Your portfolio probably will have mix of initiatives which leverages one or both of the approaches.

Governance Across Levels: Portfolio, Program, Project
A layered governance model helps ensure alignment from strategy to execution:

Portfolio Level

  • Purpose: Strategic alignment, investment decisions, and value realization
  • Key Bodies: Executive Steering Committees, Digital/AI Portfolio Boards
  • Focus Areas: Prioritization, funding, overall risk and performance tracking

Program Level

  • Purpose: Coordinating multiple related projects and initiatives
  • Key Bodies: Program Boards or Program Management Offices
  • Focus Areas: Interdependencies, resource allocation, milestone tracking, issue resolution

Project Level

  • Purpose: Delivering tangible outcomes on time and on budget
  • Key Bodies: Project SteerCos, Agile team ceremonies
  • Focus Areas: Daily execution, scope management, risk and issue tracking, delivery cadence

Connecting the Layers: How Governance Interacts and Cascades
Effective governance requires more than clearly defined levels—it demands a dynamic flow of information and accountability across these layers. Strategic priorities must be translated into executable actions, while insights from execution must feed back into strategic oversight.

  • Top-down alignment: Portfolio governance sets strategic objectives, funding allocations, and key performance indicators. These are cascaded to programs and projects through charters, planning sessions, and KPIs.
  • Bottom-up reporting: Project teams surface risks, status updates, and learnings which are aggregated at the program level and escalated to the portfolio when needed.
  • Horizontal coordination: Programs often interact and depend on each other. Governance forums at program level and joint planning sessions across programs help manage these interdependencies.
  • Decision and escalation pathways: Clear routes for issue resolution and decision-making prevent bottlenecks and ensure agility across layers.

Organizations that master this governance flow operate with greater transparency, speed, and alignment.

Tools and Enablers for Good Governance
Governance is not just about structure—it’s also about enabling practices and tools that make oversight effective and efficient:

  • Terms of Reference (ToR): Define the mandate, decision rights, and meeting cadence for each governance body.
  • Collaboration & Transparency Tools: Use of platforms like Asana, Confluence, Jira, MS Teams for sharing updates, tracking decisions, and managing workflows.
  • Standardized Reporting: Leverage consistent templates for status, risks, and KPIs to create transparency and drive focus.
  • RACI Matrices: Clarify roles and decision-making authority across stakeholders, especially in cross-functional setups.
  • Governance Calendars: Synchronize key reviews, steerco meetings, and strategic checkpoints across layers.

Lessons from the Field
From my experience, common governance pitfalls include over-engineering (which stifles agility), under-resourcing (especially at the program level), and slow/unclear decision making. Successful governance relies on:

  • Aligned executive sponsorship
  • Clear ownership at all levels
  • Integration of risk, value, and resource management
  • Enabling people to act

Conclusion
In digital and AI transformation, effective governance is not about control—it’s about enablement. It provides the structure and transparency needed to drive transformation, align stakeholders, and scale success. As your organization moves toward Anchored Agility, governance becomes less of a bottleneck and more of a backbone.

Where is your organization on the governance journey—and what would it take to reach the next phase?

Why Centres of Excellence Are the Backbone of Sustainable Transformation

Real-world lessons from building CoEs across domains

In every transformation I’ve led—whether in supply chain, commercial, innovation, or enabling functions—one thing has remained constant: transformation only sticks when it becomes part of the organizational DNA. That’s where Centres of Excellence (CoEs) come in.

Over the years, I’ve built and led CoEs across foundational disciplines, transformation approaches, and specific capabilities. When set up well, they become more than just support groups—they build skill, drive continuous improvement, and scale success.

This newsletter shares how I’ve approached CoEs in three distinct forms, and what I’ve learned about setting them up for lasting impact.


What is a CoE? (From Theory to Practice)

In theory, a CoE is a group of people with expertise in a specific area, brought together to drive consistency, capability, and performance. In practice, I’ve seen them evolve into vibrant communities of practitioners where people connect to:

  • Share business challenges and solutions
  • Scale learnings and continuously evolve best practices
  • Facilitate exchange between experts and users
  • Build a knowledge base and provide education

The most successful CoEs I’ve led were about enabling people to learn from each other, work smarter, and operate more consistently.


Three Types of CoEs I’ve built and led

1. Foundational CoEs – Building Core Capabilities

These are the bedrock. Without them, transformation initiatives often lack structure and miss out on leveraging proven approaches. Examples from my experience include:

  • Program & Project Management CoE
    Built on PMI (PMBoK) and Prince2 standards, this CoE offered training, templates, mentoring, and coaching. It became the go-to place for planning and executing complex programs and projects.
  • Process Management CoE
    Using industry frameworks (e.g., APQC), platforms (ARIS, Signavio), and process mining tools (Celonis, UiPath, Signavio), this CoE helped standardize processes and enabled teams to speak a shared process language and identify improvement opportunities through data.
  • Change Management CoE
    Drawing from Kotter’s principles and other industry best practices, we developed a change playbook and toolkit. This CoE played a critical role in stakeholder alignment and adoption across transformation efforts.
  • Performance Management CoE
    Perhaps less commonly named, but highly impactful. We developed strategy-linked KPI frameworks and supported teams in embedding performance reviews into regular business rhythms.
  • Emerging: AI Enablement CoE
    Looking ahead, I believe the next foundational capability for many organizations will be the smart and responsible use of AI. I’ve begun shaping my thinking around how a CoE can support this journey—governance, tooling, education, and internal use case sharing.

2. Transformation-Focused CoEs – Orchestrating Change Across the Enterprise

Unlike foundational CoEs, these focus on embedding transformation methodologies and driving continuous improvement across functions. In my experience, they’re essential for changing both mindsets and behaviors.

  • Continuous Improvement | Lean CoE
    Anchored in Toyota’s principles, our Lean CoE supported everything from strategic Hoshin Kanri deployment to local Kaizens. It equipped teams with the tools and mindset to solve problems systemically, and offered structured learning paths for Lean certification.
  • Agile CoE
    Created during our shift from traditional project models to Agile, this CoE helped scale Agile practices—first within IT, then into business areas like marketing and product development.
  • End-to-End Transformation CoE
    One of the most impactful setups I was part of. At Philips, in collaboration with McKinsey, we created a CoE to lead 6–9 month E2E value stream transformations. It brought together Lean, Agile, and advanced analytics in a structured, cross-functional method.

3. Capability & Process CoEs – Scaling New Ways of Working

These CoEs are typically created during the scaling phase of transformation to sustain newly introduced systems and processes.

  • Supply Chain CoEs
    I’ve helped build several, covering Integrated Planning, Procurement (e.g., SRM using Coupa/Ariba), and Manufacturing Execution Systems (e.g., SAP ME). These CoEs ensured continuity and ownership post-rollout.
  • Innovation CoE
    Focused on design thinking, ideation frameworks, and Product Lifecycle Management (e.g., Windchill). It enabled structured creativity, process adoption, and skill development.
  • Commercial CoEs
    Anchored new ways of working in e-Commerce, CRM (e.g., Salesforce), and commercial AI tools—helping frontline teams continuously evolve their practices.
  • Finance CoEs
    Supported ERP deployment and harmonized finance processes across regions and business units. These CoEs were key in driving standardization, transparency, and scalability.

Lessons Learned – How to Build an Effective CoE

Having built CoEs in global organizations, here’s what I’ve found to be essential:

  • Start with a Clear Purpose
    Don’t set up a CoE just because it sounds good. Be explicit about what the CoE is solving or enabling. Clarify scope—and just as importantly, what it doesn’t cover (e.g., handling IT tickets).
  • Design the Right Engagement Model
    Successful CoEs balance push (structured knowledge and solutions) with pull (responsiveness to business needs). Two-way communication is critical.
  • Build the Community
    Experts are crucial, but practitioners keep the CoE alive. Foster interaction, feedback, and peer-to-peer learning—not just top-down communication.
  • Leverage the Right Tools
    Teams, SharePoint, Slack, Yammer, newsletters, and webcasts all support collaboration. Establish clear principles for how these tools are used.
  • Measure What Matters
    Track adoption, usage, and impact—not just activity. Set CoE-specific KPIs and regularly celebrate visible value creation.

Closing Thought

CoEs aren’t a magic fix—but they are one of the most effective ways I’ve found to institutionalize change. They help scale capabilities, sustain momentum, and embed transformation into the organization’s ways of working.

If you’re designing or refreshing your CoE strategy, I hope these reflections spark new ideas. I’m always open to exchanging thoughts.

Scaling Digital Transformation: From MVP Success to Enterprise-Wide Impact

In many organizations, digital transformation starts strong but struggles to scale. Initial pilots show promise, yet momentum often stalls when leaders attempt to replicate success across the broader enterprise. Based on my own experience leading global transformations, I’ve found that the key to scaling lies in a phased, learning-driven approach—one that combines fast wins, co-creation, and structured rollout methodology with rigorous focus on adoption and impact.

I’ve applied this phased scaling approach in multiple global transformations—ranging from a Lean-based, end-to-end value chain transformation developed with McKinsey, to process standardization and operating model redesign in Innovation, Commercial, and Supply Chain domains. It also proved highly effective during large-scale ERP implementation programs, where structured rollout and local ownership were critical to success. These experiences shaped the approach outlined below.


1. Start with a Strategic MVP

Every transformation needs a spark—a Minimum Viable Product (MVP) that proves value fast. This is not about testing technology in isolation but selecting a use case that matters to the business and can deliver tangible outcomes quickly.

In practice:
During a global ERP rollout, we launched the MVP in smaller countries in South America. These markets had limited existing system support, so the gains from automation and standardized processes were immediate and highly visible. It was the ideal environment to test the new design while delivering clear early value to the business.

The goal: build credibility and show that “this works.” More importantly, use this phase to create internal advocates who will help carry the message forward.


2. Expand to a Representative Deployment

Once momentum is created, the next step is critical: move beyond the pilot phase and test your approach in a part of the organization that reflects the broader complexity and scale.

In practice:
In the same ERP program, we selected North America as the next deployment—arguably the toughest environment due to its scale, criticality, and history of customized automation. If we could succeed there, we knew we could scale globally. Through co-creation with local business and functional teams, we built the assets, governance model, and deployment methodology that became the foundation for all subsequent rollouts.


3. Codify Learnings and Create the Scaling Roadmap

After two deployments, you’ll have rich insight into what works, what doesn’t, and what needs to adapt. This is the time to capture and codify those learnings into a repeatable transformation playbook.

In practice:
In the end-to-end Lean transformations, we developed a comprehensive playbook that included standard tools, templates, and real-world examples. We also created resource models and scaling scenarios to support portfolio-level roadmap planning. This gave the transformation program clarity on how to scale with confidence.


4. Industrialize the Rollout

With a clear roadmap in place, the focus shifts to scaling with speed and consistency. This requires an empowered central team that operates as an enabler, that works closely with the business, regional and functional teams.

In practice:
As we scaled our end-to-end transformations from 2 to more than 70 deployments in three years, we enhanced the playbook continuously and established an internal academy. This academy enabled employees to follow a structured learning journey and certify their transformation skills. At the same time, we implemented clear communication strategies and dashboards to track adoption, progress, and value realized—keeping everyone aligned and accountable.


5. Drive Momentum

Scaling isn’t just about delivery—it’s about sustaining energy and belief. Even the best plans will stall without a mechanism to drive momentum and reinforce impact.

In practice:
In the ERP program, we implemented regular health checks to assess deployment progress and identify adoption risks early. Senior leaders were actively involved in championing each phase, reinforcing the “why” and spotlighting success stories across regions. We knew that technical go-live was just the beginning—the real success comes from business adoption, continuous improvement and realized value.


Final Reflections

Many transformations fail at different phases of the journey. Some stumble at the very start by choosing a pilot that is too complex to succeed quickly. Others struggle to make the leap from isolated success to broader relevance. Some lose steam by over-engineering the scale-up, while others fail to sustain momentum once the initial excitement fades.

Next to the examples of what worked, I’ve also seen where things went wrong—or could have gone significantly better. These lessons are just as valuable, and they underscore the importance of staying pragmatic, agile, and focused on value throughout the scaling process.

A successful digital-enabled transformation moves from MVP to full scale through a deliberate series of learnings, adjustments, and investments—not just in technology, but in people, process, and governance.

By starting with fast, meaningful wins, expanding through co-creation, and scaling with discipline, organizations can move from isolated experiments to enterprise-wide impact—with clarity, confidence, and speed.

Why Systems Thinking Is Crucial in Designing Digital Transformations

Digital transformations are hard. Despite bold ambitions, most still fall short—projects stall, teams get overwhelmed, and new technologies fail to deliver lasting value.

One major reason is that many organizations approach transformation in a fragmented, linear way—missing the underlying complexity of the systems they’re trying to change.

Systems thinking offers a powerful alternative. It equips leaders to design transformations that are more coherent, more efficient, and more likely to succeed. The approach helps teams see connections, anticipate ripple effects, and align initiatives across silos. Systems thinking increases the odds that transformation efforts won’t just launch—they’ll land, scale, and sustain real impact. It helps leaders see and communicate the bigger picture, connect the dots across silos, and design smarter interventions that actually stick.


What Is Systems Thinking?

Think of a transformation not as a set of initiatives, but as a connected system. Systems thinking helps you:

  • Spot how parts are connected—across tech, people, data, and processes.
  • Understand ripple effects—how changes in one area can help or hurt others.
  • Design smarter interventions—by targeting the right pressure points, not just the most obvious problem.

Example: Adding AI to customer support won’t drive impact unless you also rethink workflows, retrain staff, align incentives, and adjust how performance is measured. Systems thinking shows you the whole picture.


When to apply Systems Thinking?

Most transformations are cut into initiatives and get locked into “project mode” too early—jumping to solutions before fully understanding the system they aim to change.

That’s why systems thinking is most valuable during the Design and Scoping phase—from shaping strategy to turning it into an actionable plan. It helps:

  • Identify where the real bottlenecks are—not just symptoms.
  • Avoid siloed planning.
  • Create a roadmap that aligns required resources, impact, and ownership.

In the transformations I was involved in the Design and Scoping phase, I always used four angles: People, Process, Data and IT to look at the systems. In combination with bringing in the mindset to think through each of the topics, End to End, we looked across the silo’s and approached the transformation holistically.

In several of the transformations we leveraged experts in this field (like McKinsey, BCG, Accenture and Deloitte) and also publications from leading institutions were used as inspiration. Below a couple of excerpts of what they write about the topic.


What Thought Leaders Say applying System Thinking in Digital Transformation

MIT Sloan highlights that true digital transformation requires more than upgrading technology. Success depends on aligning tech, data, talent, and leadership—treating them as parts of one evolving system. Their research urges leaders to think integratively, not incrementally.

Harvard Business Review stresses the need for adaptive leadership in complex environments. Traditional linear planning falls short in today’s dynamic systems. Leaders must learn to coordinate across organizational boundaries and steer transformation with responsiveness and curiosity.

McKinsey & Company argues that managing transformation complexity demands a systems mindset. They emphasize the importance of understanding how processes, technologies, and people influence one another—revealing hidden dependencies that can derail progress if left unaddressed.

Deloitte offers practical tools to tackle so-called “messy problems” using systems thinking. They advocate mapping interactions and identifying root causes rather than reacting to surface-level symptoms. This approach is especially useful in large-scale enterprise or public-sector transformations.

Boston Consulting Group (BCG) connects systems thinking with platform innovation and agile ways of working. Their work emphasizes the importance of thinking in flows rather than functions—designing transformations around end-to-end customer, data, and value journeys.

Stanford University’s d.school and HAI combine systems thinking with design and AI ethics. Their research underscores the importance of aligning technology, people, and social systems—especially when integrating AI into existing structures. They promote a holistic view to ensure responsible and sustainable change.


How to Apply Systems Thinking in 5 Practical Steps

1. Define the Big Picture
What system are you trying to change?
Start by mapping the environment you want to influence. Identify key players, teams, technologies, and processes involved. Look at how value is created today, where it flows, and where it gets stuck. This helps frame the real scope of the challenge and ensures you don’t miss critical pieces.

2. Spot Key Connections
What influences what?
Once the landscape is clear, explore how the elements interact. Look for patterns, cause-and-effect relationships, and feedback loops. For instance, increasing automation may speed up service but also drive new types of demand. These dynamics are crucial for anticipating second-order effects.

3. Find the Pressure Points
Where can a small change make a big difference?
Focus on areas where a strategic adjustment could generate disproportionate impact. This might be a policy that shapes behavior, a workflow bottleneck, or a metric that drives priorities. The goal is to shift the system in ways that amplify positive change and reduce resistance.

4. Design the Roadmap Around the System
Move the whole system, not just parts.
Align your initiatives across domains—technology, data, process, people, and culture. Sequence interventions so that early wins unlock momentum for deeper shifts. Consider how one change enables another, and make sure efforts reinforce rather than compete with each other.

5. Build in Feedback and Learning
How will you measure and adapt?
Transformations unfold over time. Equip your teams with ways to detect what’s working, what’s not, and where unintended consequences arise. This includes system-level KPIs, qualitative insights, and space for reflection. The ability to course-correct is what makes a systems approach resilient.

Conclusion: The Payoff of Thinking in Systems

When systems thinking becomes a consistent practice, the result is not just better-designed transformation programs—it’s a smarter, more adaptive organization. Leaders begin to anticipate change instead of reacting to it. Teams work across boundaries instead of within silos. And investments create compounding value rather than isolated wins. Ultimately, systems thinking enables transformation efforts to scale with clarity, resilience, and lasting impact.