Go Back Up

LuminateCX Blog

From Search to Source: Why AI Is Rewriting the Role of the Website

Posted by Steven Muir-McCarey on Jun 11, 2025 9:16:42 AM
8 min read

Executive Summary: AI search engines are fundamentally changing how websites function, shifting from digital storefronts designed for human browsing to structured intelligence feeds optimised for AI consumption. This transformation requires organisations to adopt Generative Engine Optimisation (GEO) strategies, focusing on being cited rather than clicked, to maintain visibility in an increasingly AI-mediated digital landscape.

The website is no longer the front door: it's the intelligence feed. Here's what that means.

The End of Traditional Search and the Rise of AI Discovery

For two decades, we built websites to win the Google game. SEO was the operating system. Rank well, earn the click, convert the visitor. That funnel was gospel.

But 2025 has changed the rules. AI search engines (ChatGPT, Gemini, Perplexity) aren't just surfacing content. They're synthesising, deciding, and delivering. They're answering the question for the user, often without a single click.

Search Shift

"

60%+ of AI-powered queries now end in zero clicks. Discovery is abstracted. Conversion begins before your site is even seen.

Source: Adobe / Perplexity usage study, Q2 2025

What we're witnessing is not just a search engine evolution: it's a complete reordering of how digital visibility is earned.

AI Doesn't Browse. It Consumes.

Websites today are no longer scanned for layout; they're mined for meaning. AI platforms like Perplexity and Gemini read your site semantically, not visually. They quote schema, not slogans.

And if your content isn't clear, structured, and credible? You don't exist in the AI result.

This is GEO, not SEO

Welcome to Generative Engine Optimisation. Where the win condition isn't "rank me": it's "cite me." Your brand becomes the source, not the destination.

And the battlefield isn't search engines: it's AI interfaces.

Fragmented Journeys, Compressed Funnels

Today's digital journeys aren't linear: they're conversational, compressed, and AI-guided. A user researching a loan doesn't visit five bank sites. They ask ChatGPT for options, get a summary, and maybe click once to verify or transact.

What used to be:

Search → Browse → Evaluate → Decide

Is now:

Ask → Understand → Act

AI-referred traffic converts faster, with users averaging 10.4 minutes on-site versus 8.1 from Google. They arrive deeper in the funnel, primed and decisive.

The Website's New Role: Source Layer, Not Brochure

Leading organisations aren't rebuilding websites; they're repositioning them.

Instead of digital storefronts, they're becoming structured intelligence feeds designed to:

  • Train the AI: Feed high-fidelity, semantically tagged content to LLMs
  • Surface accurately: Prevent hallucinations by publishing "truth objects"
  • Convert contextually: Optimise the page the user lands on, not the homepage

What They're Doing Differently

1

Expedia

Embedded ChatGPT into its booking flow, capturing AI-driven travel planners before they bounce to a general answer.

2

HubSpot

Launched "AI Search Grader" to measure brand visibility inside AI results, not just Google rankings.

3

CommBank

Restructured product pages and deployed an AI assistant ("Ceba") to serve answers as conversations, lifting customer satisfaction and digital engagement.

4

Atlassian

Turned documentation into an AI-ready support layer, exposing structured help data to both users and LLMs.

These aren't UI tweaks. They're structural changes to compete in the new arena: AI discovery.

Being Seen Without the Click: Why AI Visibility Matters

In a zero-click world, being mentioned is as powerful as being clicked.

  • AI assistants cite sources, even if users don't click them
  • Brand trust is inferred from citations, not sessions
  • Some businesses are tracking AI impressions as the new KPI of digital visibility

The takeaway? You're not just optimising for the user; you're optimising for the AI that introduces you.

Governance, Signals and the Battle for Authority

As AI becomes the lens through which customers see the internet, websites need to think about their machine-readability and data rights:

  • llm.txt is emerging as a "treasure map" to help LLMs find your best content
  • Schema, APIs, and structured FAQs are now AI visibility multipliers
  • Meta tags like noai and selective robots.txt directives let brands control how they're used, but also risk invisibility if overused

The new question isn't just: "What pages do we publish?"

It's: "What data do we want AI to know, say, and cite?"

Strategic Questions for Digital Leaders

If you're leading MarTech, CX, or digital strategy, here's what should be on your radar:

1

Are we cited by AI models?

Run prompts in Gemini, ChatGPT, Perplexity and see if your brand shows up.

2

Can AI understand our offering?

Are our services, products, and differentiators machine-readable and well-structured?

3

Are we treating AI as a distribution channel?

Is your website training the intelligence your customers trust?

Final Thought: From Visibility to Veracity

AI won't stop at summarising. It will increasingly recommend, compare, transact.

Which means your website isn't just a content hub: it's a source of truth in a web of synthesis.

"

If you're not feeding the intelligence that feeds your customers, you may not even be part of the consideration set.

Want to Know Where You Stand?

At LuminateCX, we help digital leaders reimagine their web presence as an AI-optimised asset, fit for this new era of discovery and influence. Book a Spark Workshop with us to audit your AI visibility and citation readiness, design structured content frameworks, and plan for GEO and AI-distributed engagement.

Start the Conversation

Let's turn your website into the engine that fuels your next 1,000 conversions without waiting for the click.

Tags: MarTech, AI Search, AI Discovery, Generative Optimization

The New Rules of Getting Stuff Done

Posted by Steven Muir-McCarey on Jun 10, 2025 6:00:01 AM
6 min read

Executive Summary: Despite countless productivity tools, most professionals feel overwhelmed and unproductive due to misalignment between their energy and impact. The Productivity Blueprint addresses this by helping individuals identify friction zones and align their work with their strengths, potentially saving 3-4 hours per week through strategic elimination rather than optimisation.

If productivity was just about having better tools, we'd all be crushing it by now.

Instead, most of us are paddling through 12-hour days, 275 notifications, 30 tabs and still feeling behind. The apps are multiplying. The meetings keep stacking. And AI? So far, it's mostly just helping us write emails faster so we can get to… more emails.

So here's the hard truth: our productivity problem isn't effort. It's misalignment.

We've mistaken movement for momentum.

In 2024, Microsoft interviewed leaders across operations, marketing, and customer experience. What they found was startling but not surprising:

  • 87% of employees believe they're productive
  • Only 12% of CEOs agree

(Source: Microsoft Work Trend Index, 2024)

It's not just a communication gap; it's a productivity paranoia loop. People feel like they're falling short, leaders can't see the impact, and AI gets wheeled in like a miracle fix.

But instead of fixing the friction, we're automating it.

Most tools weren't built for the real problem.

When we talk to teams, we hear the same things:

"

I'm in meetings all day and can't find time to do actual work.

I use all the tools, but everything still feels reactive.

I'm exhausted. I'm doing my best work after hours.

Sound familiar?

That's not a time management issue. That's a misalignment between where your energy goes and where your impact lives. And no amount of AI integrations are going to fix that until we get clear on what matters and what doesn't.

Productivity isn't about doing more. It's about doing you better.

We built the Productivity Blueprint to help people get back to what they're good at. And more importantly, get out of what's draining them.

It takes five minutes. Four simple questions. No jargon, no signup, no fluff.

We don't ask how many emails you answer. We ask:

1

What kind of work actually energises you?

2

What work do you do just because it's expected?

3

What's draining you?

4

What would you happily never do again?

The result? A personal report that shows:

  • Where your time is going (and where it shouldn't be)
  • What work you can automate, re-assign or ditch
  • Where AI and systems might help, but only after the human work is sorted

It's not a to-do list. It's a stop-doing list.

The big unlock: alignment.

Most productivity frameworks obsess over time saved. Ours focuses on value created.

What we've found is this: when you align people with their sweet spot—high-value, energising work, everything else starts to flow. Tasks get delegated smarter. AI actually helps, because it's supporting the right work, not masking the wrong ones. Teams communicate better because they're not drowning in noise.

And the wild part? Just identifying the friction zones (we call them your "Energy Drain" and "Elimination List") saves an average of 3–4 hours per week.

That's a half day back every week.

If it feels broken, it's not you.

One of the most consistent emotions we've seen in our work is guilt.

"

Why can't I stay on top of everything?

Why does this tool make me feel more behind?

Why can't I focus like I used to?

Here's the thing: it's not because you're lazy or distracted or bad at time management.

It's because most systems are designed to optimise the process, not the person.

You weren't built to be a Slackbot. You were built to solve problems, create ideas, lead people, build things. But if your calendar's full of noise, you'll never get to the signal.

So what now?

If you've made it this far, I'm not going to give you five steps to inbox zero or some templated morning routine.

I'm going to suggest this instead:

  • Take 4 minutes.
  • Answer 4 questions.
  • Get a clear blueprint of how you work best.

It's free. No email capture. No spin.

Just a tool that helps you get out of the friction and back into flow.

Because productivity doesn't start with AI, or apps, or tools. It starts with you.

Ready to Reclaim Your Productivity?

Take the Productivity Blueprint and discover how to align your energy with your impact. In just 4 minutes, you'll get a personalised report showing exactly where to focus—and what to eliminate.

Get My Blueprint

Tags: Productivity, AI Productivity, Blueprint, Meaningful Work

Unanswered but Not Forgotten: Your Questions After Humanising AI

Posted by Steven Muir-McCarey on Jun 3, 2025 5:00:01 AM
10 min read

Executive Summary: Following our Humanising AI event in Brisbane, we address the most compelling audience questions about responsible AI adoption, maintaining critical thinking skills, and preserving human value in an increasingly automated world. This reflection explores practical approaches to AI governance, cultural transformation, and the balance between convenience and creativity.

Some of the best questions from our Humanising AI event in Brisbane didn't come from Dan Shaw, our MC to the panel. They came from you.

They were submitted via Slido, or asked quietly in conversations and captured. And while we didn't get to respond to them all in the moment, they still hold value and are worth reflecting on. They reflected the tension we're all grappling with right now, not just how we adopt AI, but how we do it responsibly, practically, and without losing the very things that make our organisations human.

So I wanted to take a moment, not to answer everything, but to respond with what we're thinking, learning, and still exploring at LuminateCX.

1. What happens to our ability to think for ourselves?

One of the big themes that emerged was fear, not of robots taking jobs, but of humans forgetting how to think critically. There were questions around generational EQ, university readiness, and whether AI convenience is making us lazy.

And I get it. There's a risk we're not talking about enough.

We've all had that moment where we tried to force an AI output into shape and thought, "I should have just done this myself." That's not just a UX issue... it's a red flag.

If we let convenience override creativity, we're not augmenting intelligence. We're outsourcing it.

At LuminateCX, we see this as a design responsibility. AI should be a catalyst, not a crutch. The value is in co-creation, where human thinking is enhanced, not replaced. And yes, that means protecting the time and space for ideas to breathe, not just execute.

2. Should AI be governed like nuclear energy?

There were pointed questions about ethics, regulation, and global oversight. Should there be AI treaties? International guardrails?

The short answer: probably yes. The better answer: don't wait for them.

AI is already behaving more like infrastructure than tooling. And that means governance can't be an afterthought. At LuminateCX, we look at responsible AI through a multi-layered lens and risk isn't one size fits all. We assess based on exposure, access, and consequence. What data is in play? What decision could this output trigger? Who or what could it affect downstream?

Responsible AI is values-led, context-aware, and risk-adjusted. And just like document classification, it needs clear categories. Public. Internal. Confidential. Critical.

These boundaries shouldn't slow teams down, they should give them the confidence to move faster, within the right lane.

3. Why is AI adoption moving faster than our culture can catch up?

Many asked: Why are organisations letting users self-learn AI without support? Why aren't leaders more involved in guiding this shift?

"

This is the AI paradox. We buy fast and adopt slow. We chase potential but forget permission.

Culture doesn't scale because you rolled out Copilot. Change happens when people are brought along, when their questions are heard, their fears addressed, and their creativity respected.

We always tell clients: don't start with a platform or technology. Start with your people. Find one team, one task, one problem worth solving. That's where adoption becomes trust, and trust becomes momentum.

4. Isn't it just a matter of time before AI builds and runs everything?

There was a question about websites running themselves. Another about Klarna rolling back its AI play. And plenty about the future of jobs.

Here's where I land. It's still day zero. No one is going to accurately predict AI's impact two years out, not with the speed of change we're seeing. But what we can say is this: the energy you used to spend on repetitive tasks is going to shift. Autonomy will replace drag. The next challenge is deciding where you'll apply that energy instead.

The work isn't going away. It's just moving further along the process, into strategy, judgement, empathy, and nuance. That's where humans win. That's where value lives.

5. What makes a tool worth trusting?

We had several asks for tool recommendations. My honest answer? Tools are changing faster than you can learn them. If you're picking tools without building capability, you're not scaling, you're stacking problems.

At LuminateCX, we focus on foundational literacy:

1

Understand the architecture

What's powering this tool?

2

Know the role

Is it an assistant, orchestrator, or decision-maker?

3

Stress-test governance

What data is it pulling from, and where is it pushing to?

4

Align value zones

Where is the human in the loop, and where should they be?

Good tools disappear into good systems. Great tools make your people more confident, not confused. That's the test.

6. What are we missing? What are we ignoring?

Someone asked, "What lesson from the printing press or the internet are we at risk of ignoring?"

That's the question that's still ringing in my head for varying reasons.

If I had to answer today, I'd say: we're underestimating the emotional layer of transformation. In every technological shift, we focus on speed, cost, and efficiency. But what about trust? What about the sense of meaning people get from their work? This is also an immense opportunity for all of us to find a path to leveraging this technology to drive ideas, personal lives and businesses further.

"

That's what we can't afford to bypass. Because when those things erode, culture collapses quietly, long before your AI strategy fails.

The conversation doesn't end here

We won't have all the answers. But we'll keep asking better questions. We'll keep learning out loud. And we'll keep building frameworks that allow organisations and the people inside them to do their best work alongside AI, not in spite of it.

If one of these questions is alive in your organisation, let's talk. Book a Spark session. Reach out. Or just send a note.

We're not just building strategies. We're shaping what comes next.

Ready to Navigate Your AI Transformation Journey?

Connect with Steven Muir-McCarey and the LuminateCX team to explore how we can help your organisation adopt AI responsibly whilst preserving the human elements that drive real value.

Start the Conversation

Tags: AI Adoption, Responsible AI, Humanising AI, AI Culture, Critical Thinking

The Hidden Cost of Single AI Model Thinking

Posted by Steven Muir-McCarey on May 12, 2025 7:00:00 AM
"

Is your AI strategy delivering results, or just running up the bill?

OpenAI has announced it will deprecate GPT-4.5 in July. It costs 30 times more than GPT-4o, yet offers only marginal gains. This isn't just a pricing change. It's a signal.

Many organisations are still deploying AI like it's 2022, one large model, stretched across every use case. That approach isn't scalable or cost-efficient. It's a liability.

Today, there's increasing access to fit-for-purpose models from open-source options like Mixtral and LLaMA to enterprise tools like Claude, Gemini and GPT-4o. Each brings different strengths depending on the task, domain or cost to value.

The companies getting this right aren't choosing "the best model." They're orchestrating across many. They route simple prompts to efficient models and reserve premium power for high-value use cases.

This article breaks down the real cost of single-model thinking and what it takes to build a smarter, leaner AI strategy in 2025.

The Era of One-Model AI Is Over

Not long ago, deploying AI meant choosing a model and pushing it across as many use cases as possible. That strategy might have been passable in 2022, but with what we have available now, it's not just an outdated approach, it's an expensive one at that.

We've moved from only having access to a handful of generalist models to many purpose built models and although this could be seen as noise, it's more of a lever waiting to be orchestrated. Even within just the big players  we see that OpenAI offers GPT-4o, o4-mini and 5+ other models on their platform. Google provides Gemini 2.5 in both "Pro" and "Flash" editions as well as 10+ other variants. Anthropic's Claude 3.7 Sonnet introduces hybrid reasoning variants and flexible pricing and access to 3+ other variants of its models. 

What matters isn't just the name or brand. It's that the cost-performance delta between these models is now massive.

  • GPT-4.5: $75 per million input tokens, $150 for output, with only 7% performance improvement over GPT-4o (Ojha, 2025; TechCrunch, 2025)
  • GPT-4o: $2.50 input / $10 output per million
  • Claude 3.7 Sonnet: $3 input / $15 output, and it handles multi-step reasoning (Anthropic, 2025)
  • Gemini 2.5 Pro: $2.50 input / $15 output per million
  • Gemini 2.5 Flash: $0.15 input / $0.60 output for fast-response tasks 

Ai Model Use across needsv2

Use the wrong model for the wrong task and your costs don't double, they can increase by thousands of percent.

Behind the Curtain: Smarter Models, Not Just Bigger Ones

The evolution of AI models in 2025 is marked by a strategic shift towards Mixture of Experts (MoE) architectures, with a focus on efficiency and specialisation over brute-force scale.

Meta's Llama 4 series, released in April 2025, exemplifies this approach. Models like Llama 4 Scout and Maverick use MoE designs that activate only a fraction of their total expert pathways per task. For example, Llama 4 Maverick includes 128 experts, but uses just 17 billion active parameters per inference, a configuration that significantly reduces compute costs while retaining high performance (Dataconomy, 2025).

But perhaps more importantly, enterprises are now beginning to design and fine-tune bespoke MoE models for specific domains: customer service, finance, compliance triage. These tailored models don't aim to outperform general-purpose giants like GPT-4. Instead, they deliver 90% of the value at a fraction of the cost, while aligning tightly with business objectives.

This isn't just a technical upgrade. It's a strategy shift away from "best model overall" toward "best model for this job, at this price."

As model orchestration matures, modularity becomes a competitive lever. Companies embracing smaller, task-specific MoEs are gaining precision, control, and commercial agility.

What Orchestration Looks Like in Practice

Some companies have already made the shift.

A major auto manufacturer uses Microsoft's lightweight Phi model to handle the bulk of incoming customer queries. Only the edge cases get routed to GPT-4o. The orchestration reduced their AI spend by 80 percent, while improving response quality.

Sage, the accounting software company, fine-tuned Mistral for their domain, using it as a triage layer before escalating to a larger model. It improved resolution accuracy and delivered faster responses.

This is what orchestration looks like. It's not just better performance, it's operational leverage.

Vendors are starting to catch up:

  • Azure AI Studio now offers over 1,800 models with routing capabilities
  • AWS Bedrock provides a unified interface for multiple foundation models
  • Google Vertex AI hosts more than 200 models in its Model Garden

Startups like Martian are building routing layers that dynamically switch models based on task complexity. Customers report 70 to 90 percent reductions in model costs with no drop in output quality (TechCrunch, 2025).

Two Strategies, Same Use case, Radically Different Outcomes.

Let's look at how this plays out in practice. Below, two companies deploy models across similar use cases, but only one does it with orchestration in mind.

Feature Company A: "Premium Everything" Approach Company B: "Right Tool for the Job" Approach
Primary AI Strategy Use Claude 3.7 Sonnet for all tasks, ensuring premium quality across the board. Route tasks to the most appropriate model based on complexity and performance requirements.
Models Used &
Task Allocation
Claude 3.7 Sonnet exclusively for:
• Customer Support FAQs                      (7M interactions) • Investment Research Summaries    (2M interactions) • Real-time Trading Analysis                 (1M interactions)
Optimized Mix:
Gemini 2.5 Flash for Customer Support FAQs (7M interactions) GPT-4.1 for Investment Research Summaries (2M interactions) Claude 3.7 Sonnet for Real-time Trading Analysis (1M interactions)
Token Calculation
(Monthly)
Input: 10M interactions × 500 tokens = 5,000M tokens
Output: 10M interactions × 150 tokens = 1,500M tokens
Input: (7M × 500) + (2M × 500) + (1M × 500) = 5,000M tokens
Output: (7M × 150) + (2M × 150) + (1M × 150) = 1,500M tokens
Estimated Monthly
AI Spend
Claude 3.7 Sonnet (100% volume):
Input Cost: 5,000M × $3/M = $15,000
Output Cost: 1,500M × $15/M = $22,500
Total: $37,500
Gemini 2.5 Flash (70% volume):
Input: 3,500M × $0.15/M = $525
Output: 1,050M × $0.60/M = $630
Subtotal: $1,155
GPT-4.1 (20% volume):
Input: 1,000M × $2/M = $2,000
Output: 300M × $8/M = $2,400
Subtotal: $4,400
Claude 3.7 Sonnet (10% volume):
Input: 500M × $3/M = $1,500
Output: 150M × $15/M = $2,250
Subtotal: $3,750
Total: $9,305
Outcomes &
Efficiency
Customer FAQs: Excellent quality but massive overkill. Slower response times than Flash. 26x higher cost than necessary.
Research Summaries: Excellent quality, but at premium price.
Trading Analysis: Excellent quality, appropriately resourced.
Overall: Consistently high quality but severe cost inefficiencies and suboptimal user experience for high-volume tasks.
Customer FAQs: Perfect quality for use case, 5-10x faster responses, minimal cost.
Research Summaries: Better document analysis with GPT-4.1's 1M token context and engineering focus.
Trading Analysis: Top-tier analysis with Claude 3.7 Sonnet for critical decisions.
Overall: Task-optimized performance, superior user experience, maximum ROI.
Business Impact Very high operational costs. Budget inefficiently allocated to low-complexity tasks. Customer frustration with slower response times for simple queries. Premium pricing may need to be passed to clients.
75% reduction in AI costs ($28,195 monthly savings)
Faster customer service resolution. Better research quality with specialized model. Demonstrates sophisticated AI strategy. Competitive advantage through efficiency.

*Pricing approximate & in USD. Source: OpenRouter.ai

The only difference is that Company B didn't just choose models, they built a model strategy.

What Smart AI Strategy Looks Like

The orchestration approach isn't about jumping on the next new model. It's about building a framework that supports:

  • Task audits – Start with what each part of your business needs
  • Cost-to-value mapping – Not every prompt deserves frontier-level performance
  • Routing logic – Route based on complexity and risk, not convenience
  • Ongoing optimisation – Measure performance/cost by model and task type and adjust routing rules accordingly.

This is what separates maturity from experimentation. Without it, you're not building AI systems, you're running expensive experiments at scale.

Time to Unpack What's Really Happening

Companies still relying on single-model AI deployments will be running tech strategies from a previous era. It's no longer enough to have "an AI model." You need a way to manage, optimise, and evolve your AI footprint in line with business goals.

"

The winners in this economy won't have the best model. They'll have the clearest orchestration strategy.

Let's Make Sense of It

Curious where your AI investments might be underperforming?
Our Spark Workshop is designed to help leadership teams unpack what's really happening across their AI stack.

We'll help you:

  • Identify where model decisions are driving unnecessary costs
  • Pinpoint orchestration opportunities already hiding in your business
  • Explore how intelligent routing can unlock performance and  cost control

This isn't about switching platforms. It's about getting clarity.

Book a Spark Workshop

References 

Anthropic. (2025). Claude 3.7 Sonnet and Claude Code. https://www.anthropic.com/news/claude-3-7-sonnet 

Ars Technica. (2025). "It's a lemon": OpenAI's largest AI model ever arrives to mixed reviews. https://arstechnica.com/ai/2025/02/its-a-lemon-openais-largest-ai-model-ever-arrives-to-mixed-reviews/ 

Meta / Llama 4:
Dataconomy. (2025, April 7). Meta launches new Llama 4 AI models: Scout, Maverick now available in apps. Retrieved from https://dataconomy.com/2025/04/07/meta-launches-new-llama-4-ai-models-scout-maverick-now-available-in-apps/?utm_source=chatgpt.com 

Mistral / Mixtral MoE:
Artetxe, M., Rabe, M. N., Scao, T. L., Tunstall, L., de Masson d'Autume, G., & Elbayad, M. (2024). Mixtral of MoE: Sparse mixture of experts models with open weights (arXiv:2401.04088). arXiv. https://arxiv.org/abs/2401.04088 

Google Cloud. (2025). Vertex AI Pricing. https://cloud.google.com/vertex-ai/generative-ai/pricing 

Ojha, A. (2025). The great paradox: Why OpenAI's most expensive model GPT-4.5 falls short of expectations. https://medium.com/@ayushojha010/the-great-paradox-why-openais-most-expensive-model-gpt-4-5-falls-short-of-expectations-4c3c5035a692 

TechCrunch. (2025). OpenAI plans to phase out GPT-4.5. https://techcrunch.com/2025/04/14/openai-plans-to-wind-down-gpt-4-5-its-largest-ever-ai-model-in-its-api/ 

Tags: Zero-to-Solve, AI Strategy, Ai Budget Orchestration, Multi-Model AI Strategy, Mixture of Experts, Enterprise LLM

The Human in the Machine: What AI Still Needs From Us

Posted by Steven Muir-McCarey on May 6, 2025 5:02:38 PM
"

AI doesn't know your brand. Or your people. It can learn patterns, generate content, even sound convincing. But it has no memory of what your team went through last year. It doesn't understand your customers' frustrations. It doesn't carry your culture.

We're standing at a strange intersection. AI is everywhere, embedded in our workflows, our tools, even our job descriptions. Yet for many business leaders, it still feels distant. Shiny. Unsettling. Unproven.

In the rush to integrate AI, the narrative has been hijacked by tech vendors and productivity metrics. Platforms promise acceleration, automation, and insight at scale. But the question we're wrestling with at LuminateCX is more human:

What happens to culture, creativity, and connection when we hand more of our decision-making to machines?

As we prepare for Humanising AI Brisbane 2025 and mark one year of LuminateCX, I've been reflecting on what "humanising AI" really means. Not just as a concept, but as a strategy. A leadership responsibility. A design choice.

When AI Strategies Fail, It's Rarely About the Tech

Most AI pilots don't stall because the tools weren't smart enough. They stall because the people weren't brought along.

We've seen this play out in real-time: leadership teams chasing AI initiatives without first answering why it matters to their people, how it fits their culture, or what problems it actually needs to solve.

"

An AI strategy does not start with just simply buying a piece of technology and implementing it employee-wide. The true strategy is to first look inwardly — to your people and your processes.

That's not just a philosophical stance. It's a pattern we've observed across dozens of organisations.

  • Tech teams light up a GPT tool... but nobody uses it.
  • Marketing automates content... but brand voice gets diluted.
  • HR explores AI-powered onboarding... but employee trust erodes.

Too often, AI is introduced as a layer on top of the business — not within it.

And when that happens, three things suffer:

1

Trust

Teams feel blindsided, not empowered.

2

Culture

The 'why' gets lost in the 'wow'.

3

Creativity

People defer to AI instead of developing their own voice.

"

Simply turning on AI capability doesn't mean people will use it. Unless they see the value — how it solves a problem or benefits them — it's just another tool gathering dust.

So What Does Humanising AI Look Like?

It starts small  and it starts with people.

The term "humanising AI" gets thrown around a lot. But in practice, it's not about warm fuzzy branding or philosophical debates. It's about design. Leadership. Practical choices in how AI is introduced, structured, and scaled across the organisation.

"

My initial take of what humanising AI means is really, how do I incorporate this latest technology into helping me do what I find is important — for me, for my job, for the people and systems I work with.

At LuminateCX, we've found that the most successful AI adoption stories don't begin with automation targets, they begin with understanding friction. Where are people stuck? What drains their energy? What could help them feel more effective?

Humanising AI means looking at your workflows and asking:

  • "Where can we remove complexity without losing connection?"
  • "What micro-tasks are ripe for augmentation, but still need a human in the loop?"
  • "How do we build trust in these systems through visible wins?"
"

Going into your organisation and asking 'what can we automate?' is not a recipe for success. It pushes back on your people.

Instead, the better question is:

"What's a repetitive task your team wants help with  and where they'd welcome support, not fear replacement?"

This approach does more than drive productivity. It builds momentum. You're not just integrating tech,  you're creating buy-in, enabling experimentation, and embedding AI into your culture without breaking it.

"

Where there are small wins that are of value to automate — where the expert in the middle still plays a role — that's where AI adoption becomes trusted, useful, and human.

This Isn't About the Future. It's About Right Now

We've moved past prompts. AI isn't something we experiment with on the side anymore, it's operating behind the scenes, shaping decisions, streamlining processes, and driving change whether we plan for it or not.

"

Those systems will increasingly become almost like organisational AI. They won't just respond — they'll orchestrate. Multi-step tasks, decision-making flows, process automation... carried out under the hood, with more abstraction at the surface.

That abstraction means less visibility and more risk.

If you don't actively shape how AI behaves in your organisation, it will still operate… just without your values in the loop.

That's when things start to break.

Brand gets diluted. Trust erodes. People start checking out — or worse, they bypass the system entirely.

That's why humanising AI isn't a philosophical nice-to-have. It's a strategic necessity.

You don't need to have all the answers. But you do need a plan.

"

Right now, the opportunity is to bring all of your teams to a base level of competency on the use of AI. If you're not doing it, your competitors are.

So Where Do You Start?

Not with a platform. Not with a pilot.

Start with your people.

"

AI doesn't know your company. It doesn't understand your culture, or your customers. Your people do. There's so much richness in their experience — and that's what should shape how AI is used, not the other way around.

The Humanising AI Approach: A Practical Flow

At LuminateCX, we've seen that humanising AI isn't a one-off initiative. I t's a sequence of leadership choices made over time. Here's how we frame it:

Listen → Localise → Layer → Learn

4Ls Humanising AILCX


You don't need a perfect strategy on Day One. But you do need to move and move in a way your people can believe in.

Ready to Start Your Humanising AI Strategy?

We're helping organisations move beyond pilots and into meaningful, people-first AI integration. And it doesn't start with a 100-page report. It starts with a conversation.

Book a complimentary 90-minute Pulse session to explore your current state, uncover where AI can add value, and set the stage for a strategy that fits your people, process, and purpose.

Book a Pulse Session

Tags: Digital Transformation, AI Adoption, AI Strategy, CIOLeadership, Humanising AI, Organizational Culture

Your AI Pilot Worked. That’s Exactly Why It’ll Fail

Posted by Steven Muir-McCarey on May 6, 2025 7:00:00 AM

Let's not sugar-coat it.

If your AI pilot went well, there's a good chance it's headed straight for the scrapheap.

Gartner just made it official: by the end of 2025, 30% of generative AI projects that prove successful in pilot will be abandoned before reaching production. Not because the tech didn't work—but because the business wasn't ready.

You don't need better AI.
You need a better plan for what comes next.

The Pilot Trap: Why Early Success is Misleading

Successful AI pilots are designed to prove the art of the possible—not the rigour of the real.

They're built on:

  • Handpicked, well-labelled datasets
  • Controlled use cases
  • Isolated environments with minimal risk
  • Tight-knit project teams with deep support
  • Zero integration with legacy platforms

But once the pilot ends, reality begins.

Suddenly you're facing:

  • Disparate data quality across business units
  • Conflicting priorities between IT, ops, and execs
  • Integration complexity with platforms that weren't part of the pilot
  • Governance, compliance, and risk questions that never got asked
  • A board expecting results while your team scrambles for alignment

That's not a technical problem.
It's a framework problem.

GenAI Pilot Purgatory: The New Enterprise Bottleneck

HFS Research coined the term "GenAI Pilot Purgatory" to describe this exact scenario:
Where innovation gets stuck between validation and value.

The technology works—but the business isn't built to scale it.

We've seen this play out in both public and private sectors. Pilots get board-level airtime, but when it's time to integrate with mission-critical systems, the gaps appear. AI initiatives are derailed not by model drift, but by organisational friction.

McKinsey summed it up well:

"

Business leaders face increasing pressure to generate ROI from their GenAI deployments.

And that pressure is real. Especially when every other line item in your budget is being challenged.

Scaling AI Requires More Than a Rollout Plan

Here's the common failure pattern we see:

1

Launch a pilot to test a use case

2

Celebrate early success

3

Attempt a scale-up without foundational readiness

4

Hit resistance across governance, integration, and adoption

5

Quietly abandon the effort—or delay indefinitely

It's the equivalent of building a show home before securing zoning approval.

If you're solving for tech first, you're solving the wrong problem.

From Pilot to Production: The Ignite Approach

At LuminateCX, we built Ignite to help organisations avoid exactly this scenario.

Ignite is not about improving your pilot.
It's about creating the strategic conditions that allow pilots to scale—with confidence.

Here's how we do it:

1

Audit Where You Really Are

We don't rely on success theatre. We analyse your actual architecture, governance readiness, data posture, and stakeholder alignment—across systems, silos, and scenarios.

2

Define What You Actually Need

That means:

  • What's the measurable business impact you're targeting?
  • What operating models need to shift?
  • What technical enablers are missing for scale?
  • Where are the internal points of friction?

This isn't about use cases. It's about use value.

3

Build the Framework for Scale

We develop a structured approach to:

  • Governance and data risk models
  • Integration pathways and technical dependencies
  • Change and adoption strategy
  • Metrics aligned to business outcomes, not just technical performance

The goal isn't to get AI into production—it's to keep it in production and deliver meaningful returns.

4

Create a Strategic Roadmap

Not a tech wishlist. A pragmatic, sequenced pathway from now to next, with clear ownership, executive sponsorship, and milestones that withstand internal scrutiny.

What the Data (and Our Experience) Actually Tells Us

AI success isn't about picking the right tool—it's about building the right conditions for scale.

We've seen the pattern across sectors—whether it's a government agency piloting language models for service delivery, or a national insurer testing claims automation. The tools work in isolation. The frameworks often don't exist.

And while few organisations have mastered AI at scale, the red flags show up early:

  • Governance gaps
  • Shadow adoption without oversight
  • Mismatched expectations between tech and business units
  • No clear ROI model beyond the pilot phase

At LuminateCX,  our team have spent their careers guiding clients through complex digital transformations—across MarTech, data integration, and cloud migration. While GenAI is newer, the underlying transformation risks are familiar.

We know what it looks like when frameworks are missing—and we know how to build the ones that unlock momentum.

You've Got 8 Months

That's how long you have before Gartner's prediction becomes your reality.

Eight months to either:

  • Join the 30% of organisations that never move beyond the pilot
  • Or build a foundation that makes AI a core capability—not a proof of concept

Your Next Strategic Move

Before you double down on scaling your AI pilot, take a day to pressure-test your readiness.

Book a Spark Session—a structured, advisory-led engagement that identifies the key blockers, risks, and opportunities before you commit capital or credibility to a flawed rollout.

Let's ensure your pilot is the starting line and not the finish.

Next Step: Start With the Right Conversation

Book a Spark Session with our team to assess your AI readiness and build a framework for success. Don't let your pilot become another statistic.

Book Your Spark Session

The Navigator's Take

"Your AI pilot is working perfectly. That's the problem. The winners in this next wave won't be the ones who prototype the best—but the ones who plan for production from day one."

— Steven Muir-McCarey

Tags: Digital Transformation, Generative AI, AI Strategy, AI Governance, Enterprise AI, AIatScale, CIOLeadership

Pilot Purgatory to BAU

Posted by Steven Muir-McCarey on May 1, 2025 10:53:12 PM

Introduction

When OpenAI released its AI in the Enterprise guide, it offered a compelling blueprint for AI adoption. The seven lessons—drawn from companies like Klarna, BBVA, and Morgan Stanley—are instructive and well-articulated.

But for most enterprises, especially in regulated industries or mid-market environments, getting from AI pilot to business-as-usual (BAU) requires more than inspiration. It calls for a structured path, organisational readiness, and a mindset shift. This article offers a strategic reflection on OpenAI's lessons, adding context, caution, and concrete steps. The goal: help enterprises move from experimentation to embedded capability.

1. Start with Evals — but Choose Metrics That Matter

OpenAI recommends: Start with systematic evaluation processes to measure how AI models perform against use cases.

Why it matters
Morgan Stanley's early success with generative AI stemmed from carefully structured model evaluations (OpenAI, 2024). By rigorously assessing summarisation, translation and relevance, they created confidence to scale.

LuminateCX Insight: Many enterprises over-index on model benchmarks without linking them to operational impact. A model might hit 95 percent accuracy and still fail to change outcomes. Without alignment between eval metrics and business KPIs, evals create false confidence.

Actionable step: Develop eval frameworks that blend technical benchmarks with real-world metrics—such as cycle time, task completion rate, or reduction in manual effort. Include compliance and safety as explicit criteria.

2. Embed AI in Products — but Rethink the Journey Too

OpenAI recommends: Embed AI into your products to create smarter, more responsive customer experiences.

Why it matters
Indeed's use of AI to explain job matching improved engagement and conversion, driving a 20 percent increase in applications started (OpenAI, 2024).

LuminateCX Insight: Too often, enterprises layer AI features onto existing tools without rethinking the broader experience. This leads to confusion or degraded UX. The AI feature may be technically impressive but not functionally useful.

Actionable step: Use AI to redesign the workflow, not just add a feature. Co-design with users. Pilot with human-in-the-loop controls. Make it easier, not just more advanced.

3. Start Early — but Only If You're Ready

OpenAI recommends: Invest and start early to gain compounding value.

Why it matters
Klarna's early investment led to an AI assistant now handling two-thirds of customer chats, delivering a projected $40 million in efficiency gains (OpenAI, 2024).

LuminateCX Insight: Some organisations start before governance, security, or alignment is in place. Samsung's data leak via ChatGPT use by engineers led to an internal ban and reputational risk (TechCrunch, 2023).

"

Starting early is not the same as starting recklessly.

Actionable step: Begin AI adoption early—but with sandbox environments, policy frameworks, and risk assessments in place.

4. Fine-Tune When Necessary — but Don't Default to It

OpenAI recommends: Customise and fine-tune models for specific use cases.

Why it matters
Lowe's improved product tagging accuracy by 20 percent through fine-tuning GPT models (OpenAI, 2024).

LuminateCX Insight: Fine-tuning increases operational complexity. It creates maintenance overhead, locks in assumptions, and limits flexibility when base models evolve. Fine-tuned models can also become outdated quickly and require constant retraining (OpenAI Community Forum, 2024).

Actionable step: Use prompt engineering and retrieval-augmented generation first. Only fine-tune if performance gains justify long-term cost and effort. Maintain documentation for versioning and compliance.

5. Empower Experts — but Ensure Support Structures Exist

OpenAI recommends: Put AI in the hands of the people closest to the process.

Why it matters
BBVA enabled over 125,000 employees to create their own GPT-powered apps. This decentralised innovation accelerated adoption and surfaced unexpected value (OpenAI, 2024).

LuminateCX Insight: Giving people access to AI doesn't mean they will use it effectively. Many teams lack training, support, or incentives. AI tools without enablement lead to inconsistent adoption.

Actionable step: Provide onboarding, training, guardrails, and coaching. Launch internal AI Champion programs to ensure AI literacy and responsible use across departments.

6. Unblock Developers — and All Other Stakeholders

OpenAI recommends: Automate software development processes to accelerate AI delivery.

Why it matters
Mercado Libre built a platform (Verdi) that allowed over 17,000 developers to accelerate AI-enabled applications, improving inventory, fraud detection, and customer experience (OpenAI, 2024).

LuminateCX Insight: Bottlenecks rarely exist in development alone. Security, legal, procurement, and compliance often delay deployment—especially in regulated industries.

Actionable step: Establish an AI Delivery Council that includes compliance, security, and operations early in the design process. Map AI delivery friction points across the organisation, not just in code.

7. Set Bold Automation Goals — but Stay Grounded

OpenAI recommends: Aim high with AI. Don't limit yourself to low-hanging fruit.

Why it matters
Ambitious automation can transform processes and free capacity. OpenAI automated internal support workflows, increasing efficiency (OpenAI, 2024).

LuminateCX Insight: Overpromising AI potential without operational discipline leads to failed projects and executive fatigue. Not all processes are ready for automation.

Actionable step: Use a value-risk framework to prioritise automation. Start with rule-based, repetitive tasks with measurable ROI. Keep humans in the loop where judgment is required.

From Pilot to BAU: What It Really Takes

Moving AI into BAU doesn't just mean deploying it. It means:

  • Clear ownership and accountability
  • KPIs linked to operational or customer outcomes
  • Change management and communication
  • Ongoing model monitoring, retraining, and compliance review

LuminateCX Insight: AI in production must meet the same standards as any enterprise system: secure, supported, and sustainable.

LuminateCX BAU Readiness Checklist

Before scaling an AI use case, ask:

1

Is this project aligned to business KPIs, not just experimentation goals?

2

Do we have compliance and risk controls in place?

3

Has the impacted team been trained and consulted?

4

Do we have the infrastructure to monitor and retrain this model?

5

Is there a fallback plan or human override mechanism?

If the answer to any of the above is "no," the project may not be ready to move to BAU.

Conclusion

OpenAI's seven lessons are a valuable starting point. But AI adoption doesn't succeed through vision alone. It requires readiness, execution discipline, and clarity on what "good" looks like post-pilot.

"

For enterprise leaders, the challenge is not in launching AI pilots—it's in embedding AI in ways that are safe, valuable, and scalable.

For enterprise leaders, the challenge is not in launching AI pilots—it's in embedding AI in ways that are safe, valuable, and scalable. That's what separates AI as hype from AI as a lever for transformation.

Next Step: Turn Insight Into Action

LuminateCX works with enterprise and government leaders to assess readiness, design governance-aligned AI roadmaps, and embed AI capabilities responsibly.

  • Schedule a Pulse consultation to assess where you sit on the BAU maturity curve.
  • Or engage in an Ignite workshop to build a full operationalisation plan.
Schedule a Consultation

References

Tags: AI Strategy, LuminateCX, Business Transformation, AI Governance, Enterprise AI, Responsible AI, OpenAI

The Business Adoption Curve: Why Frameworks Beat Platforms

Posted by Steven Muir-McCarey on Apr 10, 2025 10:00:00 AM
"

If you're not climbing, you're sliding.

That's how Shopify's  CEO Tobi Lutke framed it in an internal memo to employees, declaring that effective AI usage is no longer a competitive edge. It's the baseline. AI is now an expected skill, a default tool in every knowledge worker's toolbox. The shift from optional to expected is happening in real time.

It's a sentiment I've begun to hear echo across boardrooms, client strategy calls, and executive workshops. And yet, behind the noise and momentum, there's a growing risk many businesses aren't seeing.

The real danger isn't doing nothing. The real danger is doing something without a plan.

Outperforming the average in today's AI-flooded economy, what I once described as 'beating the bell curve' for the individual, won't come from buying a shiny platform or deploying a headline-grabbing tool. It will come from building the internal capability: the thinking, the frameworks, the architecture that allows your organisation to adapt as fast as the technology itself evolves. And especially in this space, change will be constant and far more rapid than we have seen in 20 years.

Falling Behind the Business Adoption Curve

In a previous article, I explored how individuals could outperform the average by combining their lived domain expertise with AI's leverage. But my thoughts are now evolving to what this means for business in a version of this very same predicament. What started as a conversation about individuals is now a high-stakes question for entire organisations.

There is a new curve at play, the business adoption curve, and the gap between early movers and hesitant followers is widening. Organisations that have already begun integrating AI into their systems and processes are accumulating compound advantage. Every test run, policy refinement, and use case validated today builds future momentum.

By contrast, those that have delayed adoption face a reverse compounding effect. It becomes harder to catch up. Capabilities expected by customers and employees begin to feel like technical debt. AI-native competitors operate faster, leaner, and with clearer confidence. What's normal for them starts to feel unattainable for you.

The result is an exponential gap in productivity, innovation, and talent retention. And for those still sitting on the sidelines, the longer the wait, the steeper the climb.

Evolve AI Adoption Curvev2

Why Tech-First Thinking is a Strategic Trap

In times of change, the instinct to “just do something” is strong. But in AI, that instinct often leads to the wrong kind of action.

Many organisations assume that buying into a single AI platform means they’ve made a strategic leap. But not all AI integrations are equal, and many introduce more rigidity than value.

At the rigid end of the spectrum are bespoke AI tools and wrappers. These are often built around a single model or function. While they can feel impressive in a demo, they’re brittle, hard to scale, and disconnected from broader business systems.

Next are existing SaaS platforms that have added AI features as an afterthought. CRMs, ERPs, and ticketing systems now offer predictive fields or summarisation, but these capabilities are typically vendor-driven and may not align with the workflows that actually matter to your business.

Then comes the most commonly adopted tool: ChatGPT. It offers wide-ranging utility, but even within the OpenAI ecosystem, different models perform better at different tasks. The user has limited control over which model is being used. While future iterations like GPT-5 will likely introduce more dynamic orchestration, today's experience remains constrained.

Microsoft Copilot has just offered a more integrated environment. Initially reliant on OpenAI models, it has recently expanded to include models from Meta and others, allowing for more flexible reasoning across use cases. This is a positive shift, but orchestration logic remains locked inside the Microsoft ecosystem.

Further along the spectrum are orchestration-layer platforms like Manus and Genspark.ai. These tools abstract model and tool selection entirely. Users engage with a single interface, while multiple models are used behind the scenes to complete more complex tasks. The experience is powerful, but users typically have little visibility or control over how orchestration decisions are made.

At the most flexible end of the spectrum are the integrators: tools like Workato, Boomi, Make.com, and Zapier to name a few. These platforms allow teams to define how systems, tools, and AI models interact. They enable true composability and give organisations control over how intelligence is embedded across workflows.

In reality, organisations often operate with a mix of layers across this spectrum. But where the centre of gravity sits determines whether AI becomes an asset or a liability.

AI Specturm Rigidity to Flexibility_LCX

This is the real risk: not just picking the wrong tool, but building around the wrong idea. AI is not something you install. It’s something you design into your operating model. It needs to evolve with your people, your systems, and your priorities.

The businesses gaining the most from AI today are not those moving the fastest. They are the ones building with flexibility, composability, and the clarity to avoid short-term decisions that lead to long-term constraints.

What Businesses Actually Need: Composable AI Frameworks

We've seen this pattern before. In cloud adoption. In marketing technology. In ERP transformations. The lesson is always the same: platforms don't solve problems, systems thinking does.

Here's what leading organisations are doing instead.

1

Architect for change

Design your stack to be model-agnostic. Prioritise interoperability and avoid long-term vendor lock-in. This doesn't just protect you, it sets you up to experiment and evolve without structural overhaul.

2

Embed AI literacy across the business

You don't need every employee to be a prompt engineer. But you do need everyone to understand where AI can augment their decisions, workflows, and outcomes. Skills enablement should be continuous, embedded, and role-relevant.

3

Build learning loops, not fixed projects

The most successful AI programs behave more like agile experiments than enterprise rollouts. Build in feedback, share learnings across teams, and make iteration your core capability.

4

Apply a governance mindset from day one

Ethics, security, and explainability must be part of every conversation, not bolt-on compliance later. Build frameworks for safe experimentation and responsible deployment.

5

Align AI with business value, not hype

Focus on real problems: customer experience gaps, efficiency gains, decision-making acceleration. Use cases should be prioritised not by technical possibility, but by strategic relevance.

Why Frameworks Beat Big Platforms

There's a fundamental difference between building for adaptability and buying for convenience. Big AI platforms promise integration, automation, and acceleration. But what they often deliver is rigidity.

They're hard to replace. They limit optionality. They centralise decision-making in systems rather than people. And they rarely keep pace with open model innovation.

By contrast, a composable framework gives your organisation the ability to swap, integrate, and extend capabilities as needed. It's the same logic behind modern integration platforms and composable MarTech stacks, applied to AI.

You don't build around one vendor's roadmap. You build for your own.

Operationalising AI Without Overcommitting

Some will argue that doing something is better than doing nothing. But rushed adoption without architecture creates more problems than it solves.

The goal is not just to introduce AI into the business. It's to make it operational. That means:

  • Mapping use cases to workflows
  • Equipping teams to use the tools effectively
  • Measuring real impact, not superficial engagement
  • Ensuring responsible usage with built-in guardrails

You don't need to choose a model today. You need to build the capacity to choose wisely tomorrow, and again the day after that.

Final Thought: The Curve Is Moving

The organisations moving today with clarity, strategy, and humility are setting themselves up for a decade of advantage. Those stuck in a holding pattern or placing big bets on singular platforms are gambling their agility.

AI is not a passing trend. It is a new operating environment.

The businesses that thrive in it will not be those who acted fastest. They'll be the ones who acted most intelligently, with frameworks that adapt, systems that scale, and teams who know why they're using it in the first place.

Next Step: Start With the Right Conversation

If you're ready to move beyond AI noise and into structured, scalable adoption, we can help. Our Spark Sessions are designed to help businesses build clarity around AI strategy, without the pressure to buy into platforms too early.

You don't need to pick the perfect model today. But you do need to begin building the right foundation.

Let's get that started.

Book a Spark Workshop

Tags: Digital Transformation, Zero-to-Solve, AI Adoption, AI Strategy, Organisational Agility, Composable Architecture, Enterprise Technology, Framework vs Platform, Execution at Scale, LuminateCX, Steven Muir-McCarey

Zero-to-Solve: The Execution Layer Has Moved

Posted by Steven Muir-McCarey on Apr 7, 2025 10:03:00 PM

The Real Shift: It’s Already Happened

Most teams are still chasing AI disruption. But the sharpest operators have already moved on because the real disruption? It’s behind us.

While the market obsessed over prompts, copilots, and generative UI tricks, something quieter but more consequential took root: a fundamental shift in where execution happens.

The execution layer has moved. It’s no longer buried in your dev backlog, your delivery pipeline, or integration stack. It now lives at the edge, where intent meets outcome, instantly.

This is the Zero-to-Solve era: a world where business users, strategists, and domain experts don’t just generate ideas, they deliver them. Not with dev tickets, but with AI-native tools that turn prompts into products, ideas into automation, and bottlenecks into velocity.

The question is no longer “Can we build it?” It’s “How fast can we go from thought to solution?”

From Curiosity to Capability

Remember when GPT and Claude first hit the mainstream?

We treated them like clever assistants, smarter search engines, better writers, AI sidekicks to help tidy content or clean up code.

But the true inflection point wasn’t when they started speaking back. It was when they started executing unprompted, unassisted, and with intent.

Take Claude’s Artifacts feature. What looked like a UI enhancement was something far deeper: the emergence of AI as a systems thinker. The model wasn’t just answering, it was coding, styling, and structuring the output inside an interactive canvas without being explicitly told to (Anthropic, 2024).

That’s when the shift happened.

We stopped prompting, and started prototyping. In real time. With real outcomes.

Suddenly, AI wasn’t just responding. It was reasoning, building, and delivering without waiting to be told.

And just like that, we crossed the line from disruption to democratisation.

Zero-to-Solve: Execution at the Speed of Intent

Zero-to-Solve isn’t a feature. It’s a new operating model where execution moves at the speed of thought.

No more waiting on prioritisation, pipelines, or tickets. If you can describe it, the system can deploy it.

One user built and launched a fully functional note-taking app on Bolt.new in under two minutes—starting from a blank canvas, using nothing but plain language prompts (The Prompt Warrior, 2025).

This isn’t code generation. It’s product delivery.

It’s what happens when infrastructure fades into the background and intent becomes the API.

Execution is no longer gated by expertise. It’s triggered by clarity of vision.

We call this vibe coding. A new paradigm where the builder defines the outcome, and the system handles the logic, syntax, and deployment path (Garg, 2025).

Describe what you want. The stack responds. Zero-to-Solve doesn’t just accelerate workflows. It rewrites them.

The Rise of the Citizen Developer

We used to say, “If you can dream it, you can build it.” But that used to mean: funding, engineers, velocity planning, and three months of backlog wrangling.

Now, you describe it. And it ships.

AI-native platforms have collapsed the gap between idea and execution. Non-technical users, product managers, marketers, operations leads, are turning plain language into production-ready tools, workflows, and systems. No code required. No approvals queued.

The new builder stack includes platforms like:

  • Windsurf: An agentic IDE that applies multi-file edits using Flows and Cascade (Codeium, 2025).
  • Cursor: A transparent AI coding assistant where users see and approve diffs in real time.
  • Bolt.new: A browser-native environment that deploys entire apps from natural prompts (Refine.dev, 2025).
“If you can describe it, you can deploy it.” That’s not hype. That’s the new workflow.

We’re not just seeing hobbyist momentum here. Inside enterprises, this shift is already reshaping internal velocity. Business users are shipping prototypes, integrations, and internal tools in hours....not weeks....without ever touching a line of code.

It’s not a threat to engineering. It’s leverage. Execution is decentralising and accelerating.

Agentic Architecture & Orchestrated Workflows

Once you’ve built something, the next question isn’t “does it work?” it’s “how does it connect?”

This is where execution breaks or compounds.

We’ve entered the era of agentic architecture: AI-native systems that take a goal, break it into parts, and assign each part to a specialised agent, coordinated, parallel, and autonomous.

  • One agent pulls data from your CRM or product database.
  • Another drafts content or analysis based on that data.
  • One validates against compliance or brand guidelines.
  • Another pushes the result into Slack, HubSpot, or Notion instantly.

This isn’t just automation. It’s orchestration, with awareness of sequence, logic, dependencies, and outcomes.

The future of delivery isn’t point-and-click. It’s define-and-delegate.

Orchestration layers like LangChain (for chaining tools and memory), MindStudio (for no-code agent design), and Zapier, Make.com, and Flowise (for low-code automation) are making this composable by design (Joyce Birkins, 2025).

What used to take a product team, an engineer, and a delivery roadmap is now handled by a network of agents with your intent as the trigger.

This is where execution scales without bureaucracy.

Model Context Protocol: The Standard for Execution

Orchestration unlocked intent-based workflows. But without a shared foundation, every AI system still spoke its own language.

That’s where the Model Context Protocol (MCP) comes in.

Developed by Anthropic, MCP is the emerging connective tissue for AI ecosystems. Think of it as the USB-C of AI tooling: one universal interface that lets models discover, use, and coordinate tools, data, and actions natively.

Without shared context, agents are just freelancers. With MCP, they become a team.

Instead of bespoke APIs and rigid integrations, MCP introduces a common server layer exposing:

  • Tools: like send_email or query_database
  • Resources: such as files, customer records, or policies
  • Prompts: structured templates for consistent execution

This allows agents to maintain memory, select tools mid-task, and share execution context across sessions, platforms, and models.

Early adopters include Claude, Block, Zed, and Codeium, with more layering in as interoperability becomes essential, not optional (Philschmid, 2025).

MCP isn’t just about plug-and-play AI. It’s about building a networked execution layer where agents, tools, and platforms work as one.

The AI Operating Systems: Where Tools Become Teams

Citizen Developer

When orchestration, vibe coding, and context protocols converge, we don’t just get better AI tools, we get a new class of operating system.

These are full-stack AI execution environments, built to take goals, not just instructions. Designed to deliver outcomes, not just assistance. And they’re already working in the wild.

The AI operating system is no longer science fiction, it’s a strategy execution layer. Built with agents, powered by prompt logic, deployed at edge velocity.

Manus: The Agent Workforce

Manus is a cloud-native AI OS built around agent-first architecture. You assign a task like “analyse 500 CVs and generate a shortlist” and Manus handles the orchestration.

No prompt loops. No micro-managing. Each part of the job is handled by a specialised agent: coding, summarising, browsing, formatting. They work in parallel. They coordinate autonomously. And they deliver.

This isn’t a prototype. It’s a workforce.

Manus was hired on Upwork and Fiverr, completed jobs, generated deliverables, and got paid (WorkOS, 2025). It’s not pitching capabilities. It’s operating in marketplaces.

Genspark.ai: The Generalist Super Agent

Genspark takes a different path to the same destination: a mixture-of-agents model built on LLMs, integrated tools, and real-time orchestration.

Its Super Agent executes full workflows using over 80+ specialised tools—from travel planning to restaurant booking, video generation, voice calling, and even animated content creation.

"I want the AI to book all the restaurants on this trip for me..."
The agent dials, speaks with a human, considers food allergies, and requests a window seat—all autonomously. (Genspark Demo, 2025)

It doesn’t stop there:

  • Plans and books 5-day travel itineraries using map and research tools
  • Generates videos from recipes or trending news with voiceovers and sound effects
  • Supports marketers, teachers, analysts, and recruiters with fully packaged tasks

Why does it work? Because Genspark combines:

  • Large Language Models
  • Toolsets (for real-world action)
  • Datasets (for nuance and context)

Together, these make it fast, reliable, and steerable—ready to execute across everyday knowledge work.

The Bigger Signal

What both Manus and Genspark.ai show isn’t competition, it’s convergence.

They represent the next phase of AI delivery: not assistants helping operators, but systems that become the operator.

For Zero-to-Solve thinkers, these platforms are not about replacing people, but about rethinking who (or what) delivers value in your execution model.

Zero-to-Solve: The Operating Model of Now

Zero to Solve - visual selection

Let’s connect the dots.

The platforms. The agents. The protocols. The citizen developers. All of it is converging toward one undeniable shift:

We’re no longer building tools. We’re building ecosystems.

And ecosystems don’t just scale, they compound. Each new agent, workflow, or integration increases your organisational surface area for execution. AI becomes less of a layer, and more of a substrate, something your operations are built on.

Zero-to-Solve isn’t about speed. It’s about proximity to action. The shortest path from intent to outcome wins.

If your delivery model still depends on:

  • Manual prioritisation
  • Velocity bottlenecks
  • Backlog gatekeeping

…then you’re not building for capability, you’re building for delay.

It’s time to redesign the way work gets done.

Final Word: Own the Execution Layer

So here’s the question every leader should be asking:

Are we building for control—or for capability?

Because the execution layer has moved. And those who see where it went will own what comes next.

Let's Cut Through the Noise

At LuminateCX, we help leaders:

  • Separate signal from hype
  • Identify execution leverage points
  • Build AI-native workflows and Zero-to-Solve roadmaps that actually ship

Let’s design your execution model for what’s real, not just what’s possible.

🧾 References

  • Garg, J. (2025). Vibe Coding: Concept, Workflow, AI Prompts, Tools. Medium.
  • Codeium. (2025). Windsurf Editor. codeium.com.
  • Refine.dev. (2025). Bolt.new – AI App Builder. refine.dev.
  • The Prompt Warrior. (2025). Bolt vs. Cursor. promptwarrior.com.
  • Joyce Birkins. (2025). 16 AI Workflow Platforms. Medium.
  • Philschmid, P. (2025). MCP Overview. philschmid.de.
  • Anthropic. (2024). Introducing the Model Context Protocol. anthropic.com.
  • WorkOS. (2025). Introducing Manus. workos.com.
  • AI Base News. (2025). Genspark Super Agent. aibase.com.

Tags: AI Revolution, AI Personalisation, AI distruption in SaaS, Zero-to-Solve, Citizen Developer, AI Execution Layer, Prompt Engineering, No-Code Development

Beyond the Hype: A Real Conversation About AI, Strategy, and Business Enablement

Posted by Steven Muir-McCarey on Apr 6, 2025 4:39:38 PM

Introduction

It’s hard to avoid the buzz. If you’re a head of marketing or digital leader in 2025, you’ve probably seen more AI webinars, thought pieces, and platform updates in the last six months than in the previous five years combined.

But as the noise ramps up, many leaders are left with more questions than answers: What should we actually be doing? Are we behind? Is AI really going to replace jobs, or is it just another overhyped trend?

In this conversation, I sat down with my Business Partner Dan Shaw as he explored his questions for me, not from the perspective of a tech evangelist, but from a grounded, business-first lens. Here’s what we uncovered.

The AI Conversation Isn’t Just About AI

Despite the flood of AI content, the biggest shifts aren’t just technological. As I shared early on:

I think my first comment is actually not about AI. Instead, the real story is workforce transformation.

Over the last five years, we’ve seen an accelerated evolution in how businesses operate from hybrid work models to restructured teams and pressure to do more with less. AI has become part of that story, but not the whole story.

We've brought people on to continue to accelerate businesses, and now we’ve got headlines around AI transforming everything and people losing jobs. I don’t love that narrative, but there is a fundamental shift happening that we need to prepare for.

Fear vs Focus: The Right Mindset for AI Adoption

Let’s get one thing clear: the fear is real, but it’s misplaced.

It’s not wrong to feel uncertain. Budgets are tightening. Teams are shrinking. And the noise around AI is relentless. But fear shouldn’t paralyse action.

I think the fear is genuinely not doing anything. If I was in a business right now pushing back on AI, I don't think I could just continue to ignore it.

2024 was the buffer year. In 2025, the window to wait and see is closing.

AI Is Not a Silver Bullet

A critical theme we unpacked was this: AI will not save you if you don’t know what problem you’re solving.

Dan shared a scenario typical for many leaders today: where a head of marketing is grappling with reduced budget, fewer people, and constant pressure to deliver more. Vendors are knocking. AI features are everywhere. But what do you actually do?

Wouldn't it be awesome if you could just buy an AI tool and it magically solves your resourcing and budget problems? But it's never that simple.

Just like rolling out a CRM or ERP, adopting AI requires understanding:

  • Your Data & internal processes
  • Your team’s IP, and how to nurture & apply that knowledge
  • Your brand voice & culture fit both internal & external
  • Your customer nuances, and CX expectations. 

AI is powerful, but it’s only as good as the data, structure, and context you give it.

Strategy First, Always

Too often, AI is treated as the strategy. But as I emphasised during the chat:

It can't be a technology-first approach. AI is just a ubiquitous word and it can mean so many things. The key question is not which tool should I buy, but rather What business problem am I solving?

In your business today, are you trying to automate parts of your workforce, speed up customer response times, generate better marketing content, or improve data visibility and decision-making? Until you define that, any AI initiative is just a shot in the dark.

Tap Into Existing Momentum

One of the most overlooked opportunities in early AI adoption? Your own people.

There are already individuals inside your organisation quietly using AI tools to speed up tasks, improve work quality, or test ideas. Rather than fearing shadow IT, bring these people in.

People are learning how to adapt AI. Bring them in. Give them a structure. Understand what’s working and scale it safely.

This doesn’t require enterprise-wide investment, just curiosity, frameworks, and an appetite for experimentation.

Apply What You Already Know: CRM Lessons for AI

We often think of AI as something entirely new. But the blueprint for adoption already exists.

Think back to your CRM transformation. You didn’t just install software, you defined processes, cleaned up data, trained users, and built trust across departments. AI should be no different.

Businesses are capable of leaning into existing frameworks—like their CRM rollout process, and applying that same lens to AI. 

It’s not about reinventing the wheel. It’s about reapplying proven practices to a new wave of capability.

What Comes Next: AI as the New Operating System

Perhaps the most forward-looking part of our conversation was this: AI is moving from ‘nice to have’ to ‘core business infrastructure.’

We’ve gone from ERP in the ’90s to CRM in the 2010s. Now, I really see AI becoming the core operating system of the business into the next decade.

This doesn’t mean every business needs to rush out and build custom models or replace their SaaS stack. But it does mean we’re seeing a shift from AI as a bolt-on tool to AI as an integrated enabler across marketing, operations, customer service, and more.

Key Takeaways

  • Acknowledge the shift – AI isn’t just hype. It’s part of a broader workforce transformation.
  • Don’t wait – The time for “watch and see” has passed. Start exploring.
  • Start with strategy – Define the business problem before picking a tool.
  • Leverage internal champions – Find the people already experimenting and build on their momentum.
  • Use existing frameworks – Apply what you learned from past tech rollouts (like CRM).
  • Think long-term – AI is becoming foundational. Lay the groundwork now for what’s coming next.

Final Thoughts

This isn’t about panic. It’s about progress. AI isn’t here to take your job, but it might just accelerate your future, if you’re willing to engage with it intentionally. It’s a moment of opportunity for those who choose to lead, not just react.

 

 

Not sure where AI fits in your business strategy?



Start with a Pulse Session — a no-commitment, high-value conversation designed to help you cut through the noise, define the real opportunity, and get clear on what matters next.

At LuminateCX, we don’t sell technology — we help you make it work. Independently. Strategically. Sustainably.

 

Tags: Content Strategy, AI and Knowledge Management, Generative AI

Changes in the Marketing Unit? Here's how your team can navigate 2025 and maintain performance.

Posted by Dan Shaw on Mar 30, 2025 9:52:00 PM

Let’s not sugarcoat it...2024 was a chaotic year for some.

Budgets were slashed. Teams were shuffled. AI hype reached fever pitch. And while everyone tried to sound excited about “doing more with less,” many CMOs were quietly asking: What happens when the ‘less’ keeps getting smaller?

Now it’s 2025. You’re still here. That’s the good news. The bad news? The pressure hasn’t let up, it’s just changed shape.

So, if your team is leaner, your timelines are tighter, and your performance targets haven’t budged, this post is for you. 

Let’s break down what’s actually happening in marketing teams right now, and how you can build a performance system that doesn't collapse every time there's a restructure or a new tech buzzword. 

Key takeaways:

 

Here are the main points from the conversation: 

  • Performance is a system. Sustainable growth comes from building an operating model that flexes with change, not from sprinting harder every quarter.
  • Focus is the new speed. High-performing teams aren’t faster, they’re more focused. Ruthlessly prioritise, kill vanity projects, and simplify execution.
  • AI is only half the answer. Automation only drives value when it’s embedded in how your team works, not just what tools you use.
  • Cross-functional alignment unlocks efficiency. RevOps and Marketing Ops should share goals, data, and accountability to drive consistent performance.
  • Leaner teams require smarter workflows, prioritise ruthlessly, reduce handoffs, and empower cross-skilled contributors to own full campaign loops.

Why Marketing teams are evolving: AI, RevOps, and the drive for efficiency. 

The traditional marketing model, siloed functions, big teams, bloated tech stacks, is being quietly dismantled. 

Three forces are reshaping the unit:

  1. AI and Automation: Not just the shiny stuff. Yes, generative content tools matter, but the real shift is in how AI is quietly stitching together workflows, data, and insights that used to be spread across five teams and three platforms.
  2. RevOps Thinking: Revenue Operations isn’t just a B2B thing anymore. It’s a mindset shift, less about channels, more about customer flow and commercial alignment. Teams are being measured less on brand vanity metrics and more on pipeline, conversion, and impact.
  3. Efficiency Mandates: Boards and CFOs aren’t asking “How big is your reach?” anymore. They’re asking, “Why are we spending that much to get this result?” Marketing is now a cost center under a microscope. You need to prove you're a growth engine, not just a storytelling studio. 

If it feels like you’re being asked to do more with fewer people, less budget, and tighter turnaround times, you're not imagining things.

Adapting to leaner teams: smarter workflows, better prioritisation. 

The hard truth is that most teams aren't underperforming, they're overextended. 

When everything is urgent, nothing gets done well. The fix? Ruthless prioritisation and intentional workflow design.

A few high-leverage moves:

  • Kill your vanity projects. If it doesn’t drive pipeline, productivity, or performance, then park it.
  • Design workflows around outputs, not departments. Cross-skill your team so one person can own a full campaign loop, from ideation to launch, without passing it through five stakeholders.
  • Implement sprint cycles for marketing. Agile isn’t just for developers. Two-week sprints with clear priorities beat sprawling quarterly plans that break the moment something changes. 

This isn’t about working harder, it’s about eliminating friction, ambiguity, and busywork. 

Aligning with other business units: RevOps, MOps, and cross-functional efficiency.

Marketing can’t be the lone wolf anymore. 

The orgs that are thriving in this environment have cracked one thing: cross-functional clarity. They’ve aligned their marketing operations with sales ops, CX, and finance. 

You don’t need to merge departments. You need to merge goals and data.  Things to consider: 

  • Get in the same room (or Slack channel) with Sales and CX. Share dashboards. Debate the funnel. Challenge assumptions.
  • RevOps and MOps need a shared map. One that connects customer journeys to backend processes, to revenue levers.
  • Don’t just align on metrics, align on incentives. If Sales wins and Marketing doesn’t, the model’s broken. If CX insights aren’t influencing campaign strategy, your flywheel has a flat tire.. 

This is where a lot of teams get stuck. They have great strategy but terrible execution because execution is happening in silos. Smash those silos. 

The role of automation. Doing more with less (without burning out). 

Automation isn’t just about speeding things up, it’s about making space for higher-value work.

Here’s how smart teams are using it:

  • Tactical automation. Think campaign cloning, triggered journeys, dynamic content generation. Anything that eliminates copy/paste work.
  • Strategic automation. Using AI to identify patterns in performance data, surface next-best actions, or simulate scenarios based on customer behaviour.
  • Cultural automation. Embedding automation into how you work, not just what you use. Example: auto-prioritising requests based on business value. Or using AI to draft briefs that get 80% of the way there. 

The best automation feels like a natural extension of your team’s brain, not a bolt-on solution.  

Performance is a system, not a sprint. 

Even though perfect MOps is tailored to each organisation, one thing is clear: 

The teams that will win this year aren’t the ones with the biggest budgets. They’re the ones who: 

  • Know how to focus under pressure, 
  • Build systems that don’t break under change, 
  • And collaborate like their bonus depends on it (because it probably does). 

This isn’t about survival. It’s about building a marketing team that can thrive in disruption. That knows how to pivot, prioritise, and perform, without running itself into the ground. 

You don’t need to move faster. You need to move smarter. 

And if that sounds like a cliché, it’s only because most teams still haven’t figured out how to do it. 

Want help turning your MOps into a performance engine (even with a leaner team)?

We help teams align strategy to execution, build automation that matters, and unlock performance through smarter processes.

Contact us today to book a strategy session, and get started on strengthening your Marketing Operations.

 

STAY ACROSS INDUSTRY TRENDS

Register for Updates

Register to our regular email to stay across industry updates and changes in the AI, Cyber Security, MarTech, Customer and Marketing world.

Tags: Marketing, Operations, Digital Transformation, Strategy, MarTech, Digital Engagement, CX, Customer Experience, Customer Experience Innovation, MOps

How to Convert Your Board or Executive Team on a DXP Project

Posted by Dan Shaw on Mar 19, 2025 9:51:28 AM

Getting executive buy-in on a DXP project is challenging...especially in the current climate.

Digital leaders know a Digital Experience Platform (DXP) can transform customer engagement, streamline content management, and future-proof digital operations. But getting the board or executive team to approve the investment? That’s where things get tough.

Dan Shaw and Anthony Hook, recently sat down to discuss the challenges digital leaders face in securing executive buy-in for DXP projects. Their conversation highlighted the real reasons decision-makers hesitate, how to frame the business case, and what digital leaders can do to push their projects over the line.

Here is the video and below a summary of their conversation.

 

Key takeaways:

 

Here are the main points from the conversation: 

  • Executives don’t care about the technology, only the business impact. Focus on revenue growth, cost reduction, or risk mitigation.
  • Avoid vendor driven decisions. If a software provider is pushing you to upgrade without a clear internal business need, stop and reassess.
  • Make the investment feel manageable. Break the project into phased investments with quick wins in the first 3-6 months.
  • Align internal teams before going to the board. Marketing, IT, and finance must be on the same page to avoid internal roadblocks.
  • Find an internal board or executive champion. Having a senior advocate within the organisation increases the chances of approval.
  • Decisions are made over time, not in a single meeting. Build support through informal conversations before the formal pitch.
  • Use real business data to support your case. Instead of vague benefits, quantify the impact.

Executives are skeptical, and for good reason. 

There’s a reason DXPs face resistance at the executive level: decision-makers have been burned before. 

According to Anthony, many senior executives and board members have seen previous “all-in-one” digital platforms promise big results, only to fall short. 

“The promises were the same five or ten years ago: better personalisation, reduced costs, streamlined marketing ops, and a platform that would do it all. But now, with the push toward headless, composable, and API-first architectures, those same promises are being dressed up in new terminology.” 

This has led to decision fatigue and skepticism. Many executives are not technologists, so when vendors push composable DXPs, API-first stacks, or headless CMS, it all starts to feel like another expensive experiment. 

The three big objections to DXPs.

When boards and executive teams push back against a DXP investment, their concerns usually fall into three categories: 

  1. Cost vs. ROI"We’ve spent millions on digital platforms before. How is this different?"
  2. Complexity"We don’t have the resources to take this on."
  3. Risk Aversion"If this goes wrong, it’s my head on the chopping block." 

Dan emphasised that these concerns aren’t just excuses, they reflect a real problem in how DXPs are pitched. 

“If you walk into the boardroom talking about ‘headless CMS’ or ‘API-first architecture,’ you’ve already lost them. They don’t care about the technology. They care about business impact: Will this make us more money? Will it reduce costs? Will it lower our risk?” 

The mistake many digital leaders make? They focus on the technology instead of the business outcome. 

Reframing the conversation: Make it about business value. 

To get a DXP project approved, the pitch needs to shift from technology to business impact. 

Executives care about three things:

  • Revenue Growth – How does this drive sales or improve retention?
  • Cost Reduction – Can this reduce operational inefficiencies?
  • Risk Mitigation – Will this make compliance, security, or governance easier? 

“If you can’t link your DXP investment to at least one of these three things, it’s going to be a tough sell,” said Anthony. 

A soft pitch sounds like this: 
“A composable DXP will unify our digital channels and create an omnichannel content strategy.” 

A strong pitch sounds like this: 
“Right now, 30% of customer service calls are from people struggling to find answers online. A personalised self-service experience powered by a DXP could cut that in half—saving us $5M per year.” 

Don’t let software vendors set the agenda. 

Another common issue is that too many DXP projects are vendor-driven, not business-driven. 

According to Anthony, executives often feel pressured into upgrades because vendors create “compelling events”—whether it’s an end-of-life product, a new software trend, or a fear-based sales pitch. 

“Just because your current CMS or DXP is ‘outdated’ doesn’t mean it’s actually a business problem. Vendors will always tell you that you need to upgrade. But is this upgrade solving a real problem, or are you just reacting to external pressure?” 

Dan agreed, adding that businesses need to take back control of the conversation: 

“Before you even start talking to vendors, make sure you’ve done your own internal analysis. What do you actually need? What problems are you solving? If you don’t control the narrative, you’ll end up solving the wrong problem.”  

Break the investment into key phases. 

Another major reason DXPs fail to get approval is because they’re seen as too big and too risky.

The solution? A phased investment strategy.

Dan explained that boards don’t want massive multi-year projects—they want to see quick wins before committing to long-term investments.

“The most successful DXP projects don’t ask for everything up front. Instead, they start with a small, high-impact use case that delivers ROI in 3-6 months.”

Example of a Phased DXP Investment:

  1. Phase 1 (Quick Win) – Implement AI-powered personalisation for top customers, leading to a 5% lift in repeat purchases.
  2. Phase 2 (Scaling the Impact) – Expand automation into marketing and sales, reducing manual effort and improving conversion rates.
  3. Phase 3 (Full Integration) – Roll out across all customer touchpoints, creating a truly seamless experience.

“This de-risks the investment and gives executives confidence in the long-term vision,” Anthony added.

Winning internal support - it’s not just about the board. 

Even if the board is convinced, other stakeholders inside the organisation can block your DXP project.

“Most big organisations are political ecosystems. Even if the CEO supports the project, if marketing and IT aren’t aligned, it’s going to hit roadblocks,” said Dan.

Marketing teams want agility and flexibility.
IT teams want control and security.
Finance teams want predictable costs and ROI.

Before going to the board, align these groups internally. Find an internal champion—someone who will advocate for the project at the executive level.

“It’s not a one-and-done conversation. These things are won in the hallways, in one-on-one chats, in coffee meetings. The board meeting is just the final step.” 

Final thought: Play the long game. 

Securing executive buy-in for a DXP investment isn’t about selling technology. It’s about framing the conversation around business value, managing risk, and proving ROI in small, strategic steps.

“Winning approval isn’t about a single meeting. It’s about shaping the conversation over time,” said Dan.

“Executives don’t need to be convinced that DXPs are amazing. They need to see how this will drive business success,” added Anthony.

Need help making your DXP business case?

We help digital leaders translate complex technology into board-ready business cases. If you need help getting your project approved, we can guide you through the process.

Contact us today to book a strategy session, and get started on moving forward on your DXP project.

 

STAY ACROSS INDUSTRY TRENDS

Register for Updates

Register to our regular email to stay across industry updates and changes in the AI, Cyber Security, MarTech, Customer and Marketing world.

Tags: Marketing, Operations, Digital Transformation, Strategy, MarTech, Digital Engagement, CX, Customer Experience, Customer Experience Innovation, MOps

Something Powerful

Tell The Reader More

The headline and subheader tells us what you're offering, and the form header closes the deal. Over here you can explain why your offer is so great it's worth filling out a form for.

Remember:

  • Bullets are great
  • For spelling out benefits and
  • Turning visitors into leads.

Recent Posts

Posts by Tag

see all