I don’t know if you’ve noticed, but we’re in an “AI or die!” era right now. Every organization seems desperate to figure out how AI fits into their business strategy. Over the past year, we’ve talked with several teams at large enterprises that span industries – everything from global financial services to established automakers with heavy regulation burdens – to see how they’re actually using and managing their AI-embedded applications. Right now, there’s a flurry of definitions floating around about what makes an application “intelligent” but only a handful of use cases, best practices, and patterns have been established.
When it comes to the “enterprise” in enterprise applications – meaning the safety, compliance and performance we consumers have come to expect from apps that are built and maintained by reputable brands – we’ve seen two big hurdles to delivering Ai-embedded apps. First, teams need to acquire new AI skills and tools. Second, they must prove a solid business case—ROI and all—if they want their AI adoption to actually stick and scale beyond early PoCs and experiments.
When we talk about “enterprise” applications, there are two obstacles that pop up frequently. First is compliance: I’m seeing teams scramble to update long-standing security and governance frameworks just to accommodate AI models, even making temporary exceptions to keep projects moving. Worse, maybe they just ignore them, practicing “shadow AI.” Then there’s the execution barrier—teams need new AI skills and tooling, plus a solid business case with clear ROI. Otherwise, any pilot project risks becoming a one-and-done experiment instead of a sustainable practice.
We want to share one approach that can help organizations avoid getting caught up in blockers: framing your approach to add “intelligence” to applications with an enterprise AI maturity model can help guide you toward your goals. This framework clarifies what it means to add AI to your applications and, more importantly, shows how to ensure you get a real payoff. In other words, it’s your path to those elusive “business outcomes” we’re all chasing with AI.
Using a maturity model as a tool for an emerging technology can seem stodgy at first. But I see it more as a map of what to expect and a “default” plan to start with. As you’ll see, it also clarifies how to navigate the skills and ROI challenges.
This AI app delivery maturity model maps out four stages toward more “intelligent” apps: (1) pre-AI applications that follow a three-tier model, (2) adding machine learning, (3) adding generative AI (, and (4) creating agentic AI apps and workflows. However, each of these stages relies on the same basic ingredients—servers, frameworks, and clear business requirements. As you move up the model, it’s really about how much agency you add to the app.
Let’s take a look.
Stage 1: Three-Tiered Apps
Before you even get to using AI, you start here: a classic three-tier architecture consisting of a UI, app frameworks and services, and a database. Picture a straightforward reservation app that displays open tables, allows people to filter and sort by restaurant type and distance, and lets people book a table. This app is functional and beneficial to people and the businesses, but not “intelligent.”
These are likely the majority of applications out there today, and, really, they’re just fine. Organizations have been humming along thanks to the fruits of a decade of “digital transformation” for a long time. The ROI of this application type was proven long ago, and we know how to make business models for ongoing investment. Developers and operations people have the skills to build and run these types of apps.
This provides the baseline we’re operating from and helps us judge what “better” is* and* if it’s worth the price.
Stage 2: Adding Machine Learning Intelligence
Before generative AI (GenAI), machine learning was the next stage. This type of intelligence relies on data scientists looking through past user behavior to create predictive systems. You could be predicting simple things like food people would order or the chance they might cancel a reservation. But, you tend to want to target high-value tasks like predictive maintenance and fraud detection because adding machine learning is expensive and time-consuming.
One reason is that the skills needed for machine learning are different from standard application development. Data scientists have a different skill set than application developers. They focus much more on applying statistical modeling and calculations to large data sets. They tend to use their own languages and toolset, like Python. Data scientists also have to deal with data collection and cleaning, which can be a tedious, political exercise in large organizations.
Meanwhile, application developers tend to have very different skill sets, tools, and even working relationships with the rest of the organization. More than likely, you’re not going to find application developers who can be machine learning experts, nor is it common to have data scientists who can develop applications.
This skills difference means it can be difficult and costly to add machine learning to your applications. It’s not as easy as having an application developer learn how to add new UI elements or incorporate new frameworks. This separation created a bottleneck, limiting how often machine learning (ML)was embedded into enterprise applications.
My theory is that this cost is why we haven’t seen a huge wave of the type of intelligent applications we’re creating with generative AI. Unlike traditional ML, which required specialized teams and complex integration, GenAI arrived with pre-trained models and user-friendly APIs. Businesses could suddenly generate text, code, and even images without hiring specialized teams or adding new infrastructure, staff, and process. The ease of implementation made it more accessible and practical than earlier ML approaches. Where traditional ML struggled to become ubiquitous in enterprise applications, GenAI is rapidly finding a place because it reduced the complexity and cost barriers that had held ML back for years.
Stage 3: Applying Generative AI
And this is where generative AI enters the picture, changing the cost-benefit equation. From a business standpoint, it’s generally cheaper and faster to add AI using generative models than it is with traditional machine learning. That means you can apply AI more broadly across your portfolio, not just to high-value apps that justify the time and money.
Let’s revisit our restaurant reservation example: imagine letting users say, “Book me something like my last Italian spot, but with vegan options.” Generative AI tackles the heavy lifting—interpreting human language, filtering data, and responding—without the specialized data science work that a traditional ML workflow approach would require.
The nature of generative AI also tackles the “people problem.” While application developers might not have data-science skills, generative AI feels far more accessible. As my colleague Adib Saikali puts it, “You don’t need to know anything about linear algebra, or calculus, or any of the math you learned in university. All you need to know is how to call a REST API.”
As a result, you can start layering more advanced intelligence into more of your apps. In larger organizations, your developers likely have many of the tools and know-how to do that already. This “AI broker” approach ensures security, governance, and guardrails, rather than leaving each team to figure that out on their own, and, too often, security teams to figure out how to secure and govern it all after the fact.
Stage 4: Agentic Applications
What’s beyond generative AI? At the moment, the AI community has settled on a concept called “agentic AI.” I think about this as AI that autonomously “does stuff.” While generative AI capabilities are amazing, they’re more “read only,” responding to your prompts with answers, suggestions, and predictions. In contrast, agentic AI doesn’t just answer queries; it acts on them… autonomously.
This is done by introducing a new pattern to generative AI. When you put in a request, an agentic AI application will:
- Break down a request into a plan.
- Execute each step (calling APIs, running microservices, executing code).
- Analyze the outcomes, adjusting if needed.
Think back to the restaurant reservation app. First, the agentic application would reason out that it needs to find your preferences for a restaurant. This might mean asking you directly, but it might also mean looking at your past reservations and reviews you gave. Perhaps you like sitting outside, but is it raining? The intelligent application would look up weather for your area. And so forth. Once this “research” is built up, the agentic AI will then use tools it has to look up available tables at matching restaurants, and if it finds a good match, make the reservation for you, again by making an API call. Of course, a human (that’s you!) can step in at any point to move the workflow along or modify it.
Doing this requires an additional set of services on-top of the base level generative AI toolset. You need to manage the “context” and “memory” of this ongoing process, you need a framework for making those API calls, and so forth.
The good news here, again, is that your application developers don’t need to learn a completely new skill set like machine learning. They can use the existing tools and skills they have to start using agentic AI to add intelligence to their applications.
This agentic AI concept has been evolving over the past year or so, and many frameworks have been developed. We’ve been thinking through the stack of services needed here: the data stores, the execution frameworks, and the basics we already have with Spring AI and AI-brokering in the Tanzu stack. Our advice here is to make sure you have the basics of generative AI apps in place (see above). With that, you can start learning and experimenting with agentic AI apps. For example, you can check out the work we’ve been doing with Anthropic’s Model Context Protocol in Spring AI.
Intelligently escape the fog of opportunity
Hopefully this maturity model highlights that you don’t need to start from scratch—or spend a fortune—to add AI to your existing apps. Many of you can begin right now with the people and tools you already have in place.
There’s one last point to consider. So far, we’ve focused on application developers. Obviously, they’re critical to building apps. But operations teams matter just as much. The good news is that they already know how to run, manage, and secure most parts of these applications, because you’re using the same core building blocks. Sure, there’s nuance and a few new things to learn, but from an ops perspective, AI is largely just another service to oversee. With the right platform, they don’t have to retool everything or break the budget. You can simply keep evolving your apps to become even smarter.