Enterprises shift from building to operating Artificial Intelligence agents

Enterprises are moving from rapidly building Artificial Intelligence agents to the harder discipline of governing, observing, and securing them in production, as agent-based automation spreads across core business functions.

Enterprises are entering a new phase of generative Artificial Intelligence adoption in which the main challenge is no longer how to build agents but how to operate them safely, predictably, and at scale inside real business systems, according to Maryam Ashoori, vice president of product and engineering at IBM watsonx. She said that since 2023, the market has cycled from early experimentation with models and prompts to a surge of enthusiasm for agents, and is now settling into a more sober focus on control, visibility, and governance. In 2023, most companies treated generative Artificial Intelligence as an exploratory investment concentrated on narrow use cases such as summarization, classification, question answering, information extraction, code generation, and content creation, and when organizations had a clear production-relevant use case, the value was immediate but otherwise pilots produced insight without measurable outcomes.

The landscape shifted in early 2024 when large language models gained the ability to take actions by calling application programming interfaces, a capability often described as agentic Artificial Intelligence that let models interact directly with enterprise software and legacy systems. Ashoori said the enterprise reaction was immediate, with chief information officers requesting agents across the business even when concrete tasks were not fully defined, and speed frequently winning out over structure. By late 2025, she said enterprises had accumulated dozens or even hundreds of agents built by developers, business users, and external providers, often running across different platforms with varying assumptions, and while an agent can be built in less than five minutes, the real problem is what happens once it is deployed and connected to sensitive systems.

As agents multiplied, so did exposure to risk, because agents inherit model limitations and those weaknesses are amplified when systems can act rather than just respond. Ashoori noted that hallucinations, which are familiar at the model layer, can become operational failures at the agent layer, and if the model hallucinates and takes the wrong tool and that tool has access to unauthorized data, then you have a data leak. She said enterprises have therefore shifted their focus from build time to run time as they move from experimentation to production and discover that managing and governing a collection of agents is significantly more complex than creating them, and this has made observability unavoidable.

Ashoori defined observability in terms of tracing, which records every action an agent takes, including which model was involved, which tool was called, the inputs and outputs, how long each step took, and what it cost. She illustrated the need for tracing with a customer support scenario in which an agent diagnosing a non-working camera might consult internal manuals via retrieval, then escalate to online search tools or public forums, and each action is a decision point where incorrect or inappropriate information can surface. With tracing, enterprises can go back and see exactly what happened, support audits and compliance, and identify optimization opportunities by seeing where the latency was caused, where the cost was triggered, and which models or workflows should be replaced. Despite these benefits and growing risks, she cited figures indicating that only about 19% of organizations currently focus on observability and monitoring in production, even as costs and exposure increase.

Looking ahead, Ashoori said that this gap will close quickly, pointing to Gartner estimates suggesting that by 2028, roughly one-third of interactions with generative Artificial Intelligence systems will occur through agents, reinforcing her view that agents are going to be everywhere. However, she stressed that observability is only one part of the challenge, and the next phase of adoption will be shaped heavily by policy enforcement and security. As agents gain autonomy, assigning responsibility becomes more difficult because the builder, model provider, tool owner, and end user may all share accountability when something goes wrong, and she described the security of agents as a very hot topic as enterprises create non-negotiable policies from risk, compliance, and security teams to govern behavior, especially in highly regulated environments.

Fragmentation further complicates governance because agents are created on different platforms and frameworks both inside and outside organizations, which is driving interest in approaches that are agnostic to how agents are built or where they run. From Ashoori’s perspective, best practice involves separating systems that build agents from systems that govern them so that enterprises can monitor, evaluate, and optimize agent behavior regardless of origin. Even with heightened concern around risk, she said momentum behind agents is not slowing but rather becoming more practical, and the next wave of adoption will emphasize domain-specific automation in areas like HR, sales and marketing, procurement, commerce, and IT support, where workflows are repeatable and the value of automation is easier to measure.

Ashoori pointed to IBM’s internal HR agent “AskHR” as an example of this more grounded approach, explaining that it is used to provide complementary analytics during performance evaluations and other sensitive processes involving highly personal information. She expects that successful workloads such as HR agents will increasingly be reused across enterprises because if an agent works well in one large organization, it is likely to be relevant to many others facing similar processes and constraints. In her view, the market is moving away from pure invention toward disciplined execution, and she said the energy has shifted so that the core priority is now learning how to operate Artificial Intelligence agents with confidence and at scale rather than simply launching new pilots.

65

Impact Score

South Korea’s Artificial Intelligence Basic Act takes effect with new obligations for developers and users

South Korea’s Act on the Development of Artificial Intelligence and Establishment of Trust has come into force, introducing transparency, safety, and local-representative duties for a wide range of Artificial Intelligence operators, including those outside the country. The law sets up new oversight bodies and a framework for detailed technical rules that will be finalized by the Ministry of Science and Information and Communication Technology.

How global R&D spending growth has shifted since 2000

Global research and development spending has nearly tripled since 2000, with China and a group of emerging economies driving the fastest growth. Slower but still substantial expansion in mature economies highlights a world that is becoming more research intensive overall.

Finance artificial intelligence compliance in European financial services

The article explains how financial firms can use artificial intelligence tools while meeting European, United Kingdom, Irish and United States regulatory expectations, focusing on risk, transparency and governance. It details the European Union artificial intelligence act, the role of cybersecurity, and the standards and practices that support compliant deployment across the financial sector.

Contact Us

Got questions? Use the form to contact us.

Contact Form

Clicking next sends a verification code to your email. After verifying, you can enter your message.