Curated by THEOUTPOST
On Sat, 30 Nov, 12:01 AM UTC
2 Sources
[1]
Safety and reliability concerns temper agentic AI mania - SiliconANGLE
When information services and technology firm Thomson Reuters Corp. acquired accounting software firm Materia in October, it paved the way for what it sees as the future of its business. Materia, which is the business name of Credere Technologies Inc., builds software agents that can break down complex queries and consult numerous information sources to deliver a compound answer. Such automation could be a godsend for the accounting profession, which is facing a shortage of up to 3.5 million accountants by 2025. One reason often cited is the volume of tedious information-gathering the job involves. Materia is tackling the problem with a new type of generative artificial intelligence that dispatches autonomous agents to identify the most important sources of information, look up the required material and deliver a consolidated summary. Unlike generative AI, which sorts through large data corpora to produce outputs that mimic human creativity but do not independently set or pursue goals, agentic AI acts autonomously with decision-making capabilities, often guided by goals or objectives. Agentic systems can plan, reason and execute tasks across multiple steps, adapting to changing environments and contexts. Thomson Reuters sees agents as a perfect match for its business of providing actionable information to business professionals. "You'd think a question about the accounting treatment of sales and revenue is straightforward, but you need to consult standards information from a source like Thomson Reuters' Checkpoint products, internal accounting policies and various online resources," said David Wong, Thomson Reuters' chief product officer. "You need to pull the information relevant to that particular question, reconcile it and produce a response to the user." Materia deconstructs such problems into individual steps that an agent can tackle. "The approach that they took was to think of every problem as being essentially a multistep one, even if the steps were really, really simple to begin with," Wong said. Autonomous agents are potentially far more powerful than the current breed of information-gathering generative models. In theory, agents can work independently, tackling complex tasks with little need for supervision. OpenAI Chief Executive Sam Altman has called them "AI's killer function." The online dictionary Whatis.com defines agentic AI as "systems capable of autonomous action and decision-making [that] can pursue goals independently, without direct human intervention." Agents have generated enormous enthusiasm recently because of their potential to pull information from multiple sources and use that intelligence to automate processes and deliver finished plans or products. Large language models have solved the information retrieval problem. The next step is to apply automation to the results. However, agentic AI has also raised questions about accountability, reliability and safety. At its worst, a generative AI model dispenses misleading or incorrect information. Agentic models can do far more damage if given too much decision-making latitude, potentially including financial damage or threats to human safety. That's why most experts believe AI agents will be limited to noncritical business decisions in clearly defined domains for the foreseeable future. "Autonomy is negatively correlated with the stake of the decision," said James Wall, product manager at causal AI software firm causaLens, the business name of Impulse Innovations Ltd. "Much of the talk about agentic AI today actually refers to rudimentary automations that are far from agentic systems that can dramatically transform workflows," said David Vellante, chief analyst at theCUBE Research. "Enterprise leaders should view agentic AI the same way they should view quantum computing," said Kjell Carlsson, head of AI strategy at Domino Data Labs Inc. "Like quantum, agentic AI has the potential to be very powerful and disruptive, but there are too many fundamental challenges on a technological and governance level that need to be solved before leaders should spend any time thinking about it." Those risks have done little to quell a gold rush as software vendors scramble to add agentic capabilities to their products. Gartner Inc.'s 2024 Emerging Technology Hype Cycle has autonomous agents climbing the expectation curve just behind artificial general intelligence. The research firm predicts agents will be embedded in one-third of enterprise software applications by 2028. A recent survey of 100 information technology executives by Forum Ventures LLC found that 48% are beginning to adopt AI agents, and 33% are actively exploring them. "When you look at doing things much more efficiently, agentic AI is the way," said Bhaskar Roy, chief of AI products and solutions at Workato Inc., which recently added agentic capabilities to its data integration platform. In October, Salesforce Inc. launched Agentforce, a set of tools that enables customers to build and customize agents that can access company data and take actions on behalf of employees in areas like sales, service, marketing and commerce. Moveworks Inc., Cisco Systems Inc., Thoughtful Automation Inc. and UiPath Inc. have recently joined the agentic parade. And at Amazon Web Services Inc.'s re:Invent conference starting Monday, AI agents are expected to get top billing on multiple keynotes. Just this past week, a team of prominent developers from Google, Meta Platforms Inc., Stripe Inc. and Dropbox Inc. said they've raised more than $50 million to build an operating system for agents. "Just as Android made mobile development accessible to virtually any developer, we're building the platform that will help make AI agents mainstream," said David Singleton, co-founder and chief executive of a startup called /dev/agents. Startup Thoughtful Automation Inc. has raised more than $38 million for its AI agents that handle claims processing, patient eligibility verification and payment posting for healthcare companies, claiming its customers see between five and nine times return on investment. Construction management software firm Procore Technologies Inc. just rolled out a line of agents that are said to enforce project efficiency, improve safety, enhance decision-making and streamline workflows. Human resources software provider UKG Inc. just introduced agents that can monitor regulatory changes and alert HR professionals. A July report by CB Information Services Inc. estimated that more than 50 companies are building agents, agentic workflows and agent infrastructure. The number is no doubt much larger now. The concept of autonomous agents is nothing new. "This type of architecture has been around for several decades; object-oriented software used many of these principles," said Bern Elliot, vice president and distinguished analyst at Gartner. "Think of it as a subroutine dressed up in a tuxedo." Krishna Tammana, chief technology officer at conversational messaging platform GupShup, the business name of Webaroo Inc., sees many potential everyday uses for agents. They include monitoring data from wearables and medical records to identify potential health risks and recommend interventions, intervening to prevent fraud in financial transactions, automating routine customer service interactions and conducting initial job interviews. "Instead of simply retrieving search results, agentic AI can actively guide the user, offering product options and explaining each recommendation," he said. Software development is primed for an agentic revolution, said Kevin Cochrane, CMO of Vultr, the business name of cloud computing provider The Constant Company LLC. "These models will not only generate functional code but also automate testing steps that adhere to organizational coding standards," he said. "Autonomous coding agents can potentially manage entire coding cycles for tasks such as troubleshooting bugs or optimizing legacy systems." One well-known example of agentic AI is autonomous vehicles. They integrate input from various sources and continually make decisions that adapt to changing road and traffic conditions. Their decision-making domain is well-defined, and humans can always override their choices. However, autonomous vehicles also exemplify some of the ethical and moral dilemmas of assigning too much control to a machine. Despite excellent safety records, occasional mistakes have inhibited broader adoption. Agentic AI will encounter many of the same objections. LLMs have rekindled interest because of their striking success in delivering sophisticated responses to questions posed in everyday language. Generative AI has removed the need for users to master sophisticated programming languages to create agentic interactions. "You don't have to be explicit with an LLM," Elliot said. "It concludes and decides what information is needed for the task." That potentially opens up huge new use cases. The distinction between agentic and generative AI is fuzzy enough that the term is being applied loosely. Materia, for example, barely mentions agents on its website, calling itself instead the "generative AI platform for intelligent accounting." Thomson Reuters' Wong acknowledged that agentic capabilities are more of a roadmap than a current deliverable. "Eventually, we want these agents to take action," he said. An LLM that culls responses from a defined set of sources and delivers a recommendation can appear to be acting independently but is actually only following instructions. "My experience is that the agent label will be slapped on anything," said Matt McLarty, chief technology officer at integration specialist Boomi Inc. Domino Data Labs' Carlsson advises IT leaders to look out for what he called "agentic AI washing," or exaggerated claims that products appear more innovative, advanced or competitive than they are. "Since practical agentic AI is infeasible for enterprises, vendors will increasingly use these terms as fancier-sounding ways to refer to the real and important capabilities necessary for orchestrating AI workflows, which is called MLOps," he said. An example of how fine a line exists between programmatic and agentic behavior is Reserve with Google, an assistant Google LLC introduced in 2017 that can make restaurant, event and other reservations using online reservation forms or dialing the establishment and placing the order using a synthesized voice. That's an automation, but not an agent, said Nenshad Bardoliwalla, Google Cloud's director of product management for Vertex AI and cloud AI industry solutions. "If you asked it to book a restaurant reservation but didn't specify the restaurant, that would be agentic behavior," he said. The question that will loom over agents for a long time is how much autonomy to give them. The current crop mostly works within a defined set of applications, as with Salesforce's Agentforce. That allows interactions to be strictly defined and carefully monitored. "We're not at the autonomous auto stage; we're more at cruise control," said Harry Wang, vice president of growth and new ventures software quality firm Sonar, the business name of SonarSource SA. "You need people to have their hands on the steering wheel." Workato's agentic orchestration platform, introduced in August, narrowly defines agents' domains to competencies it calls skills and allows agents only to take actions with permissions granted by human users. "Use cases need to be deterministic; you don't want an agent to go rogue, so autonomy needs to be controlled," Roy said. Agentforce exemplifies what agents will likely look like for at least the next few years. They can be programmed to access functions within the portfolio of Salesforce applications and selected partners but are limited in the actions they can take and how they interact with the outside world. That's another major challenge of making agentic AI more broadly useful. Internet services are built with various technologies, protocols and application programming interfaces with little standardization. Public APIs may have inconsistent documentation, rate limits or updates that can break integrations. Services use diverse data formats that complicate communication. Scalability, predictable performance and security are other major concerns. "To make agents work in concert, several pieces have to fall into place, not the least of which is the ability to harmonize disparate data types and formats," said theCUBE Research's Vellante. "This is nontrivial and something often glossed over." In a perfect world, a person would ask an agent to book a trip to a resort destination on certain days, with certain activities and within a stated budget and receive a full proposed itinerary in return. With one click, the agent could make all the necessary reservations. But it's one thing to instruct an agent to book a window seat on a particular flight on a specific day and another to let it choose the airline and reservation on its own. "The question is the degree of agency and the authority you wish to delegate," said Gartner's Elliot. "If they can autonomously do things, then you have agents without accountability because they don't understand the context." The travel booking example would involve the agent conversing with multiple airlines and hotels, figuring out rates and discounts and achieving the optimal balance of adventure and relaxation. That could include interacting with dozens of other travel agents. Though most travel services expose APIs, the interoperability standards that permit full autonomy don't exist yet. "There is a technical challenge to translating what the models know into speaking to the APIs that glue the action systems to the model so they can talk to each other," said Google's Bardoliwalla. "You need the model to think and respond in a step-by-step manner that is reasonable and makes sense. Not only do they have to understand API signatures, but they have to adhere to a very specific format." CB Insights' report said reservations about agents' ability to execute complex tasks across multiple services will limit broad enterprise adoption, at least in the short term. "Despite the rising momentum, agents remain limited in their ability to execute tasks reliably across the internet and software apps," the report said. Google, Sabre Corp., Alaska Air Group Inc., InterContinental Hotels Group PLC, Sonderbase Technologies PLT and Mindtrip Inc. are just a few of the established and emerging companies building AI-powered travel planners. Still, none allows its bots to book reservations on its own. "The future of agentic AI depends on balancing its potential with responsible deployment, guided by ongoing research and evolving regulations," said Unmesh Kulkarni, head of generative AI at data science firm Tredence Inc. Even in scenarios that lend themselves to automation, such as customer service, placing too much trust in AI can backfire. Customer service interactions often involve emotion and conflict, necessitating a human touch. "Agents don't have empathy," said Workato's Roy. Bala Kumar, chief product and technology officer at identity management firm Jumio Corp., sees darker possibilities. Unregulated activity on e-commerce and dating platforms risks "fundamentally disrupting user engagement and sales," he wrote in an e-mail message to SiliconANGLE. "E-commerce bots will wreak havoc on online shopping, particularly during high-demand events like concert ticket sales, creating a significant threat to consumer trust and market integrity." Improving model transparency could be a big step toward increasing trust in agents, but that goal has proved elusive in generative AI because of the nature of the complex and nonlinear interactions between potentially billions of parameters. Agentic AI is even more complicated. "I don't think the industry talks enough about the black box problem," said Daniel Avancini, chief data officer at data services company Indicium Tech Corp. "If you don't have well-defined processes, there are a lot of ethical risks, such as hallucinations and unpredicted outcomes. Answers have to be well-defined." CausaLens' Wall believes familiarity will breed acceptance. Although bad actors will undoubtedly co-opt agents in nefarious new ways, the vast majority will help knowledge workers be more productive. He recalls the introduction of spreadsheets to the workplace. "People were skeptical about the 'mean' function; was it really taking an average?" he said. "Spreadsheets started in a low-trust environment, and we've slowly come to rely on them. Automation will be as good as people's ability to understand it." Vellante sees agents improving over time to become ubiquitous in the workplace. "Agents will observe and learn from human reasoning," he said. "The vast amount of work that is non-automated today will begin to be streamlined in ways that will drive dramatic improvements in productivity."
[2]
AI agents are practical. Reliability is another matter. - SiliconANGLE
Building artificial intelligence agents that can interact with each other reliably across services presents technical challenges that have never been tackled before. According to one AI researcher, that will limit their use to a narrow set of business processes for the next few years. Niloufar Salehi (pictured below), an assistant professor at the University of California at Berkeley, said traditional data processing systems are built to deliver predictable results. But nothing is predictable about machine learning. The same algorithm may produce entirely different results depending on context. Building trustworthy agentic AI systems involves solving problems that have never before existed. "The way databases work is that you have a data schema and you know what sort of things are there, like accounts and deals," she said. "What agents are doing is much more unstructured and adapts over time. It's a completely different way of thinking about data structure and schemas, and we don't have the right way to build that yet." The mathematical term is stochastic, a reference to systems or processes that are inherently random or involve uncertainty. One well-known example of a stochastic algorithm is the Facebook newsfeed, which varies based on the preferences of the person reading it. "The moment systems become stochastic, you no longer open your Salesforce application and see the same thing you saw yesterday," she said. "That means reliability is a huge issue." It's also an extremely difficult problem because many variables are involved. In addition to the billions of parameters used to train machine learning models, agentic AI involves interactions among multiple agents, each operating on stochastic algorithms. Solving the problem requires creating persistent, shared memory that agents can tap to learn from past actions, Salehi said. Current AI models approach each situation as if it's entirely new, even if it has encountered the same scenario repeatedly in the past. "If an agent has a problem, it tries multiple ways to solve it," she said. "If it comes across the same problem tomorrow, it goes through that process all over again. Building out the shared memory that lets these agents coordinate is an extremely difficult feat of engineering." Researchers are working to extend the context length or maximum amount of short-term memory a model can employ to make actions more predictable and repeatable. Salehi pointed to MemGPT, a research project attempting to build a memory manager for large language models, as a potential solution. "Giving agents a longer-term memory they can share will be a big unlock," she said. Salehi became interested in human-computer interaction shortly after receiving her Ph.D. in computer science from Stanford University in 2018. Working with a team of physicians using machine translation in emergency rooms to communicate with non-English-speaking patients, she was surprised to discover that error rates were alarmingly high. "Chinese translation had a 20% error rate for commonly used sentences, such as emergency room discharge instructions," she said. "It was 8% for errors that caused potentially significant harm, such as telling someone to continue kidney medication they were supposed to stop. That's what got me thinking about how to make agentic systems more reliable." Salehi recently joined a new venture, Across.ai, to build agentic memory for enterprise workflows, starting with sales operations. Its agents, called Sparkles, "do one task well using a combination of techniques designed to achieve higher reliability," she said. For example, Its Pain Point Tracker Sparkle culls data from sources like social media and customer support calls to identify customers' most common difficulties and track how they change over time. Insight Sparkles present data and recommendations. Action Sparkles can automatically generate emails, presentations, proposals and other sales-related materials. A human must approve actions. She is optimistic that agents that can coordinate across services and companies will be practical within the next five years, but they probably won't be used in the ways people expect. The reasons have little to do with technology. For example, training agents to schedule meetings is difficult because people's time is so personal and many of us keep mental notes about managing our schedules. "It's hard to automate when human factors aren't known or not written anywhere," she said. Better candidates for automation are well-known processes that could benefit from better coordination. "I think we're pretty close to making those systems happen," she said. "The next five to 10 years will be all about solving extremely difficult engineering problems to get agents to work together and reliably."
Share
Share
Copy Link
Agentic AI is gaining traction in enterprise software, promising autonomous decision-making capabilities. However, safety, reliability, and technical challenges temper the enthusiasm, limiting its current applications to non-critical business processes.
Agentic AI, a new breed of artificial intelligence capable of autonomous action and decision-making, is gaining significant attention in the enterprise software landscape. Unlike traditional generative AI, agentic AI can independently set and pursue goals, plan, reason, and execute tasks across multiple steps, adapting to changing environments 1. This technology is seen as a potential game-changer for businesses, with Gartner predicting that agents will be embedded in one-third of enterprise software applications by 2028 1.
Companies are already beginning to integrate agentic AI into their products and services. Thomson Reuters, for instance, acquired Materia to leverage its agentic AI capabilities in breaking down complex queries and consulting numerous information sources to deliver compound answers 1. This approach could be particularly beneficial in fields like accounting, where there's a projected shortage of professionals.
Other major players entering the agentic AI space include:
Despite the enthusiasm, agentic AI faces significant technical challenges that limit its current applications. Niloufar Salehi, an assistant professor at UC Berkeley, highlights the unpredictable nature of machine learning algorithms as a major hurdle 2. Unlike traditional data processing systems, agentic AI systems are stochastic, meaning their outputs can vary based on context and other factors.
Key challenges include:
While agentic AI holds promise for automating complex tasks, it also raises important questions about accountability and safety. Unlike generative AI, which may produce misleading information, agentic models have the potential to cause more significant damage if given too much decision-making latitude 1. This concern has led experts to suggest that AI agents will likely be limited to non-critical business decisions in clearly defined domains for the foreseeable future 1.
Despite the challenges, researchers and companies are working on solutions. Projects like MemGPT aim to extend the context length of AI models, potentially improving their predictability and repeatability 2. Salehi is optimistic that agents capable of coordinating across services and companies will be practical within the next five years, though their applications may differ from current expectations 2.
As the field progresses, the focus will be on solving the engineering problems necessary to make agents work together reliably. While the potential of agentic AI is significant, its widespread adoption in critical business processes will likely require overcoming substantial technological and governance challenges in the coming years 12.
Reference
AI agents are gaining widespread adoption across industries, but their definition and implementation face challenges. Companies are rapidly deploying AI agents while grappling with issues of autonomy, integration, and enterprise readiness.
5 Sources
5 Sources
AI agents are emerging as the next frontier in artificial intelligence, promising to revolutionize how businesses operate and how technology is developed and utilized. This story explores the current state of AI agents, their potential impact, and the challenges that lie ahead.
4 Sources
4 Sources
AI agents are emerging as autonomous systems capable of handling complex tasks across various industries, from customer service to software development. While promising increased efficiency, their deployment raises questions about job displacement, privacy, and trustworthiness.
8 Sources
8 Sources
AI agents are emerging as a powerful force in business automation, combining the capabilities of large language models with autonomous decision-making to revolutionize workflows across industries.
7 Sources
7 Sources
AI agents are emerging as the next wave of AI technology, offering autonomous task completion and decision-making capabilities beyond traditional chatbots and large language models.
7 Sources
7 Sources
The Outpost is a comprehensive collection of curated artificial intelligence software tools that cater to the needs of small business owners, bloggers, artists, musicians, entrepreneurs, marketers, writers, and researchers.
© 2025 TheOutpost.AI All rights reserved