From IEEE Spectrum, November 19:
Here's how AI agents work, why people are jazzed about them, and what risks they hold
The artificial intelligence world is abuzz with talk of AI agents. Microsoft recently released a set of autonomous agents that could help streamline customer service, sales, and supply chain tasks. Similarly, OpenAI unveiled Swarm, an experimental framework to explore better coordination between multi-agent systems. Meanwhile, Claude, the large language model (LLM) from Anthropic, is taking agentic AI to the next level with the beta stage of its computer use skills—from moving a mouse cursor around the screen to clicking buttons and typing text using a virtual keyboard.
So, what exactly are AI agents?
“AI agents are advanced artificial intelligence systems that are able to complete a task or make a decision,” says Adnan Ijaz, director of product management for Amazon Q Developer, an AI-powered software development assistant from Amazon Web Services (AWS). “Humans set the goal, and agents figure out on their own, autonomously, the best course of action.” The agents can interface with external systems to take action in the world.
In addition to this autonomy, agentic AI can also receive feedback and continually improve on a task, says Yoon Kim, an assistant professor at MIT’s Computer Science and Artificial Intelligence Laboratory.
Think of AI agents as a more capable version of generative AI. While both technologies rely on LLMs as their underlying model, generative AI creates new content based on the patterns it learned from its training data. Agentic systems, on the other hand, are not only able to generate content but are also able to take action based on the information they gain from their environment. “So all of that is essentially a step further than generative AI,” Ijaz says.
How AI Agents Work
To fulfill a particular task, AI agents usually follow a three-part workflow. First, they determine the goal through a user-specified prompt. Next, they figure out how to approach that objective by breaking it down into smaller, simpler subtasks and collecting the needed data. Finally, they execute tasks using what’s contained in their knowledge base plus the data they’ve amassed, making use of any functions they can call or tools they have at their disposal.
Let’s take booking flights as an example, and imagine a prompt to “book the cheapest flight from A to B on Y date.” An AI agent might first search the web for all flights from A to B on Y date, scan the search results, and select the lowest-priced flight. The agent then calls a function that connects to the application programming interface (API) of the airline’s flight booking platform. The agent makes a booking for the chosen flight, entering the user’s details based on the information stored in its knowledge base.
“The key point of agentic interaction is that the system is able to understand the goal you’re trying to accomplish and then operate on it autonomously,” says Ijaz. However, humans are still in the loop, guiding the process and intervening when required. For instance, the flight-booking AI agent might be instructed to notify the user if the cheapest flight has no available seats, allowing the user to decide on the next step. “If at any point humans don’t think the system is going in the right direction, they can override it—they have control,” Ijaz adds.Promises and Pitfalls of Agentic AI....
....MUCH MORE