top of page
InnerMined-Pitch-Deck.pptx.png

Exclusive

Agentic AI Frameworks Are Failing 86% of the Time. The Smartest Enterprises Know Exactly Why.

  • Writer: Nivedita Chandra
    Nivedita Chandra
  • 9 hours ago
  • 9 min read

The most advanced agentic AI coding tool available today resolved only 14% of real-world GitHub issues without human help (Deloitte, 2025). Sit with that number for a moment. This is not a mid-market product or an early prototype. This is one of the most capable autonomous agents ever benchmarked in a production context, and it needed a human to handle 86% of the complexity it encountered. That is not a reason to slow down AI adoption. It is a precise description of where human judgment is still the competitive edge, and it is the most clarifying number you can bring into your next strategy conversation.


Agentic AI frameworks are the software infrastructure layers that enable AI agents to plan multi-step tasks, use external tools, maintain memory across interactions, and coordinate with other agents to achieve a goal. They are the operating layer between a large language model and a real-world enterprise workflow. Understanding what they are, and which type of agent each workflow actually needs, is the decision that determines whether your AI investment produces results or liability. By the end of this post, you will know what these frameworks are, the four types of agents they power, and how to position yourself as the human layer that makes the whole system work.


Agentic AI Frameworks

What Are Agentic AI Frameworks (and Why the Choice Matters)

A large language model on its own is a powerful thinking tool. It cannot, on its own, remember what it did three steps ago, connect to an external system, hand work off to another agent, or recover when something goes wrong mid-task. Agentic AI frameworks add those capabilities on top: memory, the ability to use tools, step-by-step planning, and coordination between multiple agents working together. The framework is what turns a model into a system that can take action.


Several frameworks are now ready for serious enterprise use. LangGraph, from LangChain, structures agent workflows as a series of connected steps that can loop back and retry when needed, making it well-suited to tasks that require multiple rounds of reasoning. Microsoft's AutoGen lets multiple specialised agents hold a structured conversation, check each other's work, and collaborate toward a shared goal. CrewAI assigns each agent a specific role and set of responsibilities, working like a small team where every member has a defined job. Amazon Bedrock Agents offers a cloud-managed option for organisations already running on AWS, with ready-made connections to enterprise data and built-in infrastructure for running agents at scale.


Each framework is designed for a different way of working, and the right choice depends entirely on which type of agent your workflow requires. That makes understanding the five agent types the first decision, not the last.


Understanding the 5 Main Types of AI Agents

According to IBM, there are five main types of AI agents: simple reflex agents, model-based reflex agents, goal-based agents, utility-based agents, and learning agents (IBM, 2024). Each type differs in how much it can perceive, how it makes decisions, and how much human oversight it needs to operate safely in an enterprise environment. All five can be combined within a single multi-agent system, with each handling the part of the task it is best equipped for.


Simple Reflex Agents are the most straightforward type. They look at what is happening right now and respond according to a fixed set of rules, with no memory of what came before and no ability to plan ahead. A thermostat is the classic example: if the temperature drops below a threshold, it turns on the heat. In a business context, a simple reflex agent might route incoming support tickets to the right team based on keywords in the subject line. These agents are reliable and easy to audit, but they break down the moment a situation falls outside their predefined rules. Human oversight is needed at setup and at regular intervals to make sure the rules still reflect how the business actually works.


Model-Based Reflex Agents work the same way as simple reflex agents, except they also maintain a basic internal picture of the world around them. This allows them to handle situations where not all the relevant information is immediately visible. A customer service agent that tracks what a user said earlier in the same conversation, and uses that context to interpret an ambiguous follow-up question, is a model-based reflex agent. The rules still drive the responses, but the agent is no longer blind to everything that happened before the current moment. Human oversight remains relatively low, but someone needs to monitor whether the internal model is staying accurate as conditions change.


Goal-Based Agents go a step further. Rather than simply reacting to inputs, they are given a specific target and choose their actions based on what is most likely to get them there. They can evaluate different possible paths and select the one that best serves the goal, which makes them far more flexible than reflex agents. An agent tasked with scheduling a complex product launch across multiple teams, choosing the sequence of actions that avoids conflicts and meets the deadline, is operating at this level. The human role here shifts: rather than maintaining a set of rules, leaders need to define the goal clearly and verify that the agent's chosen path actually makes sense in the real-world context it is operating in.


Utility-Based Agents operate similarly to goal-based agents, but with an added layer of judgment. Instead of simply reaching a goal, they evaluate how good different outcomes are and choose the path most likely to produce the best result overall. This matters in situations where there are trade-offs. A fraud detection system that assigns a risk score across a range of possibilities, rather than applying a simple pass-or-fail threshold, is working like a utility-based agent. So is a procurement tool that balances cost, delivery time, and supplier reliability when recommending a vendor. These agents require careful human oversight of what "best" actually means in practice, because their output quality depends entirely on how well the success criteria have been defined.


Learning Agents are the most capable and the most unpredictable. They improve their own performance over time by storing what they have done before, observing what worked, and adjusting their behaviour accordingly. They do not just follow instructions or pursue a fixed goal: they get better at the task the more they do it. A recommendation engine that gradually refines its suggestions based on user behaviour, or a predictive maintenance system that becomes more accurate as it processes more equipment data, are both examples of learning agents in enterprise use. The human role here is significant and ongoing. Leaders must define what the agent is learning from, check regularly that its behaviour is improving in the right direction, and intervene when it starts optimising for the wrong thing.


Most enterprise deployments in 2025 combine elements of several agent types within a single system. Misidentifying which type a given workflow requires, most commonly by applying a learning or autonomous agent to a process that has not been clearly defined, is one of the most common and costly mistakes in enterprise AI rollouts.


The Human in the Loop: The Deployment Layer Most AI Strategies Skip

Most enterprise AI frameworks are evaluated on capability: what the system can do, how fast it runs, and how easily it connects to existing tools. Far fewer are evaluated on a more important question: where does human judgment need to stay in this workflow, and what happens when it is removed? The framework selection conversation almost always comes before the human oversight conversation. That order is wrong, and the consequences are not theoretical.


The Deloitte Technology, Media and Telecom Predictions 2025 report puts a number to the gap. In a benchmarking test, the agentic coding tool Devin resolved approximately 14% of real-world GitHub issues on its own, roughly twice the rate of standard AI chatbots, but still well short of operating without human support (Deloitte, 2025). The 14% is the headline. The 86% is where the strategy lives.


The tasks that advanced agents cannot resolve on their own are not spread evenly across a workflow. They pile up around unclear instructions, unusual situations, and decisions that require judgment based on context the agent does not have. These are exactly the situations where an experienced technical leader is most valuable. When deciding where human oversight is non-negotiable, start with the workflows where the requirements are most ambiguous, not the ones with the highest volume. High volume is where agents perform well. Ambiguity is where they need you.


Agentic AI frameworks also produce poor results when they are dropped into messy, poorly structured processes. The agent will execute the inefficiency faster, not fix it. The most valuable thing a technical leader can do before any deployment is step back and clean up the workflow first: cut out steps that only exist for historical reasons, clarify what a correct output actually looks like, and define exactly where a human decision is required rather than a mechanical one. This is not a technical task. It is a leadership decision that determines whether the AI produces return on investment or just faster noise.


The third argument follows from the first two. Some versions of this conversation treat human oversight as a temporary measure, something that will become unnecessary as AI improves. That framing is wrong and costly to believe. Technical leaders who know how to deploy agentic AI frameworks, set the right constraints, and read the outputs critically are the people their organisations will be least able to replace. The orchestrator role is not a stepping stone. It is where the real work of enterprise AI happens.


How to Deploy Different Types of Agents in AI for Maximum ROI

Every agent deployment must start with a clear, specific goal, written down before any framework is chosen or any vendor is contacted. An agent cannot be measured against a vague objective, and a vague objective almost always means the underlying workflow has not been thought through carefully enough.


Once the goal is clear, match the agent type to how much the workflow can tolerate a wrong answer. Simple and model-based reflex agents belong in processes where errors are expensive and the correct response is already well defined. Utility-based and goal-based agents suit situations where the task involves trade-offs and optimising across options is more important than following a fixed script. Learning agents and full multi-agent systems should only be introduced where human review points are already designed into the process, not added after something goes wrong.


Redesign the workflow before deploying the agent. Running both at the same time makes it impossible to tell whether a poor result came from a gap in the process or a limitation of the agent, which makes both harder to fix.


Define where humans need to be involved before the system goes live. Every deployment needs clear, agreed answers to three questions: which decisions need a human sign-off before the agent moves forward, what triggers an escalation, and who is responsible when the agent gets it wrong. These answers belong in the design document, not in the incident report.


The Critical Role of AI Governance in Autonomous Systems

Governance is not a brake on AI capability. It is the structure that makes it safe to run autonomous systems at scale without the risk quietly building up until something serious goes wrong. Skipping governance at the start is how organisations end up with agents that do exactly what they were told, in a situation where the instructions were not good enough.


The first governance requirement is a clear record of what the agent did and why. Every decision an agent makes should be logged in a way that a person can read and trace back to the original instruction. If a regulator, a board member, or a legal team asks why the system made a particular call, the answer needs to exist already. If it does not, the organisation cannot show it was in control, regardless of how well the technology performed.

The second requirement is limiting what each agent can access. An agent should only be able to reach the data and systems it genuinely needs for its specific task. Giving agents broader access than necessary is one of the most common sources of data risk in early deployments. It usually happens because it is faster at setup, and it becomes a serious problem when agent behaviour moves in an unexpected direction.


The third requirement is knowing in advance which decisions need a human to approve before the agent acts. These boundaries should be written into how the agent operates, not left as informal guidance that gets overlooked under time pressure. Common examples include any action above a certain financial value, changes to data the agent did not create, and any output going to an external party without prior review.


Mature agentic AI frameworks offer varying levels of built-in governance support. LangGraph allows teams to inspect what an agent is doing at each stage of a workflow. AutoGen supports human review interrupts that pause agent activity at defined points. CrewAI lets teams build approval steps directly into task sequences. When evaluating any agentic AI framework, governance capability should carry the same weight as performance. A system that works well but cannot be audited or controlled is not ready for enterprise use.


The Work That Is Already Yours

The gap between what agentic AI frameworks can do and what they can do reliably without human oversight is not closing fast enough to wait out. It is exactly where experienced technical leaders are most needed right now.


Before the next AI budget conversation in your organisation, do one specific thing: identify the three workflows where human judgment is currently the only thing preventing a bad outcome. Those are your first deployment opportunities, because they are where choosing the right agent type matters most, where governance design will be tested first, and where your role as the person holding it all together is already understood. The frameworks are ready. The question is whether the human layer around them is being built with the same care as the technology itself.

Comments


bottom of page