Millennial AI
Book a Call
Development

Agentic AI workflows explained: what works, what breaks, and what to build first

Tarun SharmaMarch 3, 202618 min read

TL;DR

  • --Agentic AI means multi-step systems that use tools, maintain state, and make decisions autonomously — the architecture matters more than the model powering it.
  • --Single-agent workflows with clear tool boundaries are production-ready today. Multi-agent orchestration remains fragile for most use cases.
  • --Reliability degrades exponentially with step count: 95% per-step reliability across 20 steps yields only 36% end-to-end success.
  • --Start with document processing, code generation, or support triage. Skip anything requiring real-time guarantees or ambiguous success criteria.

What "agentic" actually means in engineering terms

Every AI vendor now calls their product agentic. Chatbots with retrieval-augmented generation get the label. Simple API wrappers that call a model once and return the result get it too. The word has been stretched so thin it barely communicates anything useful to an engineering team making architecture decisions. So let us define it with precision.

An agentic AI system is one that plans a sequence of actions, uses external tools to execute those actions, maintains state across steps, and makes decisions about what to do next without a human directing each move. The key distinction is autonomy across multiple steps. A system that takes a question, searches a knowledge base, and returns an answer is a RAG pipeline. It follows a fixed path every single time. An agent decides what path to take based on what it observes along the way. That capacity for dynamic routing is what earns the label.

Three levels of agentic behavior exist in practice, and they carry fundamentally different engineering requirements. The distinction matters because each level implies a different team size, timeline, and infrastructure commitment. Gartner projected in August 2025 that 40% of enterprise applications will have task-specific AI agents by end of 2026, up from under 5% in 2025. That projection covers all three levels. Most of what ships will be level one and level two. Level three is where the ambition concentrates, and it is also where most of the failures will accumulate.

When someone tells you they are building an agentic system, ask which level they mean. The answer determines whether you are talking about a focused project or a platform initiative.

LevelPatternEngineering RequirementsTimeline
1 — Simple chainFixed sequence, no branchingError handling, one engineerWeeks
2 — Single agentDynamic tool selection, state across stepsState management, tool integration, eval metrics1–2 months
3 — Multi-agentSpecialized agents coordinating on shared tasksInter-agent protocols, shared memory, platform teamQuarter+

The architecture under the hood

Every agentic system, regardless of framework or model provider, shares four core components. Understanding these components helps you evaluate vendor claims, debug production failures, and make grounded architecture decisions.

The first component is the planning and reasoning loop. This is the model deciding what to do next based on the current state and the original goal. In a single-agent system, the model receives a prompt that includes the goal, the actions taken so far, and the outputs of those actions. It generates the next action. This loop repeats until the model determines the goal is met or it hits a termination condition like a step limit or a timeout. The quality of this reasoning loop depends on the model's ability to maintain coherent plans across many iterations. Most models start to drift after eight to twelve steps on complex tasks, losing track of what they were trying to accomplish or repeating actions they already completed.

The second component is tool use. Tools are the agent's interface with the external world: APIs, databases, file systems, web browsers, code execution environments, calculators, search engines, other models. Each tool has a defined interface specifying what inputs it accepts and what outputs it returns. The agent selects tools based on the current sub-task, formats the input correctly, calls the tool, and parses the output. Tool integration is where a surprising amount of engineering time goes. APIs return unexpected formats. Rate limits get hit during peak usage. Authentication tokens expire mid-workflow. Schema changes break parsing logic. Every tool represents a potential failure point, and the agent needs explicit handling for each failure mode.

The third component is memory and state. Short-term memory covers the current task: what actions have been taken, what intermediate results exist, what remains to be done. Long-term memory persists across tasks: user preferences, past interactions, learned patterns about recurring inputs. Short-term memory is typically implemented as a growing context window or a structured state object that gets updated after each step. Long-term memory usually involves a vector database or a key-value store. The challenge is deciding what to remember and what to discard. An agent that retains everything runs into context length limits and pays escalating inference costs. An agent that forgets too aggressively loses context that would have prevented errors in later steps.

The fourth component is observation and evaluation. After each tool call, the agent needs to parse the output and assess whether it moves closer to the goal. Did the API return valid data or an error? Does the extracted text match the expected format? Is the generated code syntactically correct? This observation step is where agents frequently go wrong. A model that cannot reliably evaluate its own intermediate outputs will propagate errors through every subsequent step, compounding small inaccuracies into large failures.

In a single-agent architecture, one model instance orchestrates all four components. In a multi-agent architecture, each agent handles its own planning, tool use, and observation, while a separate orchestration layer manages coordination. That orchestration layer decides which agent runs next, what information passes between agents, and how conflicts or contradictions get resolved. The Agent-to-Agent (A2A) protocol from Google and the Model Context Protocol (MCP) from Anthropic are emerging standards for this inter-agent communication. A2A focuses on how agents discover each other and exchange messages. MCP standardizes how agents discover and interact with tools. Both are early in adoption, and neither has become a universal default.

The orchestration layer is where most of the complexity in multi-agent systems concentrates. Getting individual agents to perform their specialized tasks is tractable engineering. Getting them to coordinate without duplicating work, contradicting each other, or falling into infinite delegation loops requires careful design, extensive testing, and production monitoring that goes well beyond what most teams budget for initially.

Where agentic workflows deliver today

Google Cloud's AI Agent Trends report for 2026 found that 88% of early adopters are seeing positive ROI from agentic AI deployments. That number needs context. Early adopters are companies that selected appropriate use cases, invested in engineering infrastructure, and measured outcomes carefully. But the signal is clear: specific categories of agentic workflow are producing measurable value in production right now.

Document processing pipelines are the most mature category. An agent reads an incoming document, classifies its type, extracts structured fields, validates those fields against business rules, and routes the result for human review or automated processing. Insurance claims processing illustrates the pattern well. The agent receives a claim submission, extracts claimant information along with injury details and dollar amounts, checks those against the relevant policy terms, flags inconsistencies or coverage questions, and routes straightforward claims for automated approval while sending edge cases to a human adjuster. Each step involves a different tool: OCR or document parsing for extraction, a rules engine for validation, a database query for policy lookup. The agent ties these tools together and handles the branching logic that previously lived in a twenty-page business process document nobody wanted to maintain. Companies running these pipelines report 60-70% reductions in manual processing time for standard cases, with human effort concentrated on the exceptions that genuinely require judgment.

Code generation and review is the second production-ready category, and it has progressed faster than most observers expected two years ago. The workflow spans multiple steps: understand a requirement from a ticket or specification, generate code, run the test suite, interpret failures, fix the code, and submit a pull request or merge request. Tools like Claude Code, Cursor, and GitHub Copilot Workspace implement variations of this pattern at different levels of autonomy. The agent architecture works well for code because code has clear success criteria built in. Tests either pass or they fail. The code either compiles or it does not. Type checks either succeed or they surface errors. This clarity makes the observation and evaluation component straightforward compared to tasks where success is subjective or ambiguous.

Research and analysis workflows form the third category. An agent gathers information from multiple sources, cross-references findings, synthesizes patterns, and produces a structured report. Due diligence for M&A transactions, competitive landscape analysis across market segments, and regulatory change monitoring are all use cases where agents perform well in production. The task is naturally decomposable into steps: search for relevant information, extract key facts, organize them into a coherent structure, identify gaps in coverage, search again to fill those gaps. A human analyst doing the same work spends the large majority of their time on the gathering and organizing steps and a smaller fraction on the judgment calls that require domain expertise. The agent handles the high-volume gathering work and presents the judgment calls to the human in a structured format with supporting evidence attached.

Customer support triage rounds out the production-ready categories. The agent reads an incoming ticket, classifies intent and urgency, searches relevant knowledge base articles and historical ticket resolutions, drafts a response, and either sends it directly for high-confidence cases or escalates to a human agent for anything uncertain. The key design decision is where to set the confidence threshold for autonomous responses. Too high and the agent escalates everything, providing no efficiency gain. Too low and customers receive incorrect or unhelpful responses that erode trust. Companies that tune this threshold carefully, starting conservative and lowering it gradually as they accumulate data about the agent's accuracy, see strong results within the first quarter of deployment.

Across all four categories, a consistent pattern emerges. The agentic workflows that succeed in production have clear inputs with known formats, well-defined tools with predictable interfaces, measurable success criteria that the agent can evaluate programmatically, and a human review step for edge cases and low-confidence outputs. When evaluating whether a potential use case is a good candidate for agentic automation, check it against those four properties. If any of them is missing, the project will struggle.

Where they break

The compound reliability problem is the single most important concept in agentic AI engineering, and it remains underappreciated by almost everyone building these systems for the first time.

Assume each step in an agentic workflow succeeds 95% of the time. That sounds excellent for any individual step. But reliability compounds multiplicatively across steps, and the results are worse than most teams expect.

StepsPer-step reliability 95%Per-step reliability 99%
577%95%
1060%90%
2036%82%
508%61%

A demo runs the happy path once and looks impressive. Production runs every path thousands of times per day, and the long tail of failures becomes the dominant engineering challenge.

Benchmark data confirms the severity. The best agents available today achieve less than 55% goal completion on CRM tasks that involve navigating interfaces, filling out forms, and coordinating actions across multiple systems. These are tasks that a trained human employee handles correctly 98% of the time. The gap is enormous, and it closes slowly because each improvement to the underlying model only moves the per-step success rate by a few percentage points. As the table shows, even small per-step improvements compound meaningfully, but twenty-step workflows remain fragile at any realistic per-step reliability.

Real-time constraints create a second category of failure. Agents are inherently slow. Each step requires a model inference call, a tool invocation, and an observation parse. A single step takes two to ten seconds depending on the model size and the tool's response time. A ten-step workflow takes thirty seconds to two minutes. For a research task running in the background while a user does other work, that latency is acceptable. For a customer waiting on a chat response or a transaction that needs to complete within a session timeout, it is a dealbreaker. Companies that attempt to put agentic workflows in the critical path of user-facing interactions discover this constraint quickly and painfully.

Ambiguous success criteria cause a third failure mode. When an agent generates a research report, how does it determine the report is thorough enough? When it drafts a marketing email, how does it evaluate whether the tone matches the brand? Without criteria that the model can assess programmatically, agents either over-refine by running additional steps that add marginal value while burning through tokens, or they terminate too early and deliver mediocre output. Defining success criteria that are specific, measurable, and evaluable by the model itself is a design problem that most teams underestimate by a wide margin.

Error cascading is the fourth and most insidious failure mode. One agent produces slightly incorrect output. The next step or agent uses that output as input and introduces additional inaccuracy. By the third or fourth step, the accumulated errors make the final output unreliable or outright wrong. In a human workflow, each person applies independent judgment and often catches errors introduced by previous contributors. Current agents lack that independent judgment about the quality of their inputs. They process what they receive with high confidence regardless of whether the input is accurate.

Non-determinism rounds out the major failure modes. Run the same agent with identical input twice, and you will frequently get different outputs. For creative tasks like writing or brainstorming, variability is a feature that users appreciate. For compliance reviews, financial calculations, or medical record processing, it is a serious liability. Lower temperature settings and structured output formats reduce non-determinism, but they cannot eliminate it entirely with current model architectures.

[Gartner's prediction that over 40% of agentic AI projects will be canceled by end of 2027](https://www.gartner.com/en/newsroom/press-releases/2025-06-agentic-ai-projects) maps directly to these failure modes. Projects that select use cases with too many sequential steps, latency requirements that agents cannot meet, success criteria that resist quantification, or zero tolerance for output variation are the projects heading for cancellation. Understanding these constraints before you start building is the difference between joining the 60% that delivers value and the 40% that does not.

Framework landscape: an honest assessment

Six frameworks dominate the agentic AI development space as of early 2026, and choosing between them is one of the first decisions engineering teams face. Here is a candid evaluation of each, including the tradeoffs and limitations that vendor marketing tends to omit.

LangGraph is the most mature option for stateful workflow orchestration. Built on top of LangChain, it models agent workflows as directed graphs where nodes represent processing steps and edges define transitions between them. Checkpointing is built in, which means you can pause a workflow at any node, inspect the full state, persist it to storage, and resume later. Human-in-the-loop approval gates are straightforward to implement as special nodes that block execution until a human provides input. The tradeoff is complexity. LangGraph's learning curve is steep, and even simple three-step agents require more boilerplate code than feels justified. The graph abstraction is powerful for complex workflows with conditional branching, retry logic, and parallel execution paths. For a linear five-step pipeline, it adds overhead without proportional benefit.

CrewAI takes the opposite approach, optimizing for speed from concept to working prototype. You define agents by role, giving each a description of its expertise and the tools it can access. Assemble agents into crews, define the tasks they should accomplish, and the framework handles coordination between them. You can go from zero to a functioning multi-agent demo in an afternoon. The cost of that speed is reduced visibility into execution flow. When agent interactions produce unexpected results, debugging is harder because the coordination logic is managed by the framework rather than defined explicitly in your code. Teams that prototype in CrewAI and then need production-grade control often end up migrating to a more explicit framework.

Microsoft Agent Framework replaced AutoGen in late 2025 and targets enterprise deployments within the Azure ecosystem. Identity management, compliance logging, deployment infrastructure, and monitoring all connect natively to existing Azure services. If your organization already runs on Azure and your security team requires AI workloads to stay within the Microsoft perimeter, this framework eliminates substantial integration effort. For teams outside the Microsoft ecosystem, the Azure coupling becomes a constraint that limits flexibility. The framework's abstractions are enterprise-oriented, which translates to more configuration and more ceremony around deployment.

Claude Agent SDK from Anthropic is built around the Model Context Protocol (MCP), which provides a standardized way for agents to discover and interact with tools. Tool definitions follow a shared schema, which means tools built for Claude Agent SDK can potentially be reused across any framework that adopts MCP. The SDK is opinionated toward Claude models, and the integration patterns prioritize clean, minimal code for single-agent workflows with well-defined tool boundaries. The MCP foundation positions it well for interoperability as the protocol gains adoption. The tradeoff is that multi-agent orchestration patterns require more custom implementation work compared to frameworks that provide orchestration primitives out of the box.

OpenAI Agents SDK provides a streamlined abstraction for building agents on OpenAI models. The core concepts are an agent loop that handles tool use, a handoff mechanism that lets you route specific sub-tasks to specialized agents, and a guardrails system for output validation. The emphasis is on incremental adoption: start with a single agent using one tool and gradually add complexity as your understanding of the problem grows. Documentation is thorough and the developer experience is polished. The tradeoff is model dependency. The SDK is designed for OpenAI models and does not provide abstractions for working across model providers.

The honest assessment across all six frameworks: none of them are mature in the way that web application frameworks or database drivers are mature. All are evolving rapidly. APIs change between minor versions. Documentation has gaps that only surface when you hit edge cases in production. Behaviors that are not covered in getting-started guides reveal themselves under real workloads.

FrameworkBest ForLearning CurveMaturityEcosystem Lock-in
LangGraphComplex stateful workflowsSteepMost maturePython
CrewAIRapid prototypingLowEarlyPython
Microsoft Agent FrameworkAzure enterprisesMediumGrowingAzure/Microsoft
Claude Agent SDKMCP-native tool useLowEarlyClaude models
OpenAI Agents SDKIncremental agent featuresLowEarlyOpenAI models

Pick the framework that aligns with your existing model provider, your team's primary programming language, and your cloud infrastructure. Do not reorganize your stack to accommodate a framework that might look completely different in twelve months. Budget engineering time for framework upgrades and breaking changes, because they will arrive more frequently than you would prefer. And remember that the framework matters less than the quality of your agent design. Clear tool boundaries, proper state management, robust error handling, and comprehensive evaluation will determine your outcomes more than framework choice.

Do you even need an agent?

The most valuable decision in an agentic AI project is sometimes the decision to build something simpler.

Use a fixed pipeline when the steps are sequential and predictable, when no branching logic depends on intermediate results, and when the output format is consistent across inputs. Extract data from a PDF, map it to a schema, validate the fields, write the record to a database. Every document follows the same path. A pipeline is easier to test because the execution path is deterministic. It is easier to debug because there is no reasoning loop to inspect. It is easier to monitor because you know exactly which step failed. And it is cheaper to run because you are not paying for model inference at a planning step that adds no value when the plan never changes. Many workflows that companies describe as agentic are actually pipelines with a language model handling one or two steps. That is perfectly fine engineering. Call it what it is and build it accordingly.

Use a single agent when the task requires dynamic tool selection based on what the agent discovers during execution. A single agent processing invoices might need to handle four different invoice formats, each requiring different extraction logic and different validation rules. The agent examines the document, determines the format, selects the appropriate extraction tool, validates the output against format-specific rules, and handles cases where extraction fails by trying alternative approaches. The branching decisions and error recovery logic justify the agent architecture because you cannot predict the execution path at design time. Keep the tool count under fifteen and the maximum step count under ten, and a single agent will perform reliably.

Use multi-agent orchestration when the task has genuinely distinct sub-tasks that benefit from specialization, when those sub-tasks can execute in parallel to reduce total processing time, and when your engineering team has the capacity and willingness to debug distributed systems. A due diligence workflow might involve a financial analysis agent reviewing balance sheets and cash flow statements, a legal document review agent examining contracts and regulatory filings, and a market research agent analyzing competitive dynamics and industry trends. Each agent has different tools and different domain knowledge encoded in its prompts. Running them in parallel reduces wall-clock time compared to a single agent handling everything sequentially. But coordinating their outputs, resolving contradictions between their findings, and producing a coherent final report adds significant engineering complexity that you need to staff and budget for.

The Deloitte Tech Trends 2026 report found that fewer than one in four organizations have scaled agents to production. The organizations that succeeded almost all started with single-agent workflows on narrowly defined tasks and expanded scope only after they had validated reliability on the initial use case.

Fixed sequential steps?Use a pipeline
Dynamic tool selection needed?Single agent
Genuinely distinct parallel sub-tasks?Multi-agent (if you can debug distributed systems)

A practical rule of thumb: if you can draw the entire workflow on a whiteboard with fewer than ten boxes and the branching logic fits on a single page, a single agent is sufficient. If the workflow requires multiple whiteboards to capture the full scope, multi-agent orchestration may be justified. If you need zero whiteboards because the flow is completely linear, build a pipeline and save the agent architecture for a problem that actually requires it. Most mid-market use cases fit squarely in the single-agent category. Multi-agent orchestration is overkill for 80% of what companies are trying to build today.

Governance: auditing systems that make decisions

An agent that processes invoices is making decisions about your money. An agent that triages support tickets is making decisions about your customer relationships. An agent that reviews contracts is making decisions about your legal exposure. The moment an agent acts on real business data with any degree of autonomy, governance stops being optional and becomes a core architectural requirement.

Trace logging is the foundation of agent governance. Every decision the agent makes, every tool it calls, every output it receives, and every evaluation it performs at each step must be recorded in a structured, immutable log. This is the equivalent of an audit trail for a human employee, except it needs to capture substantially more detail because the agent's reasoning process is less transparent than a human's decision-making. When something goes wrong, the trace log is how you reconstruct what happened, identify the root cause, and determine how to prevent recurrence. Build the trace logging infrastructure before you build the agent itself. Teams that add logging after the fact invariably discover that the most important events were not being captured during the period when the agent was least understood and most likely to produce unexpected behavior.

Human-in-the-loop gates define which decisions require human approval before the agent can proceed. Financial transactions above a certain dollar threshold, customer-facing communications that will be sent without further review, data deletion requests, access permission changes. The specific thresholds depend on your organization's risk tolerance and regulatory environment. The design challenge with approval gates is keeping them efficient. If every gate requires a human to open a dashboard, read three screens of context, and click an approve button, you have eliminated the efficiency gain that justified building the agent in the first place. Good gate design presents the human with exactly the information they need to make the approval decision, pre-formatted and pre-analyzed by the agent, so the review takes seconds rather than minutes.

Sandboxing limits the blast radius when an agent encounters an input it was not designed for. Agents should operate with the minimum permissions necessary for their specific task. A document processing agent needs read access to the document store and write access to the output database. It does not need access to production APIs, the employee directory, the financial ledger, or administrative systems. Principle of least privilege has been a foundational security concept for decades, and it applies with particular force when the entity making access decisions is a probabilistic model that can behave unpredictably on edge-case inputs. Every additional permission you grant an agent is an additional category of damage it can cause when it encounters something outside its training distribution.

Cost controls prevent an expensive failure mode that catches teams off guard. Agents that hit parsing edge cases or encounter malformed inputs can enter retry loops, calling APIs and consuming model tokens hundreds of times while attempting to process something that falls outside their capabilities. A document processing agent that encounters a corrupted PDF might attempt to parse it, fail, modify its approach, fail again, and repeat this cycle until it exhausts a token budget or hits a hard timeout. Without per-task and per-hour spending limits, a single problematic input can generate hundreds of dollars in API costs within minutes. Set hard budget caps at the task level and the hourly level. Monitor consumption patterns and treat any task that hits a spending limit as a signal that the agent's error handling needs improvement for that category of input.

Governance requires the same engineering rigor as building the agent itself. Organizations that treat it as a compliance checkbox to address once before launch, rather than a continuously maintained architectural layer, consistently find themselves unable to keep controls aligned as the agent's capabilities and scope evolve over time.

What I recommend building right now

If you are a CTO or engineering leader at a mid-market company evaluating agentic AI, here is a concrete path forward based on patterns I have seen work across multiple engagements and technology stacks.

Start with your most document-heavy process. Every company has one. Invoice processing, claims handling, contract review, compliance documentation, employee onboarding paperwork, vendor qualification packets. Pick the process where humans currently spend the most time on repetitive extraction, classification, and routing tasks. Build a single-agent workflow that handles the standard cases and routes exceptions to humans for judgment. Do not attempt to automate the exceptions in your first deployment. The split where the agent handles routine work and humans handle the cases requiring judgment is the right target for a first production system. That split delivers measurable time savings from day one while keeping risk contained.

Build observability from day one, before you write the first line of agent logic. Set up structured logging that captures every tool call, every model inference, every decision point, and every output evaluation. Create a dashboard that shows agent behavior in real time during development and in aggregate during production. You need to see what the agent is doing at every step before you can trust it with anything that touches real business data. Teams that skip observability during development and retrofit it later always discover behaviors they did not anticipate. Those discoveries tend to come from customer complaints or financial discrepancies rather than from proactive internal monitoring.

Use a framework that matches your existing stack. If your team writes TypeScript, do not switch to Python to use LangGraph. If you are already running workloads on Azure, evaluate Microsoft Agent Framework before exploring alternatives. Framework familiarity reduces debugging time, and debugging time will be your largest engineering cost in the first three months of agent development. The framework that lets your team move fastest with confidence is the right choice, regardless of which one generates the most discussion on social media.

Set a 90-day timeline with three distinct phases. Weeks one and two: build the agent for a narrowly scoped task with a small set of tools and clearly defined success criteria. Define what a correct output looks like for fifty representative inputs before you write any agent code. Weeks three through six: run the agent on real production data with a human reviewing every single output. Track accuracy rates, catalog failure modes, and refine prompts and tool integrations based on what you observe in the review data. Weeks seven through twelve: gradually increase autonomy. Let the agent handle high-confidence cases without human review while maintaining human oversight for edge cases and outputs where the agent's confidence score falls below your threshold. Expand the scope of inputs the agent processes as your measured accuracy supports it.

The goal of this first deployment is not full automation. The goal is a system where the agent handles the predictable, high-volume portion of the work and a human handles the portion that requires judgment, contextual understanding, or relationship awareness. That split produces measurable productivity gains without requiring the agent to be perfect. It also generates the operational data you need to make an informed decision about whether to expand the agent's responsibilities, maintain the current scope, or redirect the investment.

The engineering discipline that matters most

Agentic AI is systems engineering. The model is a component. An important component, but a component that operates within an architecture you design, build, test, and maintain. The orchestration layer, error handling, observability infrastructure, and governance controls are where the majority of production engineering effort concentrates. A team that spends 80% of its time refining prompts and 20% on systems engineering has the ratio inverted.

The Deloitte Tech Trends 2026 report found that fewer than one in four organizations have scaled agents to production. The gap between a proof-of-concept demo and a production deployment is where most projects stall or get canceled. That gap is filled with engineering work that bears no resemblance to building a demo: writing integration tests that simulate tool failures and verify the agent degrades gracefully, setting up alerting for anomalous agent behavior patterns, building rollback mechanisms for when an agent produces a batch of bad outputs, creating incident response runbooks for when the agent does something unexpected at two in the morning.

Companies that treat agent development like a data science experiment will hit the Gartner 40% cancellation wall. The pattern repeats itself across organizations. A team builds a proof of concept in a notebook or a playground environment. They demonstrate it to leadership. The demo covers the happy path and looks genuinely impressive. The team gets funding and a mandate to productionize. They discover that the demo handled 60% of real-world input variation and the other 40% triggers failure modes nobody anticipated. Months get spent adding error handling, edge case logic, and monitoring. The project either receives enough sustained engineering investment to cross the production threshold or it gets canceled when the timeline and budget blow past the original estimates.

Companies that treat agent development like production software engineering will be in the 60% that delivers measurable value. That means automated testing with input distributions that reflect production traffic, including the ugly edge cases and malformed data. It means monitoring and alerting that catches accuracy degradation before users report problems. It means gradual rollouts with kill switches that can pull the agent out of the workflow within minutes if something goes wrong. It means incident response processes that treat agent failures with the same urgency as application outages. It means measuring outcomes against a documented baseline so you have evidence, quantified evidence, that the agent produces better results than the process it replaced.

The technology is ready for production use in well-scoped domains with clear success criteria. The engineering practices needed to deploy agents reliably are well understood by anyone who has built and operated distributed systems. The gap is organizational: making the commitment to sustained systems engineering work that transforms a promising demo into a reliable production system that runs without constant supervision. That commitment is the factor that separates the projects that get canceled from the projects that deliver compounding value over years.

Tarun Sharma

Tarun Sharma

Partner, Engineering

IIT Kanpur, Jaguar Land Rover, a published paper in Elsevier, and his own company (Twinity Labs) building digital twins. Deepest technologist on the team. He decides what gets built and how.

LinkedIn
Try our free AI assessment