The successful implementation of AI agents requires more than the agent itself.
This should come as no surprise to anyone who has implemented tools in the workplace: any implementation involves much more than the tool itself. Far too many companies, however, still take a βtool firstβ approach to implementation, and the issues arising from this approach will only be heightened when implementing AI tools.
The article One Year of Agentic AI: Six Lessons from the People Doing the Work suggests that many projects are failing - companies are dismantling attempted agent deployments. It claims success requires the redesign of processes, trust, governance, and feedback mechanisms. It sets out a practical analysis of implementing AI:
-
Workflow β Focus on redesigning end-to-end processes, not on building isolated agents. β Map workflows, identify pain points, and integrate agents where they reduce friction. β Example: legal provider logged user edits in contract review to refine the agentβs knowledge base.
-
Only use agents when appropriate β Some tasks (low variance, highly standardised) are better handled by simpler automation or prompting. β Evaluate variance, decision complexity, and error profiles before building agents.
-
Evaluate output to build trust β Users lose confidence quickly if outputs are poor or inconsistent. β Treat agents like new employees: define roles, onboard them, evaluate continuously. β Use frameworks to monitor precision, recall, hallucination rates, and alignment with human judgment.
-
Track and verify every step β End-outcome metrics are insufficient; observability into intermediate steps is essential. β Logging, error detection, and feedback loops allow early diagnosis and refinement. β Example: in document review, monitoring revealed upstream data quality issues driving errors.
-
Reuse components where possible β Avoid duplicating effort by creating modular, reusable components (prompt templates, evaluation logic, retrieval tools). β Shared platforms can cut nonessential work by 30β50%.
-
Humans remain essential β Agents can execute many steps, but humans must validate, handle exceptions, and oversee performance. β Clear workflow design should define handoffs between humans and agents. β Strong UI features (highlighting, click-to-validate) build user trust and adoption.