
Jul 25, 2025
Agent Evaluation Research Exposes Critical Gaps in Production AI Systems


Conor Bronsdon
Head of Developer Awareness
Conor Bronsdon
Head of Developer Awareness


Modern LLM-based agents can draft contracts, triage customer tickets, and file code reviews, yet you rarely know whether those actions remain safe once the model goes live.
This first comprehensive survey of LLM-agent evaluation synthesized insights from more than 100 benchmarks and frameworks into 4 clear dimensions. If you've tried pushing agents beyond proof-of-concept, you know evaluation—not modeling—creates the real bottleneck.
Traditional ML metrics assume deterministic outputs and miss the erratic reasoning chains, tool calls, and emergent behaviors that define autonomous agents. The numbers tell the story: leading agents register success rates as low as 2% on the hardest tasks.
The researchers identify critical gaps—safety, cost-efficiency, and fine-grained diagnostics—setting the stage for the framework and challenges explored ahead.
Explore the Research Paper: Survey on Evaluation of LLM-based Agents
We recently explored this topic on our Chain of Thought podcast, where industry experts shared practical insights and real-world implementation strategies

Summary: Four Critical Dimensions of Agent Evaluation
The researchers neatly organize every agent testbed into four dimensions:
Fundamental capabilities
Application-specific tasks
Generalist reasoning
Evaluation frameworks themselves
This gives you a clear mental model for what was previously a scattered landscape.
The researchers' findings reveal some uncomfortable truths. On several difficult benchmarks, agents barely achieve 2% success—a stark reality captured in the survey's data visualizations. Common evaluation problems include limited automation and test planning challenges, while safety coverage and cost metrics receive less attention in existing surveys.
The good news? You can see the field evolving toward more robust, production-ready evaluation: live web environments replace static datasets, "Agent-as-a-Judge" scoring appears in advanced evaluation suites, and synthetic data generation creates endless stress tests.
These advances transform the survey into a practical roadmap for building safer, more cost-effective, and reliable agent systems in real-world applications.

The Five Challenges of Agent Evaluation
You might think your agent evaluation stack is complete after reviewing dozens of benchmarks. This research survey proves otherwise. By cataloging more than a hundred tests, this research exposes five blind spots that consistently derail production deployments.
Each gap affects your ability to trust, scale, and budget for autonomous systems. Fixing them means rethinking what you measure, how often, and under which real-world conditions.
Challenge #1: Safety and Compliance Evaluation Gaps
Modern agents can draft contracts, handle customer tickets, and review code, but do you know if they remain safe once deployed? The AIES meta-analysis of generative-AI safety work highlights three missing layers: modality coverage, risk coverage, and contextual assessment.
The research explains how most studies focus only on text models, a narrow set of ethical risks, and isolated outputs rather than actual user interactions. This leaves multimodal agents, privacy issues, and real user scenarios largely untested.
These gaps create real business risks for your organization. Without proper audit trails for toxic language, biased recommendations, or policy violations, you face potential regulatory fines and reputation damage. Safety becomes even more crucial as your agents gain autonomy—a single hallucination can cascade through an entire tool chain.
To protect your systems, you'll need multi-dimensional tests that reflect actual usage patterns. Leading teams combine automated toxicity filters with human reviews and log every decision for audits.
Modern frameworks built on accuracy, relevance, safety, and consistency principles already include such protections, flagging unsafe steps immediately. By extending these checks to image, audio, and code, you minimize potential harms while demonstrating compliance.
Challenge #2: Missing Cost-Efficiency and Resource Metrics
That impressive benchmark score looks less amazing when you see your first API bill. Most public leaderboards celebrate raw accuracy but ignore token counts, latency, or engineering costs. Few evaluations track inference cost or throughput, yet these factors often determine whether a model makes it to production.
Without cost data, you risk building agents that shine in tests but bankrupt you at scale. High token usage can cripple your real-time chat applications. Oversized context windows slow critical operations like fraud detection.
Avoiding these financial surprises requires tracking resources in every experiment you run. You should treat cost as equally important as quality and safety, recording compute, memory, and response time per scenario.
Many successful teams use "cost-normalized" scores, dividing accuracy by dollars spent, making trade-offs crystal clear to your decision-makers.
Modern frameworks handle this accounting automatically, showing token consumption alongside success rates. By standardizing these measures, you can test different temperatures, model sizes, or retrieval depths and immediately see the financial impact on your budget.
Challenge #3: Inadequate Fine-Grained Analysis
Pass/fail metrics look neat on paper but help little when your agent fails mid-conversation. The survey criticizes these broad evaluations for hiding where and why breakdowns happen. Eugene Yan illustrates this problem: end-to-end scores lump tool selection, reasoning, and execution into a single number, making it impossible for you to identify the weak link.
Debugging becomes mere guesswork without detailed traces—you end up changing prompts, temperature, or memory settings until something works. Fine-grained analysis solves this by scoring each step:
Did the agent pick the right API?
Was its reasoning logical?
Did it execute the final action correctly?
Advanced tools visualize these paths and highlight failure points in real time for your team. Framework extensions include step-by-step scoring, enabling automated reviews across thousands of runs.
Implementing this approach starts with recording every observation, thought, and action token—challenging but essential for your diagnostic process. Once captured, you can link failures to specific prompt patterns or tool uses and prioritize the most impactful fixes. Over time, these detailed metrics become early warning signs in your monitoring.
Drops in reasoning quality often precede complete task failure, giving you time to address issues before users notice.
Challenge #4: Scalability and Automation Limitations
Manual evaluations work for ten prompts but collapse when your system handles millions daily. The SIGIR LLM4Eval workshop points out that static, human-labeled tests quickly become outdated as models learn to exploit patterns in the data. Re-labeling every cycle costs too much of your time and money, undermining fast release schedules.
Automation helps, but implementing simple approaches creates new problems for your team. LLM-as-a-judge techniques scale subjective scoring but may inherit the very biases you're trying to detect.
Best practices suggest hybrid methods:
Combine automated initial grading with targeted human reviews of edge cases
Refine judge prompts to improve alignment with your quality standards
Synthetic data generation further reduces your annotation work. Advanced frameworks create scenario variations that test corner cases without exhausting reviewer capacity. By running these tests continuously—after every code change or knowledge update—you shift from occasional certification to constant quality monitoring across your deployment.
The key is balancing speed with reliability in your evaluation strategy. Automated systems must show confidence scores and uncertainty levels so you know when human review is needed. Done right, scalable evaluation becomes an always-on safety net that grows with your system rather than falling behind.
Challenge #5: Lack of Realistic, Dynamic Evaluation Environments
Many agents excel at toy problems but stumble when facing real-world use cases. The survey notes that models scoring above 80% on synthetic tests can drop to single-digit success rates when facing open-ended web environments. Static benchmarks freeze a moment in time; real websites, APIs, and your users' needs change constantly.
Leading security teams advocate for "live" benchmarks that pull fresh data or make actual API calls, though adoption remains limited. Until these resources mature, you can approximate reality by embedding evaluation in your staging systems.
Let your agent navigate a cloned customer portal, submit a test support ticket, or refactor actual GitHub issues. Each run captures rich data—response delays, unexpected page changes, authentication errors—that simple text matching misses.
Dynamic environments also reveal interaction nuances your team needs to understand: how the agent handles rate limits, retries, or partial failures. These edge cases rarely appear in curated datasets yet dominate production incidents. By prioritizing realistic simulation, you align offline metrics with real-world performance expectations.
Some evaluation frameworks are beginning to explore multi-dimensional approaches with ongoing, adaptive testing to keep measurements relevant as your systems evolve, though standardized frameworks with integrated replay environments don't yet exist.
The benefit is clear: agents that survive dynamic evaluation deliver more consistent performance, fewer surprises, and better user experiences in your production environment.
Practical Takeaways
You can apply the survey's insights to your daily engineering practice without waiting for new tools:
Many teams discover evaluation gaps only after deploying agents to production. Start by comparing your current tests against the paper's four-dimensional structure
While simple pass/fail metrics provide basic information, your production agents need deeper analysis. Add detailed, step-by-step checks—like reasoning review—to your existing evaluation system.
Trace-based tools show how granular assessment catches problems that binary scoring misses.
Safety and compliance need top priority in your evaluation pipeline, not afterthought status. The gaps in modality and context in the survey show what happens when safety evaluation takes a back seat.
Cost factors directly impact whether deployment makes sense for your business, regardless of accuracy. Your evaluation should track token usage, response time, and API expenses alongside performance to avoid models that cost more than they save.
Static evaluations quickly become obsolete as your agent capabilities grow. Implement live, continuously updated evaluations that adapt to changing needs and new challenges.
For critical applications, pair automated scoring with regular human checks. This combined approach remains the most reliable way to catch subtle failures that automated systems might miss.
Final Thoughts
The survey exposes a harsh truth: evaluation science has exploded with 100+ benchmarks testing everything from reasoning to real-world tool use, yet LLM agents still perform poorly, sometimes scoring just 2% on challenging tasks.
Safety reviews remain limited to text analysis, leaving multimodal risks largely unexamined. Cost tracking and detailed diagnostics struggle to keep pace with deployment speed.
Moving forward means treating thorough, multi-dimensional evaluation as both an ethical obligation and a business necessity for teams committed to deploying trustworthy agents at scale.
The survey's four-dimensional framework makes one thing clear: you need evaluation tools that cover fundamental skills, domain objectives, agent behavior, and daily workflow health. Relying on scattered scripts or static tests creates blind spots in safety, cost, and diagnostics highlighted throughout current research.
Explore how Galileo integrates these insights into a single, production-ready platform, so you can monitor every dimension without juggling multiple tools or manual spreadsheets.
Modern LLM-based agents can draft contracts, triage customer tickets, and file code reviews, yet you rarely know whether those actions remain safe once the model goes live.
This first comprehensive survey of LLM-agent evaluation synthesized insights from more than 100 benchmarks and frameworks into 4 clear dimensions. If you've tried pushing agents beyond proof-of-concept, you know evaluation—not modeling—creates the real bottleneck.
Traditional ML metrics assume deterministic outputs and miss the erratic reasoning chains, tool calls, and emergent behaviors that define autonomous agents. The numbers tell the story: leading agents register success rates as low as 2% on the hardest tasks.
The researchers identify critical gaps—safety, cost-efficiency, and fine-grained diagnostics—setting the stage for the framework and challenges explored ahead.
Explore the Research Paper: Survey on Evaluation of LLM-based Agents
We recently explored this topic on our Chain of Thought podcast, where industry experts shared practical insights and real-world implementation strategies

Summary: Four Critical Dimensions of Agent Evaluation
The researchers neatly organize every agent testbed into four dimensions:
Fundamental capabilities
Application-specific tasks
Generalist reasoning
Evaluation frameworks themselves
This gives you a clear mental model for what was previously a scattered landscape.
The researchers' findings reveal some uncomfortable truths. On several difficult benchmarks, agents barely achieve 2% success—a stark reality captured in the survey's data visualizations. Common evaluation problems include limited automation and test planning challenges, while safety coverage and cost metrics receive less attention in existing surveys.
The good news? You can see the field evolving toward more robust, production-ready evaluation: live web environments replace static datasets, "Agent-as-a-Judge" scoring appears in advanced evaluation suites, and synthetic data generation creates endless stress tests.
These advances transform the survey into a practical roadmap for building safer, more cost-effective, and reliable agent systems in real-world applications.

The Five Challenges of Agent Evaluation
You might think your agent evaluation stack is complete after reviewing dozens of benchmarks. This research survey proves otherwise. By cataloging more than a hundred tests, this research exposes five blind spots that consistently derail production deployments.
Each gap affects your ability to trust, scale, and budget for autonomous systems. Fixing them means rethinking what you measure, how often, and under which real-world conditions.
Challenge #1: Safety and Compliance Evaluation Gaps
Modern agents can draft contracts, handle customer tickets, and review code, but do you know if they remain safe once deployed? The AIES meta-analysis of generative-AI safety work highlights three missing layers: modality coverage, risk coverage, and contextual assessment.
The research explains how most studies focus only on text models, a narrow set of ethical risks, and isolated outputs rather than actual user interactions. This leaves multimodal agents, privacy issues, and real user scenarios largely untested.
These gaps create real business risks for your organization. Without proper audit trails for toxic language, biased recommendations, or policy violations, you face potential regulatory fines and reputation damage. Safety becomes even more crucial as your agents gain autonomy—a single hallucination can cascade through an entire tool chain.
To protect your systems, you'll need multi-dimensional tests that reflect actual usage patterns. Leading teams combine automated toxicity filters with human reviews and log every decision for audits.
Modern frameworks built on accuracy, relevance, safety, and consistency principles already include such protections, flagging unsafe steps immediately. By extending these checks to image, audio, and code, you minimize potential harms while demonstrating compliance.
Challenge #2: Missing Cost-Efficiency and Resource Metrics
That impressive benchmark score looks less amazing when you see your first API bill. Most public leaderboards celebrate raw accuracy but ignore token counts, latency, or engineering costs. Few evaluations track inference cost or throughput, yet these factors often determine whether a model makes it to production.
Without cost data, you risk building agents that shine in tests but bankrupt you at scale. High token usage can cripple your real-time chat applications. Oversized context windows slow critical operations like fraud detection.
Avoiding these financial surprises requires tracking resources in every experiment you run. You should treat cost as equally important as quality and safety, recording compute, memory, and response time per scenario.
Many successful teams use "cost-normalized" scores, dividing accuracy by dollars spent, making trade-offs crystal clear to your decision-makers.
Modern frameworks handle this accounting automatically, showing token consumption alongside success rates. By standardizing these measures, you can test different temperatures, model sizes, or retrieval depths and immediately see the financial impact on your budget.
Challenge #3: Inadequate Fine-Grained Analysis
Pass/fail metrics look neat on paper but help little when your agent fails mid-conversation. The survey criticizes these broad evaluations for hiding where and why breakdowns happen. Eugene Yan illustrates this problem: end-to-end scores lump tool selection, reasoning, and execution into a single number, making it impossible for you to identify the weak link.
Debugging becomes mere guesswork without detailed traces—you end up changing prompts, temperature, or memory settings until something works. Fine-grained analysis solves this by scoring each step:
Did the agent pick the right API?
Was its reasoning logical?
Did it execute the final action correctly?
Advanced tools visualize these paths and highlight failure points in real time for your team. Framework extensions include step-by-step scoring, enabling automated reviews across thousands of runs.
Implementing this approach starts with recording every observation, thought, and action token—challenging but essential for your diagnostic process. Once captured, you can link failures to specific prompt patterns or tool uses and prioritize the most impactful fixes. Over time, these detailed metrics become early warning signs in your monitoring.
Drops in reasoning quality often precede complete task failure, giving you time to address issues before users notice.
Challenge #4: Scalability and Automation Limitations
Manual evaluations work for ten prompts but collapse when your system handles millions daily. The SIGIR LLM4Eval workshop points out that static, human-labeled tests quickly become outdated as models learn to exploit patterns in the data. Re-labeling every cycle costs too much of your time and money, undermining fast release schedules.
Automation helps, but implementing simple approaches creates new problems for your team. LLM-as-a-judge techniques scale subjective scoring but may inherit the very biases you're trying to detect.
Best practices suggest hybrid methods:
Combine automated initial grading with targeted human reviews of edge cases
Refine judge prompts to improve alignment with your quality standards
Synthetic data generation further reduces your annotation work. Advanced frameworks create scenario variations that test corner cases without exhausting reviewer capacity. By running these tests continuously—after every code change or knowledge update—you shift from occasional certification to constant quality monitoring across your deployment.
The key is balancing speed with reliability in your evaluation strategy. Automated systems must show confidence scores and uncertainty levels so you know when human review is needed. Done right, scalable evaluation becomes an always-on safety net that grows with your system rather than falling behind.
Challenge #5: Lack of Realistic, Dynamic Evaluation Environments
Many agents excel at toy problems but stumble when facing real-world use cases. The survey notes that models scoring above 80% on synthetic tests can drop to single-digit success rates when facing open-ended web environments. Static benchmarks freeze a moment in time; real websites, APIs, and your users' needs change constantly.
Leading security teams advocate for "live" benchmarks that pull fresh data or make actual API calls, though adoption remains limited. Until these resources mature, you can approximate reality by embedding evaluation in your staging systems.
Let your agent navigate a cloned customer portal, submit a test support ticket, or refactor actual GitHub issues. Each run captures rich data—response delays, unexpected page changes, authentication errors—that simple text matching misses.
Dynamic environments also reveal interaction nuances your team needs to understand: how the agent handles rate limits, retries, or partial failures. These edge cases rarely appear in curated datasets yet dominate production incidents. By prioritizing realistic simulation, you align offline metrics with real-world performance expectations.
Some evaluation frameworks are beginning to explore multi-dimensional approaches with ongoing, adaptive testing to keep measurements relevant as your systems evolve, though standardized frameworks with integrated replay environments don't yet exist.
The benefit is clear: agents that survive dynamic evaluation deliver more consistent performance, fewer surprises, and better user experiences in your production environment.
Practical Takeaways
You can apply the survey's insights to your daily engineering practice without waiting for new tools:
Many teams discover evaluation gaps only after deploying agents to production. Start by comparing your current tests against the paper's four-dimensional structure
While simple pass/fail metrics provide basic information, your production agents need deeper analysis. Add detailed, step-by-step checks—like reasoning review—to your existing evaluation system.
Trace-based tools show how granular assessment catches problems that binary scoring misses.
Safety and compliance need top priority in your evaluation pipeline, not afterthought status. The gaps in modality and context in the survey show what happens when safety evaluation takes a back seat.
Cost factors directly impact whether deployment makes sense for your business, regardless of accuracy. Your evaluation should track token usage, response time, and API expenses alongside performance to avoid models that cost more than they save.
Static evaluations quickly become obsolete as your agent capabilities grow. Implement live, continuously updated evaluations that adapt to changing needs and new challenges.
For critical applications, pair automated scoring with regular human checks. This combined approach remains the most reliable way to catch subtle failures that automated systems might miss.
Final Thoughts
The survey exposes a harsh truth: evaluation science has exploded with 100+ benchmarks testing everything from reasoning to real-world tool use, yet LLM agents still perform poorly, sometimes scoring just 2% on challenging tasks.
Safety reviews remain limited to text analysis, leaving multimodal risks largely unexamined. Cost tracking and detailed diagnostics struggle to keep pace with deployment speed.
Moving forward means treating thorough, multi-dimensional evaluation as both an ethical obligation and a business necessity for teams committed to deploying trustworthy agents at scale.
The survey's four-dimensional framework makes one thing clear: you need evaluation tools that cover fundamental skills, domain objectives, agent behavior, and daily workflow health. Relying on scattered scripts or static tests creates blind spots in safety, cost, and diagnostics highlighted throughout current research.
Explore how Galileo integrates these insights into a single, production-ready platform, so you can monitor every dimension without juggling multiple tools or manual spreadsheets.
Modern LLM-based agents can draft contracts, triage customer tickets, and file code reviews, yet you rarely know whether those actions remain safe once the model goes live.
This first comprehensive survey of LLM-agent evaluation synthesized insights from more than 100 benchmarks and frameworks into 4 clear dimensions. If you've tried pushing agents beyond proof-of-concept, you know evaluation—not modeling—creates the real bottleneck.
Traditional ML metrics assume deterministic outputs and miss the erratic reasoning chains, tool calls, and emergent behaviors that define autonomous agents. The numbers tell the story: leading agents register success rates as low as 2% on the hardest tasks.
The researchers identify critical gaps—safety, cost-efficiency, and fine-grained diagnostics—setting the stage for the framework and challenges explored ahead.
Explore the Research Paper: Survey on Evaluation of LLM-based Agents
We recently explored this topic on our Chain of Thought podcast, where industry experts shared practical insights and real-world implementation strategies

Summary: Four Critical Dimensions of Agent Evaluation
The researchers neatly organize every agent testbed into four dimensions:
Fundamental capabilities
Application-specific tasks
Generalist reasoning
Evaluation frameworks themselves
This gives you a clear mental model for what was previously a scattered landscape.
The researchers' findings reveal some uncomfortable truths. On several difficult benchmarks, agents barely achieve 2% success—a stark reality captured in the survey's data visualizations. Common evaluation problems include limited automation and test planning challenges, while safety coverage and cost metrics receive less attention in existing surveys.
The good news? You can see the field evolving toward more robust, production-ready evaluation: live web environments replace static datasets, "Agent-as-a-Judge" scoring appears in advanced evaluation suites, and synthetic data generation creates endless stress tests.
These advances transform the survey into a practical roadmap for building safer, more cost-effective, and reliable agent systems in real-world applications.

The Five Challenges of Agent Evaluation
You might think your agent evaluation stack is complete after reviewing dozens of benchmarks. This research survey proves otherwise. By cataloging more than a hundred tests, this research exposes five blind spots that consistently derail production deployments.
Each gap affects your ability to trust, scale, and budget for autonomous systems. Fixing them means rethinking what you measure, how often, and under which real-world conditions.
Challenge #1: Safety and Compliance Evaluation Gaps
Modern agents can draft contracts, handle customer tickets, and review code, but do you know if they remain safe once deployed? The AIES meta-analysis of generative-AI safety work highlights three missing layers: modality coverage, risk coverage, and contextual assessment.
The research explains how most studies focus only on text models, a narrow set of ethical risks, and isolated outputs rather than actual user interactions. This leaves multimodal agents, privacy issues, and real user scenarios largely untested.
These gaps create real business risks for your organization. Without proper audit trails for toxic language, biased recommendations, or policy violations, you face potential regulatory fines and reputation damage. Safety becomes even more crucial as your agents gain autonomy—a single hallucination can cascade through an entire tool chain.
To protect your systems, you'll need multi-dimensional tests that reflect actual usage patterns. Leading teams combine automated toxicity filters with human reviews and log every decision for audits.
Modern frameworks built on accuracy, relevance, safety, and consistency principles already include such protections, flagging unsafe steps immediately. By extending these checks to image, audio, and code, you minimize potential harms while demonstrating compliance.
Challenge #2: Missing Cost-Efficiency and Resource Metrics
That impressive benchmark score looks less amazing when you see your first API bill. Most public leaderboards celebrate raw accuracy but ignore token counts, latency, or engineering costs. Few evaluations track inference cost or throughput, yet these factors often determine whether a model makes it to production.
Without cost data, you risk building agents that shine in tests but bankrupt you at scale. High token usage can cripple your real-time chat applications. Oversized context windows slow critical operations like fraud detection.
Avoiding these financial surprises requires tracking resources in every experiment you run. You should treat cost as equally important as quality and safety, recording compute, memory, and response time per scenario.
Many successful teams use "cost-normalized" scores, dividing accuracy by dollars spent, making trade-offs crystal clear to your decision-makers.
Modern frameworks handle this accounting automatically, showing token consumption alongside success rates. By standardizing these measures, you can test different temperatures, model sizes, or retrieval depths and immediately see the financial impact on your budget.
Challenge #3: Inadequate Fine-Grained Analysis
Pass/fail metrics look neat on paper but help little when your agent fails mid-conversation. The survey criticizes these broad evaluations for hiding where and why breakdowns happen. Eugene Yan illustrates this problem: end-to-end scores lump tool selection, reasoning, and execution into a single number, making it impossible for you to identify the weak link.
Debugging becomes mere guesswork without detailed traces—you end up changing prompts, temperature, or memory settings until something works. Fine-grained analysis solves this by scoring each step:
Did the agent pick the right API?
Was its reasoning logical?
Did it execute the final action correctly?
Advanced tools visualize these paths and highlight failure points in real time for your team. Framework extensions include step-by-step scoring, enabling automated reviews across thousands of runs.
Implementing this approach starts with recording every observation, thought, and action token—challenging but essential for your diagnostic process. Once captured, you can link failures to specific prompt patterns or tool uses and prioritize the most impactful fixes. Over time, these detailed metrics become early warning signs in your monitoring.
Drops in reasoning quality often precede complete task failure, giving you time to address issues before users notice.
Challenge #4: Scalability and Automation Limitations
Manual evaluations work for ten prompts but collapse when your system handles millions daily. The SIGIR LLM4Eval workshop points out that static, human-labeled tests quickly become outdated as models learn to exploit patterns in the data. Re-labeling every cycle costs too much of your time and money, undermining fast release schedules.
Automation helps, but implementing simple approaches creates new problems for your team. LLM-as-a-judge techniques scale subjective scoring but may inherit the very biases you're trying to detect.
Best practices suggest hybrid methods:
Combine automated initial grading with targeted human reviews of edge cases
Refine judge prompts to improve alignment with your quality standards
Synthetic data generation further reduces your annotation work. Advanced frameworks create scenario variations that test corner cases without exhausting reviewer capacity. By running these tests continuously—after every code change or knowledge update—you shift from occasional certification to constant quality monitoring across your deployment.
The key is balancing speed with reliability in your evaluation strategy. Automated systems must show confidence scores and uncertainty levels so you know when human review is needed. Done right, scalable evaluation becomes an always-on safety net that grows with your system rather than falling behind.
Challenge #5: Lack of Realistic, Dynamic Evaluation Environments
Many agents excel at toy problems but stumble when facing real-world use cases. The survey notes that models scoring above 80% on synthetic tests can drop to single-digit success rates when facing open-ended web environments. Static benchmarks freeze a moment in time; real websites, APIs, and your users' needs change constantly.
Leading security teams advocate for "live" benchmarks that pull fresh data or make actual API calls, though adoption remains limited. Until these resources mature, you can approximate reality by embedding evaluation in your staging systems.
Let your agent navigate a cloned customer portal, submit a test support ticket, or refactor actual GitHub issues. Each run captures rich data—response delays, unexpected page changes, authentication errors—that simple text matching misses.
Dynamic environments also reveal interaction nuances your team needs to understand: how the agent handles rate limits, retries, or partial failures. These edge cases rarely appear in curated datasets yet dominate production incidents. By prioritizing realistic simulation, you align offline metrics with real-world performance expectations.
Some evaluation frameworks are beginning to explore multi-dimensional approaches with ongoing, adaptive testing to keep measurements relevant as your systems evolve, though standardized frameworks with integrated replay environments don't yet exist.
The benefit is clear: agents that survive dynamic evaluation deliver more consistent performance, fewer surprises, and better user experiences in your production environment.
Practical Takeaways
You can apply the survey's insights to your daily engineering practice without waiting for new tools:
Many teams discover evaluation gaps only after deploying agents to production. Start by comparing your current tests against the paper's four-dimensional structure
While simple pass/fail metrics provide basic information, your production agents need deeper analysis. Add detailed, step-by-step checks—like reasoning review—to your existing evaluation system.
Trace-based tools show how granular assessment catches problems that binary scoring misses.
Safety and compliance need top priority in your evaluation pipeline, not afterthought status. The gaps in modality and context in the survey show what happens when safety evaluation takes a back seat.
Cost factors directly impact whether deployment makes sense for your business, regardless of accuracy. Your evaluation should track token usage, response time, and API expenses alongside performance to avoid models that cost more than they save.
Static evaluations quickly become obsolete as your agent capabilities grow. Implement live, continuously updated evaluations that adapt to changing needs and new challenges.
For critical applications, pair automated scoring with regular human checks. This combined approach remains the most reliable way to catch subtle failures that automated systems might miss.
Final Thoughts
The survey exposes a harsh truth: evaluation science has exploded with 100+ benchmarks testing everything from reasoning to real-world tool use, yet LLM agents still perform poorly, sometimes scoring just 2% on challenging tasks.
Safety reviews remain limited to text analysis, leaving multimodal risks largely unexamined. Cost tracking and detailed diagnostics struggle to keep pace with deployment speed.
Moving forward means treating thorough, multi-dimensional evaluation as both an ethical obligation and a business necessity for teams committed to deploying trustworthy agents at scale.
The survey's four-dimensional framework makes one thing clear: you need evaluation tools that cover fundamental skills, domain objectives, agent behavior, and daily workflow health. Relying on scattered scripts or static tests creates blind spots in safety, cost, and diagnostics highlighted throughout current research.
Explore how Galileo integrates these insights into a single, production-ready platform, so you can monitor every dimension without juggling multiple tools or manual spreadsheets.
Modern LLM-based agents can draft contracts, triage customer tickets, and file code reviews, yet you rarely know whether those actions remain safe once the model goes live.
This first comprehensive survey of LLM-agent evaluation synthesized insights from more than 100 benchmarks and frameworks into 4 clear dimensions. If you've tried pushing agents beyond proof-of-concept, you know evaluation—not modeling—creates the real bottleneck.
Traditional ML metrics assume deterministic outputs and miss the erratic reasoning chains, tool calls, and emergent behaviors that define autonomous agents. The numbers tell the story: leading agents register success rates as low as 2% on the hardest tasks.
The researchers identify critical gaps—safety, cost-efficiency, and fine-grained diagnostics—setting the stage for the framework and challenges explored ahead.
Explore the Research Paper: Survey on Evaluation of LLM-based Agents
We recently explored this topic on our Chain of Thought podcast, where industry experts shared practical insights and real-world implementation strategies

Summary: Four Critical Dimensions of Agent Evaluation
The researchers neatly organize every agent testbed into four dimensions:
Fundamental capabilities
Application-specific tasks
Generalist reasoning
Evaluation frameworks themselves
This gives you a clear mental model for what was previously a scattered landscape.
The researchers' findings reveal some uncomfortable truths. On several difficult benchmarks, agents barely achieve 2% success—a stark reality captured in the survey's data visualizations. Common evaluation problems include limited automation and test planning challenges, while safety coverage and cost metrics receive less attention in existing surveys.
The good news? You can see the field evolving toward more robust, production-ready evaluation: live web environments replace static datasets, "Agent-as-a-Judge" scoring appears in advanced evaluation suites, and synthetic data generation creates endless stress tests.
These advances transform the survey into a practical roadmap for building safer, more cost-effective, and reliable agent systems in real-world applications.

The Five Challenges of Agent Evaluation
You might think your agent evaluation stack is complete after reviewing dozens of benchmarks. This research survey proves otherwise. By cataloging more than a hundred tests, this research exposes five blind spots that consistently derail production deployments.
Each gap affects your ability to trust, scale, and budget for autonomous systems. Fixing them means rethinking what you measure, how often, and under which real-world conditions.
Challenge #1: Safety and Compliance Evaluation Gaps
Modern agents can draft contracts, handle customer tickets, and review code, but do you know if they remain safe once deployed? The AIES meta-analysis of generative-AI safety work highlights three missing layers: modality coverage, risk coverage, and contextual assessment.
The research explains how most studies focus only on text models, a narrow set of ethical risks, and isolated outputs rather than actual user interactions. This leaves multimodal agents, privacy issues, and real user scenarios largely untested.
These gaps create real business risks for your organization. Without proper audit trails for toxic language, biased recommendations, or policy violations, you face potential regulatory fines and reputation damage. Safety becomes even more crucial as your agents gain autonomy—a single hallucination can cascade through an entire tool chain.
To protect your systems, you'll need multi-dimensional tests that reflect actual usage patterns. Leading teams combine automated toxicity filters with human reviews and log every decision for audits.
Modern frameworks built on accuracy, relevance, safety, and consistency principles already include such protections, flagging unsafe steps immediately. By extending these checks to image, audio, and code, you minimize potential harms while demonstrating compliance.
Challenge #2: Missing Cost-Efficiency and Resource Metrics
That impressive benchmark score looks less amazing when you see your first API bill. Most public leaderboards celebrate raw accuracy but ignore token counts, latency, or engineering costs. Few evaluations track inference cost or throughput, yet these factors often determine whether a model makes it to production.
Without cost data, you risk building agents that shine in tests but bankrupt you at scale. High token usage can cripple your real-time chat applications. Oversized context windows slow critical operations like fraud detection.
Avoiding these financial surprises requires tracking resources in every experiment you run. You should treat cost as equally important as quality and safety, recording compute, memory, and response time per scenario.
Many successful teams use "cost-normalized" scores, dividing accuracy by dollars spent, making trade-offs crystal clear to your decision-makers.
Modern frameworks handle this accounting automatically, showing token consumption alongside success rates. By standardizing these measures, you can test different temperatures, model sizes, or retrieval depths and immediately see the financial impact on your budget.
Challenge #3: Inadequate Fine-Grained Analysis
Pass/fail metrics look neat on paper but help little when your agent fails mid-conversation. The survey criticizes these broad evaluations for hiding where and why breakdowns happen. Eugene Yan illustrates this problem: end-to-end scores lump tool selection, reasoning, and execution into a single number, making it impossible for you to identify the weak link.
Debugging becomes mere guesswork without detailed traces—you end up changing prompts, temperature, or memory settings until something works. Fine-grained analysis solves this by scoring each step:
Did the agent pick the right API?
Was its reasoning logical?
Did it execute the final action correctly?
Advanced tools visualize these paths and highlight failure points in real time for your team. Framework extensions include step-by-step scoring, enabling automated reviews across thousands of runs.
Implementing this approach starts with recording every observation, thought, and action token—challenging but essential for your diagnostic process. Once captured, you can link failures to specific prompt patterns or tool uses and prioritize the most impactful fixes. Over time, these detailed metrics become early warning signs in your monitoring.
Drops in reasoning quality often precede complete task failure, giving you time to address issues before users notice.
Challenge #4: Scalability and Automation Limitations
Manual evaluations work for ten prompts but collapse when your system handles millions daily. The SIGIR LLM4Eval workshop points out that static, human-labeled tests quickly become outdated as models learn to exploit patterns in the data. Re-labeling every cycle costs too much of your time and money, undermining fast release schedules.
Automation helps, but implementing simple approaches creates new problems for your team. LLM-as-a-judge techniques scale subjective scoring but may inherit the very biases you're trying to detect.
Best practices suggest hybrid methods:
Combine automated initial grading with targeted human reviews of edge cases
Refine judge prompts to improve alignment with your quality standards
Synthetic data generation further reduces your annotation work. Advanced frameworks create scenario variations that test corner cases without exhausting reviewer capacity. By running these tests continuously—after every code change or knowledge update—you shift from occasional certification to constant quality monitoring across your deployment.
The key is balancing speed with reliability in your evaluation strategy. Automated systems must show confidence scores and uncertainty levels so you know when human review is needed. Done right, scalable evaluation becomes an always-on safety net that grows with your system rather than falling behind.
Challenge #5: Lack of Realistic, Dynamic Evaluation Environments
Many agents excel at toy problems but stumble when facing real-world use cases. The survey notes that models scoring above 80% on synthetic tests can drop to single-digit success rates when facing open-ended web environments. Static benchmarks freeze a moment in time; real websites, APIs, and your users' needs change constantly.
Leading security teams advocate for "live" benchmarks that pull fresh data or make actual API calls, though adoption remains limited. Until these resources mature, you can approximate reality by embedding evaluation in your staging systems.
Let your agent navigate a cloned customer portal, submit a test support ticket, or refactor actual GitHub issues. Each run captures rich data—response delays, unexpected page changes, authentication errors—that simple text matching misses.
Dynamic environments also reveal interaction nuances your team needs to understand: how the agent handles rate limits, retries, or partial failures. These edge cases rarely appear in curated datasets yet dominate production incidents. By prioritizing realistic simulation, you align offline metrics with real-world performance expectations.
Some evaluation frameworks are beginning to explore multi-dimensional approaches with ongoing, adaptive testing to keep measurements relevant as your systems evolve, though standardized frameworks with integrated replay environments don't yet exist.
The benefit is clear: agents that survive dynamic evaluation deliver more consistent performance, fewer surprises, and better user experiences in your production environment.
Practical Takeaways
You can apply the survey's insights to your daily engineering practice without waiting for new tools:
Many teams discover evaluation gaps only after deploying agents to production. Start by comparing your current tests against the paper's four-dimensional structure
While simple pass/fail metrics provide basic information, your production agents need deeper analysis. Add detailed, step-by-step checks—like reasoning review—to your existing evaluation system.
Trace-based tools show how granular assessment catches problems that binary scoring misses.
Safety and compliance need top priority in your evaluation pipeline, not afterthought status. The gaps in modality and context in the survey show what happens when safety evaluation takes a back seat.
Cost factors directly impact whether deployment makes sense for your business, regardless of accuracy. Your evaluation should track token usage, response time, and API expenses alongside performance to avoid models that cost more than they save.
Static evaluations quickly become obsolete as your agent capabilities grow. Implement live, continuously updated evaluations that adapt to changing needs and new challenges.
For critical applications, pair automated scoring with regular human checks. This combined approach remains the most reliable way to catch subtle failures that automated systems might miss.
Final Thoughts
The survey exposes a harsh truth: evaluation science has exploded with 100+ benchmarks testing everything from reasoning to real-world tool use, yet LLM agents still perform poorly, sometimes scoring just 2% on challenging tasks.
Safety reviews remain limited to text analysis, leaving multimodal risks largely unexamined. Cost tracking and detailed diagnostics struggle to keep pace with deployment speed.
Moving forward means treating thorough, multi-dimensional evaluation as both an ethical obligation and a business necessity for teams committed to deploying trustworthy agents at scale.
The survey's four-dimensional framework makes one thing clear: you need evaluation tools that cover fundamental skills, domain objectives, agent behavior, and daily workflow health. Relying on scattered scripts or static tests creates blind spots in safety, cost, and diagnostics highlighted throughout current research.
Explore how Galileo integrates these insights into a single, production-ready platform, so you can monitor every dimension without juggling multiple tools or manual spreadsheets.
Conor Bronsdon
Conor Bronsdon
Conor Bronsdon
Conor Bronsdon