Your AI agent labored within the demo, impressed stakeholders, dealt with take a look at eventualities, and appeared prepared for manufacturing. Then you definately deployed it, and the image modified. Actual customers skilled incorrect device calls, inconsistent responses, and failure modes no one anticipated throughout testing.
The result’s a niche between anticipated agent conduct and precise person expertise in manufacturing. Agent analysis introduces challenges that conventional software program testing wasn’t designed to deal with. As a result of giant language fashions (LLMs) are non-deterministic, the identical person question can produce totally different device choices, reasoning paths, and outputs throughout a number of runs. Which means you will need to take a look at every state of affairs repeatedly to know your agent’s precise conduct patterns. A single take a look at cross tells you what can occur, not what usually occurs. With out systematic measurement throughout these variations, groups are trapped in cycles of handbook testing and reactive debugging. This burns by means of API prices with out clear perception into whether or not adjustments enhance agent efficiency. This uncertainty makes each immediate modification dangerous and leaves a basic query unanswered: “Is that this agent really higher now?”
On this submit, we introduce Amazon Bedrock AgentCore Evaluations, a completely managed service for assessing AI agent efficiency throughout the event lifecycle. We stroll by means of how the service measures agent accuracy throughout a number of high quality dimensions. We clarify the 2 analysis approaches for growth and manufacturing and share sensible steering for constructing brokers you’ll be able to deploy with confidence.
Why agent analysis requires a brand new method
When a person sends a request to an agent, a number of selections occur in sequence. The agent determines which instruments (if any) to name, executes these calls, and generates a response based mostly on the outcomes. Every step introduces potential failure factors: choosing the incorrect device, calling the fitting device with incorrect parameters, or synthesizing device outputs into an inaccurate last reply. Not like conventional functions the place you take a look at a single perform’s output, agent analysis requires measuring high quality throughout this complete interplay circulate.
This creates particular challenges for agent builders that may be addressed by doing the next:
- Outline analysis standards on what constitutes an accurate device choice, legitimate device parameters, an correct response, and a useful person expertise.
- Construct take a look at datasets that symbolize actual person requests and anticipated behaviors.
- Select scoring strategies that may assess high quality constantly throughout repeated runs.
Every of those definitions immediately determines what your analysis system measures and getting them incorrect means optimizing for the incorrect outcomes. With out this foundational work, the hole between what groups hope their brokers do and what they’ll show their brokers do turns into an actual enterprise danger.Bridging this hole requires a steady analysis cycle, as proven in Determine 1. Groups construct take a look at circumstances, run them in opposition to the agent, rating the outcomes, analyze failures, and implement enhancements. Every failure turns into a brand new take a look at case, and the cycle continues by means of each iteration of the agent.
Determine 1: The agent analysis course of follows a steady cycle of take a look at circumstances, agent execution, scoring, evaluation, and enhancements. Failures develop into new take a look at circumstances.
Working this cycle finish to finish, nevertheless, requires important infrastructure past the analysis logic itself. Groups should curate datasets, choose and host scoring fashions, handle inference capability and API charge limits, construct knowledge pipelines that rework agent traces into evaluation-ready codecs, and create dashboards to visualise developments. For organizations working a number of brokers, this overhead multiplies with every one. The result’s that agent developer groups find yourself spending extra time sustaining analysis tooling than performing on what it tells them. That is the issue Amazon Bedrock AgentCore Evaluations was constructed to deal with.
Introducing Amazon Bedrock AgentCore Evaluations
First launched in public preview at AWS re:Invent 2025, the service is now typically out there. It handles the analysis fashions, inference infrastructure, knowledge pipelines, and scaling so groups can deal with bettering agent high quality reasonably than constructing and sustaining analysis programs. For built-in evaluators, mannequin quota and inference capability are totally managed. Which means organizations evaluating many brokers aren’t consuming their very own quotas or provisioning separate infrastructure for analysis workloads.
AgentCore Evaluations study agent conduct end-to-end utilizing OpenTelemetry (OTEL) traces with generative AI semantic conventions. OTEL is an open supply observability customary for gathering distributed traces from functions. The generative AI semantic conventions lengthen it with fields particular to language mannequin interactions, together with prompts, completions, device calls, and mannequin parameters. By constructing on this customary, the service works constantly throughout brokers constructed with any Strands Agents or LangGraph, and instrumented with OpenTelemetry and OpenInference, capturing the complete context wanted for significant analysis.
The evaluations could be configured with totally different approaches:
- LLM-as-a-Judge the place an LLM evaluates every agent interplay in opposition to structured rubrics with clearly outlined standards.
- Floor Fact based mostly analysis can be utilized to check the agent responses in opposition to pre-defined or simulated datasets.
- Customized code evaluators the place you’ll be able to usher in a Lambda as a evaluator with your individual customized code.
Within the LLM-as-a-Decide method, the Decide mannequin examines the complete interplay context, together with dialog historical past, out there instruments, instruments used, parameters handed, and system directions, then gives detailed reasoning earlier than assigning a rating. Each rating comes with a proof. Groups can use these scores to confirm judgments, perceive precisely why an interplay acquired a specific ranking, and determine what ought to have occurred in a different way. This method goes past easy cross/fail judgments, offering the structured analysis and clear reasoning that allow high quality evaluation at a scale that handbook assessment can’t match.
Three rules information how the service approaches analysis. Proof-driven growth replaces instinct with quantitative metrics, so groups can measure the precise impression of adjustments reasonably than debating whether or not a immediate modification “feels higher.” Multi-dimensional evaluation evaluates totally different features of agent conduct independently. This makes it attainable to pinpoint precisely the place enhancements are wanted reasonably than counting on a single combination rating. Steady measurement connects the efficiency baselines established throughout growth on to manufacturing monitoring, ensuring that high quality holds up as real-world situations evolve. These rules apply all through the agent lifecycle, from the primary spherical of growth testing by means of ongoing manufacturing monitoring.
Analysis throughout the agent lifecycle
An agent’s journey from prototype to manufacturing creates two distinct analysis wants. Throughout growth, groups want managed environments the place they’ll evaluate options, take a look at the agent on curated datasets, reproduce outcomes, and validate adjustments earlier than they attain customers. After the agent is dwell, the problem shifts to monitoring real-world interactions at scale, the place customers encounter edge circumstances and interplay patterns that no quantity of pre-deployment testing anticipated. Determine 2 illustrates how analysis helps every stage of this journey, from preliminary proof of idea by means of shadow testing, A/B testing, and steady manufacturing monitoring.
Determine 2: From POC to manufacturing, analysis validates brokers earlier than deployment. As brokers mature, analysis helps shadow testing, A/B testing, and steady monitoring at scale.
AgentCore Evaluations map two complementary approaches to those lifecycle phases, as proven in Determine 3. On-line analysis handles steady manufacturing monitoring, whereas on-demand analysis helps managed testing throughout growth and steady integration and steady supply (CI/CD) workflows, together with evaluations in opposition to floor fact.
| On-demand Analysis | On-line Analysis | |
| Benefits |
|
|
| Use circumstances |
|
|
Determine 3: On-line analysis displays manufacturing site visitors repeatedly, whereas on-demand analysis helps managed testing throughout growth.
On-line analysis for manufacturing monitoring
On-line analysis displays dwell agent interactions by repeatedly sampling a configurable share of traces and scoring them in opposition to your chosen evaluators. You outline which evaluators to use, set sampling guidelines that management what fraction of manufacturing site visitors will get evaluated, and arrange acceptable filters. The service handles studying traces, working evaluations, and surfacing leads to the AgentCore Observability dashboard powered by Amazon CloudWatch. In the event you’re already gathering traces for observability, on-line analysis provides high quality scores with rationalization, alongside your present operational metrics with out requiring code adjustments or re-deployments. Determine 4 reveals how this course of works.

High quality points in manufacturing usually floor in ways in which conventional monitoring misses. Operational dashboards could present inexperienced throughout latency and error charges whereas person expertise quietly degrades as a result of the agent begins choosing incorrect instruments or offering much less useful responses. Steady high quality scoring catches these silent failures by monitoring analysis metrics alongside operational ones. As a result of AgentCore Observability runs on CloudWatch, you’ll be able to create customized dashboards and set alarms to get alerted the second scores drop under your thresholds.
On-demand analysis for growth
On-demand analysis is a real-time API designed for growth and CI/CD workflows. Groups use it to check adjustments earlier than deployment, run analysis suites as a part of CI/CD pipelines, carry out regression testing throughout builds, and gate deployments on high quality thresholds. Builders choose a full session and specify actual spans (particular person operations inside a hint) or traces by offering their IDs. The service considers the complete session dialog and scores particular person span/traces in opposition to the identical evaluators utilized in manufacturing. Widespread use circumstances embody validating immediate adjustments, evaluating mannequin efficiency throughout options, and stopping high quality regressions.
Determine 5: On-demand analysis permits builders to organize hint datasets, invoke evaluations by means of a CI/CD pipeline or growth surroundings, and obtain scores utilizing built-in or customized evaluators powered by Amazon Bedrock basis fashions.
As a result of each modes use the identical evaluators, what you take a look at in CI/CD is what you monitor in manufacturing, providing you with constant high quality requirements throughout all the growth lifecycle. On-demand analysis gives the managed surroundings wanted for structure selections and systematic enchancment, whereas on-line analysis maintains high quality monitoring continues after the agent is dwell. Collectively, the 2 modes type a steady suggestions loop between growth and manufacturing, and each draw from the identical set of evaluators and scoring infrastructure.
How AgentCore evaluates your agent
AgentCore Evaluations organizes agent interactions right into a three-level hierarchy that determines what could be evaluated and at what granularity. A session represents a whole dialog between a person and your agent, grouping all associated interactions from a single person or workflow. Inside every session, a hint captures every little thing that occurs throughout a single alternate. When a person sends a message and receives a response, that spherical journey produces one hint containing each step that the agent took to generate its reply. Every hint in flip comprises particular person operations referred to as spans, representing particular actions your agent carried out, equivalent to invoking a device, retrieving info from a information base, or producing textual content.
Totally different evaluators function at totally different ranges of this hierarchy, and issues at one stage can look very totally different from issues at one other. The service gives 13 pre-configured built-in evaluators organized throughout these three ranges, every measuring a definite facet of agent conduct (Determine 6). You may outline customized evaluators utilizing LLM-as-a-Decide and customized code evaluators that may work on session, hint and span ranges.
| Degree | Evaluators | Objective | Floor Fact Use |
| Session | Aim Success Price | Assesses whether or not all person targets have been accomplished inside a dialog | Person gives free type textual assertions of aim completion, that are in contrast in opposition to system conduct and measured through Aim Success Price |
| Hint | Helpfulness, Correctness, Coherence, Conciseness, Faithfulness, Harmfulness, Instruction Following, Response Relevance, Context Relevance, Refusal, Stereotyping | Evaluates response high quality, accuracy, security, and communication effectiveness | Flip stage floor fact (e.g., anticipated reply or attributes per flip) helps analysis of Correctness |
| Device | Device Choice Accuracy, Device Parameter Accuracy | Assesses device choice selections and parameter extraction precision | Device name floor fact specifies the proper device sequence enabling Trajectory Actual Order Match, Trajectory In-Order Match, and Trajectory Any Order Match |
Determine 6: Constructed-in evaluators function at session, hint, and gear ranges. Every stage measures totally different features of agent conduct. Floor Fact could be offered as assertions, anticipated response and anticipated trajectory for analysis on session, hint and gear stage.
Evaluating every stage independently helps groups to diagnose whether or not an issue originates in device choice, response technology, or session-level planning. An agent may select the fitting device with correct parameters however then synthesize the device’s output poorly in its last response. This sample solely turns into seen when every stage is assessed by itself. Your agent’s major goal guides which evaluators to prioritize. Customer support brokers ought to deal with Helpfulness, Aim Success Price, and Instruction Following, since resolving person points inside outlined guardrails immediately impacts satisfaction. Brokers with Retrieval Augmented Era (RAG) elements profit most from Correctness and Faithfulness to be sure that responses are grounded within the offered context. Device-heavy brokers want sturdy Device Choice Accuracy and Device Parameter Accuracy scores. It’s beneficial to start out with three or 4 evaluators that align along with your agent’s goal and broaden protection as your understanding matures.
Understanding evaluator distinctions
Some evaluators naturally work together with one another, so scores must be learn collectively reasonably than in isolation. Evaluators that sound related usually measure basically various things, and understanding these distinctions is necessary for analysis.
- Correctness checks whether or not the response is factually correct, whereas Faithfulness checks whether or not it’s per the dialog historical past. For instance, an agent could be trustworthy to flawed supply materials however nonetheless incorrect.
- Helpfulness asks whether or not the response advances the person towards their aim, whereas Response Relevance asks whether or not it addresses what was initially requested. For instance, an agent can reply the incorrect query completely.
- Coherence checks for inside contradictions in reasoning, whereas Context Relevance checks whether or not the agent had the fitting info out there. For instance, one reveals a technology drawback, the opposite a retrieval drawback.
Some evaluators additionally rely on or trade-off in opposition to one another. As an illustration:
- Device Parameter Accuracy is significant solely when the agent has chosen the proper device, so low Device Choice Accuracy must be addressed first.
- Correctness usually relies on Context Relevance as a result of an agent can’t generate correct solutions with out the fitting info.
- Conciseness and Helpfulness usually battle as a result of temporary responses may omit context that customers want.
Constructed-in evaluators ship with predefined immediate templates, chosen evaluator fashions, and standardized scoring standards, with configurations mounted to protect consistency throughout evaluations. They use cross-Area inference to mechanically choose compute from AWS Areas inside your geography, bettering mannequin availability and throughput whereas maintaining knowledge saved within the originating Area. Customized evaluators lengthen this basis with assist on your personal evaluator mannequin, analysis directions, standards, and scoring schema. They’re notably precious for industry-specific assessments equivalent to compliance checking in healthcare or monetary providers, model voice consistency verification, or implementing organizational high quality requirements. Customized code evaluators allow you to usher in an AWS Lambda perform to carry out the evaluations. This lets you additionally create deterministic scoring of your brokers.
To be used circumstances requiring all processing inside a single Area, customized evaluators additionally present full management over inference configuration. When constructing a customized evaluator, you outline directions with placeholders that get changed with precise hint info earlier than being despatched to the choose mannequin. The scope of data out there relies on the evaluator’s stage: a session-level evaluator can entry the complete dialog context and out there instruments, a trace-level evaluator sees earlier turns plus the present assistant response, and a tool-level evaluator focuses on particular device calls inside their surrounding context. The AWS console gives the choice to load the immediate template of any present built-in evaluator as a place to begin, making it easy to create customized variants (Determine 7).
Determine 7: The AgentCore Evaluations console gives the choice to load any built-in evaluator’s immediate template as a place to begin when making a customized evaluator.
When constructing a number of customized evaluators, use the MECE (Mutually Unique, Collectively Exhaustive) precept to design your analysis suite. Every evaluator ought to have a definite, non-overlapping scope whereas collectively overlaying all high quality dimensions you care about. For instance, reasonably than creating two evaluators that each partially assess “response high quality,” separate them into one which evaluates factual grounding and one other that evaluates communication readability. Moreover, to put in writing evaluator directions, set up the choose mannequin’s position as a efficiency evaluator to stop confusion between analysis and process execution. Use clear, sequential directions with exact language, and think about together with one to 3 related examples with matching enter/output pairs that symbolize your anticipated requirements. For scoring, select between binary scales (0/1) for cross/fail eventualities or ordinal scales (equivalent to 1–5) for extra nuanced assessments, and begin with binary scoring when unsure. The service standardizes output to incorporate a cause area adopted by a rating area, so the choose mannequin at all times presents its reasoning earlier than assigning a quantity. Keep away from together with your individual output formatting directions, as they’ll confuse the Decide mannequin.
Customized Code-based evaluators
Constructed-in and customized evaluators each use an LLM-as-a-Decide. AgentCore Evaluations additionally helps a 3rd method: code-based evaluators, the place an AWS Lambda perform can be utilized because the evaluator along with your customized code.
Code-based evaluators are ideally suited when you have got heuristic scoring strategies that don’t require language understanding to confirm. An LLM evaluator can choose whether or not a response “sounds right,” however it can’t reliably verify {that a} particular pay stub determine of $8,333.33 seems verbatim in a response, or {that a} generated request ID follows the format PTO-2026-NNN. For these deterministic checks, a customized code is quicker, cheaper, and extra dependable. There are 4 conditions the place code-based evaluators are notably useful:
- Actual knowledge validation: The agent is predicted to return particular values from a knowledge supply, equivalent to account balances, transaction IDs, or costs.
- Format compliance: Responses should conform to structural constraints, equivalent to size limits, required phrases, or output schemas.
- Enterprise rule enforcement: Insurance policies that require exact interpretation, equivalent to whether or not a response accurately applies a tiered low cost rule or cites the fitting regulatory clause.
- Excessive-volume manufacturing monitoring: Lambda invocations price a fraction of LLM inference, making code-based evaluators the fitting selection when each manufacturing session must be scored repeatedly at scale.
Making a code-based evaluator
A code-based evaluator is configured as an AWS Lambda perform along with your customized logic. AgentCore passes the agent’s OTel spans to your perform as a structured occasion and expects a end in return. Your perform extracts no matter info it wants from the spans and returns a rating, a label, and a proof.
As soon as your Lambda is deployed and granted permission to be invoked by the AgentCore service principal, you register it as an evaluator for AgentCore. As soon as registered, the evaluator ID can be utilized for on-demand analysis.
Establishing AgentCore Evaluations
Configuring the service entails three steps. Choose your agent, select your evaluators, and set your sampling guidelines. Earlier than you start, deploy your agent utilizing AgentCore Runtime and arrange observability by means of OpenTelemetry or OpenInference instrumentation. The AgentCore samples repository on GitHub gives working examples.
Configuring on-line analysis
Create a brand new on-line analysis configuration by means of the AgentCore Evaluations console. Right here, you specify which evaluators to use, which knowledge supply to observe, and what sampling parameters to make use of. For the information supply, choose both an present AgentCore Runtime endpoint or a CloudWatch log group for brokers not hosted on AgentCore Runtime. Then select your evaluators and outline your sampling guidelines.
Determine 8: The AgentCore Evaluations console for creating a web based analysis configuration, together with knowledge supply choice, evaluator task, and sampling guidelines.
You may also create configurations programmatically utilizing the CreateOnlineEvaluationConfig API with a novel configuration title, knowledge supply, checklist of evaluators (as much as 10), and IAM service position. The enableOnCreate parameter controls whether or not analysis begins instantly or stays paused, and executionStatus determines whether or not the configuration actively processes traces as soon as enabled. When a configuration is working, any customized evaluators it references develop into locked and can’t be modified or deleted. If you must change an evaluator, clone it and create a brand new model. On-line analysis outcomes are saved to a devoted CloudWatch log group in JSON format.
Monitoring outcomes
After enabling your configuration, monitor outcomes by means of the AgentCore Observability dashboard in Amazon CloudWatch. Agent-level views show aggregated analysis metrics and developments, and you may drill into particular classes and traces to see particular person scores and the reasoning behind every one.
Determine 9: The AgentCore Observability dashboard shows analysis metrics and developments on the agent stage, with drill-down into particular person classes, traces, scores, and choose reasoning.
Drilling into a person hint reveals the analysis scores and detailed explanations for that particular interplay, so groups can confirm choose reasoning and perceive why the agent acquired a specific ranking.
Determine 10: The trace-level view shows analysis scores and explanations immediately on particular person traces, exhibiting the choose mannequin’s reasoning for every metric.
Utilizing on-demand analysis
For growth and testing, you should use on-demand analysis to investigate particular interactions by choosing the traces or spans that you simply need to study, making use of your chosen evaluators, and receiving detailed scores with explanations. Outcomes return immediately within the API response, restricted to 10 evaluations per name, with every end result containing the span context, rating, and reasoning. If an analysis partially fails, the response contains each profitable and failed outcomes with error codes and messages. On-demand analysis works effectively for testing customized evaluators, investigating particular high quality points, and validating fixes earlier than deployment.
Evaluating brokers with floor fact
LLM-as-judge scoring tells you whether or not responses appear right and useful by the requirements of a general-purpose language mannequin. Floor fact analysis takes this additional by letting you specify the reply, the instruments that ought to have been referred to as, and the outcomes the session ought to have achieved. This helps you measure how intently the agent’s precise conduct matches your reference inputs. That is notably precious throughout growth, when you have got area information about what the fitting conduct is and need to take a look at for particular eventualities.
AgentCore Evaluations helps three sorts of floor fact reference inputs, every consumed by a selected set of evaluators:
| Reference Enter | Evaluator | What it measures |
expected_response |
Builtin.Correctness |
Similarity between the agent’s response and the known-correct reply |
expected_trajectory |
Builtin.TrajectoryExactOrderMatch, Builtin.TrajectoryInOrderMatch, Builtin.TrajectoryAnyOrderMatch |
Whether or not the agent referred to as the fitting instruments in the fitting sequence |
assertions |
Builtin.GoalSuccessRate |
Whether or not the session happy a set of natural-language statements about anticipated outcomes |
These inputs are elective and impartial. Evaluators that don’t require floor fact equivalent to Builtin.Helpfulness and Builtin.ResponseRelevance could be included in the identical name as ground-truth evaluators, and every evaluator reads solely the fields it wants. You may provide all three reference inputs concurrently for a complete analysis, or provide solely the subset related to a given state of affairs.
The bedrock-agentcore Python SDK gives two interfaces for floor fact analysis: EvaluationClient for assessing present classes and OnDemandEvaluationRunner for automated dataset analysis.
Analysis Consumer: Evaluating present classes
Analysis Consumer is the fitting selection when you have already got agent classes recorded in CloudWatch and need to consider particular interactions. You present the session ID, the agent ID, your chosen evaluators, a glance again window for CloudWatch span retrieval, and elective Reference Inputs. The consumer fetches the session’s spans and submits them for analysis. That is effectively suited to growth evaluation, debugging particular agent failures, and validating recognized interactions after immediate or mannequin adjustments.
Analysis Consumer works equally effectively for multi-turn classes. Once you cross a session ID from a multi-turn dialog, the consumer fetches all spans for that session and evaluates the entire dialogue. Trajectory evaluators confirm device utilization throughout all turns, aim success assertions apply to the session, and correctness evaluators rating every particular person response in opposition to its corresponding anticipated reply.
On-Demand Analysis Dataset Runner: Automated dataset analysis
On-Demand Analysis Dataset Runner is the fitting selection once you need to consider your agent systematically throughout a curated dataset by invoking the agent for each state of affairs, gathering CloudWatch spans, and scoring leads to a single automated workflow. You outline a Dataset containing multi-turn eventualities with per-turn and per-scenario floor fact and supply an agent_invoker perform that the runner requires every flip. The runner manages session IDs and handles all coordination between invocation, span assortment, and analysis.
On-Demand Analysis Dataset Runner is effectively suited to CI/CD pipelines the place the identical dataset runs in opposition to each construct, regression testing after immediate or mannequin adjustments, and batch analysis throughout a big corpus of take a look at circumstances earlier than a launch.
The 2 interfaces share the identical evaluators and Reference Inputs schema, so you’ll be able to develop and validate floor fact take a look at circumstances interactively with Analysis Consumer in opposition to present manufacturing classes, then promote those self same eventualities into your Analysis Runner dataset for systematic regression testing. The hands-on tutorial within the AgentCore samples repository demonstrates each interfaces end-to-end utilizing an instance agent throughout single-turn and multi-turn eventualities with all three sorts of floor fact reference inputs.
Finest practices
Success standards on your agent usually mix three dimensions: the standard of responses, the latency at which customers obtain them, and the price of inference. AgentCore Evaluations focuses on the standard dimension, whereas operational metrics like latency and price can be found by means of AgentCore Observability in CloudWatch. The next greatest practices are organized across the three analysis rules described earlier, and mirror patterns that emerge from working with agent analysis at scale.
Proof-driven growth
- Baseline your agent’s efficiency with each artificial and real-world knowledge, and experiment rigorously. Measure earlier than and after each change in order that enhancements are grounded in proof, not instinct. Begin testing early with the take a look at circumstances that you’ve, and construct your corpus repeatedly. The analysis loop described in Determine 1 makes positive that failures develop into new take a look at circumstances over time.
- Run A/B testing with statistical rigor for each change. Whether or not you’re updating a system immediate, swapping a mannequin, or including a device, evaluate efficiency throughout the identical evaluator set earlier than and after deployment.
- Run repeated trials (not less than 10 per query) organized by class to benchmark reliability and determine specialization alternatives. Variance throughout repeated runs reveals the place your agent is constant and the place it wants work.
Multi-dimensional evaluation
- Outline what success seems like early, utilizing multi-dimensional standards that mirror your agent’s precise goal. Contemplate which analysis ranges matter most (session, hint, or device) and choose evaluators that map to your corporation targets.
- Consider each step within the agent’s workflow, not simply last outcomes. Measuring device choice, parameter accuracy, and response high quality independently offers you the diagnostic precision to repair issues the place they really happen.
- Contain material specialists in designing your metrics, defining process protection, and conducting human-in-the-loop critiques for high quality assurance. SME enter retains your evaluators grounded in real-world expectations and catches blind spots that automated scoring alone can miss.
- Begin with built-in evaluators to determine baseline measurements, then create customized evaluators as your wants mature. Calibrate customized evaluator scoring with SMEs for automated judgments align with human expectations in your area.
Steady measurement
- Detect drift by evaluating manufacturing conduct to your take a look at baselines. Arrange CloudWatch alarms on key metrics so that you catch regressions earlier than they attain a broad set of customers.
- Keep in mind that your take a look at dataset evolves along with your agent, your customers, and the adversarial eventualities you encounter. Replace it usually as edge circumstances emerge in manufacturing and necessities shift.
Troubleshooting frequent Analysis patterns
- The evaluator relationships described earlier helps you interpret scores diagnostically. The next patterns are described for particular eventualities chances are you’ll encounter as you scale your utility together with steps to resolve them.
- In the event you discover low scores throughout all evaluators, the difficulty is usually foundational. Begin by reviewing Context Relevance scores to find out whether or not your agent has entry to the knowledge it wants. Test your agent’s system immediate for readability and completeness; obscure or contradictory directions have an effect on each downstream conduct. Confirm that device descriptions precisely clarify when and tips on how to use every device.
- In the event you discover inconsistent scores for related interactions, it normally factors to analysis configuration points reasonably than agent issues. If you’re utilizing customized evaluators, test whether or not your directions are particular sufficient and whether or not every rating stage has clear, distinguishable definitions. Contemplate reducing the temperature parameter in your customized evaluator’s mannequin configuration to provide extra deterministic scoring.
- In the event you see excessive Device Choice Accuracy however low Aim Success Price, your agent selects acceptable instruments however fails to finish person targets. This sample suggests that you simply may want further instruments to deal with sure person requests, or your agent struggles with duties requiring a number of sequential device calls. Test Helpfulness scores as effectively; the agent may use instruments accurately however clarify outcomes poorly.
- If evaluations are sluggish or failing as a result of throttling, decrease your sampling charge to guage a smaller share of classes. Scale back your evaluator rely. For customized evaluators, request quota will increase on your chosen mannequin, or change to a mannequin with increased default quotas.
Conclusion
On this submit, we confirmed how Amazon Bedrock AgentCore Evaluations helps groups transfer from reactive debugging to systematic high quality administration for AI brokers. As a completely managed service, it handles the analysis fashions, inference infrastructure, and knowledge pipelines that groups would in any other case must construct and preserve for every agent. With on-demand analysis anchoring the event workflow and on-line analysis offering steady manufacturing perception, high quality turns into a measurable and improvable property all through the agent lifecycle. The evaluator relationships and diagnostic patterns give a framework not simply to attain brokers however for understanding the place and why high quality points happen and the place to focus enchancment efforts.
To discover AgentCore Evaluations intimately, watch the public preview launch session from AWS re:Invent 2025 for a walkthrough with dwell demos. Go to the Amazon Bedrock AgentCore samples repository on GitHub for hands-on tutorials. For technical particulars on configuration and API utilization, see the AgentCore Evaluations documentation. You may also assessment service limits and pricing.
Concerning the authors


