ChanlChanl
Best Practices

Scorecards vs. Vibes: How to Actually Measure AI Agent Quality

Most teams 'feel' their AI agent is good. Here's how to build structured scoring with rubrics, automated grading, and regression detection that holds up.

Chanl TeamAI Agent Testing Platform
March 5, 2026
15 min read
Laptop and smartphone displaying data charts and metrics dashboards on a dark surface

Your AI agent had a great week. Calls sounded smooth. The team felt good about it. Nobody complained. So you shipped the new prompt to production.

Three days later, a customer complained that the agent had given completely wrong refund policy information — twice. The week before that, it had been perfect. Nothing changed in the model. Nothing changed in the data. But something changed in the agent's behavior, and nobody caught it because "feeling good about it" was the only measurement in place.

Sound familiar? This is the vibes-based quality trap, and almost every team building AI agents falls into it at some point. The fix isn't complicated — but it does require trading gut instinct for structure.

Why "Feels Good" Isn't a Quality Strategy

Here's the thing about human intuition: it's excellent at catching catastrophic failures and terrible at catching gradual degradation. When an agent completely refuses to answer a question, you notice. When it starts subtly hedging on refund eligibility — becoming 10% less helpful over a series of prompt iterations — you probably won't, at least not until a customer escalates.

Missing slow regressions is just the start. Vibes-based evaluation has a few other structural problems:

It doesn't scale. When you're handling 50 test conversations a week, a human can reasonably skim them. At 500, coverage collapses. At 5,000, you're sampling at best, flying blind at worst.

It's inconsistent. The same conversation evaluated by two different people on two different days will often get different verdicts. That's not a character flaw — it's how human judgment works without a rubric to anchor it.

It can't catch cross-dimensional failures. An agent might be great at empathy but consistently miss capturing key information. Or it might follow your policy script perfectly while completely failing to actually resolve the customer's problem. You need to measure both dimensions independently.

And perhaps most practically: when something goes wrong and leadership asks "how is our agent performing?" — "it felt pretty good last week" is not a defensible answer.

You can't improve what you can't measure. With AI agents, 'it seemed fine' is not a measurement — it's a hope.
Industry ObservationCommon wisdom in production AI teams

What a Scorecard Actually Is

A scorecard is just a structured rubric — a set of dimensions you evaluate every conversation against, each with defined criteria for what good looks like at each score level.

The key word is structured. A scorecard isn't a 10-point overall rating. It's a breakdown of the distinct things that matter about agent quality, evaluated separately so you can see exactly where things are strong and where they're breaking down.

A well-designed scorecard for a customer service agent might look like this:

DimensionWhat it measures
AccuracyDid the agent provide factually correct information?
CompletenessDid it address all parts of the customer's question?
Policy AdherenceDid it stay within defined business rules?
Resolution QualityDid the customer's issue actually get resolved?
Tone & EmpathyWas the interaction appropriately warm and professional?
Escalation JudgmentDid it correctly identify when to hand off to a human?

Each of these gets its own score — usually on a 1–5 or 0–100 scale with defined rubrics. And here's where the magic is: the dimensions are independent. An agent can score 5/5 on accuracy while scoring 2/5 on completeness. That tells you something specific and actionable.

You can see how Chanl's scorecard system is built around exactly this principle — multi-dimensional evaluation with rubrics that travel with your agent, not locked in someone's spreadsheet.

Designing Rubrics That Actually Work

A rubric is worthless if evaluators (human or automated) can't apply it consistently. Vague criteria produce vague, inconsistent scores. "Was the agent helpful?" is not a rubric — it's a question that ten people will answer ten different ways.

Good rubrics have three properties:

They're behavioral, not evaluative. Instead of "was the agent empathetic?" try "did the agent acknowledge the customer's frustration before offering a solution?" One describes a behavior that can be observed. The other asks for a judgment that varies by the judge.

They have defined score levels. Every score on the scale needs a description. What does a 3/5 on completeness look like vs. a 4/5? Write it out. If you can't describe the difference, your scoring will drift over time.

They're calibrated across evaluators. If you're using human reviewers, run calibration sessions where multiple people score the same conversation and then discuss disagreements. This is how you discover that "policy adherence" means different things to different people — before the inconsistency corrupts your data.

A minimal rubric structure for "Completeness" might look like:

ScoreCriteria
5Addressed all stated and implied questions; proactively covered likely follow-up
4Addressed all stated questions; minor implied question missed
3Addressed primary question; secondary questions partially or not addressed
2Partially addressed primary question; significant gaps in coverage
1Failed to address the primary question or provided irrelevant response

Do this for each dimension. It takes time upfront, but it's the difference between data you can trust and data that drifts.

Quality analyst reviewing scores
Score
Good
0/100
Tone & Empathy
94%
Resolution
88%
Response Time
72%
Compliance
85%

Automating the Grading Layer

Manual scoring at scale is expensive and slow. The good news: LLM-based automated grading has gotten genuinely good at applying structured rubrics to conversation transcripts — good enough to serve as your primary scoring layer for many dimensions, with human review reserved for escalations, edge cases, and calibration.

The mechanics are straightforward. A transcript gets passed to an evaluator LLM along with your rubric and the relevant context — the agent persona, policy docs, whatever it needs to make a fair judgment. The evaluator scores each dimension and outputs a rationale. Scores get logged for trend analysis. Anything that scores below threshold, or where the evaluator signals low confidence, routes to a human reviewer.

The key design decision is: what does your evaluator need to know to score accurately? If you're scoring "policy adherence," the evaluator needs to know what the policy actually says. If you're scoring "escalation judgment," it needs to know what your escalation criteria are.

This is why evaluation and prompt management are so tightly linked — your rubric references the same policy docs and guidelines that the agent itself uses. Keeping those in sync is critical. If your agent is updated with new refund rules but your evaluator's rubric still references the old ones, you'll get garbage scores.

The Chanl analytics platform connects these dots automatically — scoring conversations against the same context that generated them, so your evaluation stays calibrated as your agent evolves.

Regression Detection: The Part Everyone Forgets

Here's a scenario. You improve your agent's prompt to make it more concise. Accuracy scores stay flat. Resolution quality improves. Tone scores drop slightly but you figure that's within noise. You ship it.

A week later you notice escalation rates have crept up. You dig into the conversations and realize your new "concise" prompt is cutting off acknowledgment steps that customers relied on to feel heard before the agent jumped to solutions. Empathy scores were the leading indicator — and you dismissed them as noise.

Regression detection is the practice of systematically monitoring your scorecard dimensions over time and alerting when anything moves in a statistically significant direction. It's the difference between catching a problem when it starts drifting and catching it when customers are complaining.

A basic regression monitoring setup needs:

  • Baseline scores established before any change
  • Rolling averages tracked per dimension (not just overall)
  • Alert thresholds — typically when a dimension drops more than X points over Y conversations
  • Change attribution — correlating score movements with specific prompt or model changes

Regressions caught before customer escalation

Vibes-basedScorecard monitoring

Time to surface a quality regression

WeeksDays

Conversations reviewed before regression identified

HundredsDozens

The pattern holds consistently across teams that make this shift. Without monitoring, regression detection depends on someone noticing something feels off — which takes days and requires luck. With structured scoring, you catch it in the first few dozen conversations that score below threshold.

Connecting Scores to Business Outcomes

Scores are a means to an end, not the end itself. The ultimate question is: do your scorecard dimensions actually predict the outcomes you care about — customer satisfaction, resolution rates, escalations, repeat contacts?

If your "resolution quality" dimension doesn't correlate with CSAT or reduced repeat-contact rate, it's either measuring the wrong thing or measuring it wrong. Calibration against real outcomes is how you know your rubric has validity.

This is worth doing explicitly:

  1. Run a sample of scored conversations through a CSAT survey or outcome tracking
  2. Look at correlation between each dimension and the outcome metric
  3. Weight your dimensions accordingly — or redesign the ones that don't predict outcomes

You might discover that "completeness" is a stronger predictor of CSAT than "tone" — or vice versa — for your specific agent and use case. That's valuable. It tells you where to invest improvement effort.

If you're already tracking conversation analytics — things like escalation rate, resolution time, first-contact resolution — you have the outcome data you need. Connecting it to scorecard dimensions closes the loop.

Progress0/7
  • Define 5–8 evaluation dimensions specific to your agent use case
  • Write behavioral rubrics with score level descriptions for each dimension
  • Run calibration sessions with human reviewers before automating
  • Set up automated grading against the same context your agent uses
  • Establish baseline scores before shipping any agent changes
  • Configure alerts for dimension-level score drops, not just overall averages
  • Correlate scorecard dimensions against CSAT or resolution outcomes quarterly

What Good Looks Like: A Practical Example

A contact center team running an AI agent for subscription billing support set up their scorecard in three phases over about six weeks.

Phase 1: Define dimensions. They identified six dimensions based on what actually drove escalations in their call logs: accuracy, completeness, policy adherence, tone, escalation judgment, and a domain-specific dimension they called "billing literacy" — whether the agent could explain invoices clearly without relying on jargon.

Phase 2: Build rubrics and calibrate. They wrote rubric levels for each dimension and ran two calibration sessions where three team members scored the same 20 conversations. Disagreements surfaced two areas where the rubric was ambiguous — both got rewritten before automated grading was turned on.

Phase 3: Automate and monitor. Automated grading was set to run on every conversation. Any conversation scoring below 3/5 on accuracy or policy adherence triggered a Slack alert for human review. Dimension-level weekly averages were tracked in a dashboard.

Within the first month, they caught a regression in billing literacy that was traced to a prompt change that had inadvertently removed a key instruction about how to explain proration. It had only moved the average by 0.3 points — completely invisible without dimension-level monitoring — but correlated with a small uptick in "didn't understand my bill" escalations that had been chalked up to noise.

They found it in 48 hours. Without the scorecard, they'd estimated that kind of issue would take two to three weeks to surface through escalation patterns alone.

Common Failure Modes

Building a scorecard is not hard. Building one that your team actually uses and trusts — that's where most programs quietly die. A few patterns that cause this to happen:

Too many dimensions. Scoring 15 dimensions per conversation is exhausting for human reviewers and creates noise in automated systems. Start with 5–8. Add dimensions when you have a specific reason to track something new.

Dimensions that overlap. If "accuracy" and "policy adherence" both capture factual correctness, your scores will be correlated noise. Make sure each dimension is measuring something distinct.

No action on low scores. If conversations score 2/5 on resolution quality week after week and nothing changes, your team will stop trusting the scorecard. Every alert needs an owner and a response.

Ignoring evaluator confidence. Automated graders should output confidence signals — some conversations are ambiguous. Low-confidence scores should route to human review, not go straight into your trend charts.

Treating scores as absolute truth. A scorecard is a measurement instrument, not a judgment system. If a score seems wrong, investigate why the rubric produced that result — don't just override the score and move on. Rubrics should evolve based on what you learn.

Getting Started Without Building Everything at Once

You don't need a full platform to start measuring agent quality. Here's the minimum viable scorecard:

Pick three dimensions that matter most to your use case. Write a simple 1–5 rubric for each. Have one person score 50 conversations against the rubric. That's your baseline.

Then pick the two conversations with the lowest combined scores and investigate them. That investigation will usually reveal either a real agent problem or an ambiguity in the rubric — both are useful. Fix whichever it is, and run the cycle again.

As you scale, you can layer in automation, alerting, and richer analytics. But the discipline of structured, multi-dimensional evaluation is the foundation — and it starts with a rubric written in a document, not a system.

Platforms like Chanl's testing and scoring infrastructure exist to make the automation and monitoring parts easier to maintain as you grow. But the thinking — what dimensions matter, what good looks like, how scores connect to outcomes — that's work that pays off from day one.


Quality measurement in AI agents is still a relatively young discipline, but the pattern is becoming clear: teams that operate on structured data beat teams operating on vibes, and the gap widens as agent complexity and conversation volume increase. Regressions get caught faster. Prompt experiments produce real signal. Conversations with customers get better on purpose rather than by luck.

The good news is you don't need a research team to get there. You need a rubric, some calibration, and the discipline to look at the numbers every week.

Ready to move beyond vibes?

Chanl's scorecard system gives you multi-dimensional agent evaluation, automated grading, and regression alerting — built to travel with your agent as it evolves.

See How Scorecards Work

Chanl Team

AI Agent Testing Platform

Building the platform for AI agents at Chanl — tools, testing, and observability for customer experience.

Get AI Agent Insights

Subscribe to our newsletter for weekly tips and best practices.