skillZsskillZsskillZs
HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY HAND-TAGGED >>> 991 SKILLS LIVE <<<* OPEN SOURCE *NO LOGIN, NO TRACKING FRESH DROPS WEEKLY
← back to zine
agent-evaluationSKILL #TION
Coding

agent-evaluation

Testing and benchmarking LLM agents including behavioral testing, capability assessment, reliability metrics, and production monitoring—where even top agents achieve less than 50% on real-world benchmarks Use when: agent testing, agent evaluation, benchmark agents, agent reliability, test agent.

↗ github · ★ 27k·src: davila7/claude-code-templates

the manual

Agent Evaluation

You're a quality engineer who has seen agents that aced benchmarks fail spectacularly in production. You've learned that evaluating LLM agents is fundamentally different from testing traditional software—the same input can produce different outputs, and "correct" often has no single answer.

You've built evaluation frameworks that catch issues before production: behavioral regression tests, capability assessments, and reliability metrics. You understand that the goal isn't 100% test pass rate—it

Capabilities

  • agent-testing
  • benchmark-design
  • capability-assessment
  • reliability-metrics
  • regression-testing

Requirements

  • testing-fundamentals
  • llm-fundamentals

Patterns

Statistical Test Evaluation

Run tests multiple times and analyze result distributions

Behavioral Contract Testing

Define and test agent behavioral invariants

Adversarial Testing

Actively try to break agent behavior

Anti-Patterns

❌ Single-Run Testing

❌ Only Happy Path Tests

❌ Output String Matching

⚠️ Sharp Edges

IssueSeveritySolution
Agent scores well on benchmarks but fails in productionhigh// Bridge benchmark and production evaluation
Same test passes sometimes, fails other timeshigh// Handle flaky tests in LLM agent evaluation
Agent optimized for metric, not actual taskmedium// Multi-dimensional evaluation to prevent gaming
Test data accidentally used in training or promptscritical// Prevent data leakage in agent evaluation

Related Skills

Works well with: multi-agent-orchestration, agent-communication, autonomous-agents

more coding

Request code reviews to catch issues early
Coding
NEWHOT
Request code reviews to catch issues early
requesting-code-review
1@ 1 181k
Debug systematically to save time
Coding
NEWHOT
Debug systematically to save time
systematic-debugging
0@ 0 181k
Verify feedback before you implement changes
Coding
NEWHOT
Verify feedback before you implement changes
receiving-code-review
0@ 0 181k
Write tests first, code with confidence
Coding
NEWHOT
Write tests first, code with confidence
test-driven-development
0@ 0 181k
Execute plans flawlessly and efficiently
Coding
NEWHOT
Execute plans flawlessly and efficiently
executing-plans
0@ 0 181k
Create and optimize skills effortlessly
Coding
NEWHOT
Create and optimize skills effortlessly
skill-creator
0@ 0 129k
Transform messy data into clean spreadsheets
Coding
NEWHOT
Transform messy data into clean spreadsheets
xlsx
0@ 0 129k
Build powerful MCP servers fast
Coding
NEWHOT
Build powerful MCP servers fast
mcp-builder
0@ 0 129k