judge
Quality Assurance
Pattern: Meta-Judge → LLM-as-Judge with Context Isolation
Phase 1: Context Extraction
Review conversation history
Identify work to evaluate
Extract: Original task, output, files, constraints, artifact type
│
Phase 2: Meta-Judge (sadd:meta-judge)
┌─────────────────────────────────────────┐
│ Receives extracted context + artifact │
│ type + evaluation focus │
│ │
│ Generates evaluation specification YAML: │
│ - Tailored rubrics per artifact type │
│ - Checklists │
│ - Scoring criteria and weights │
└─────────────────────────────────────────┘
│
Phase 3: Judge Sub-Agent (sadd:judge, Fresh Context)
┌─────────────────────────────────────────┐
│ Receives ONLY extracted context │
│ + exact meta-judge specification YAML │
│ (prevents confirmation bias) │
│ │
│ For each criterion from meta-judge spec: │
│ 1. Review evidence │
│ 2. Write justification │
│ 3. Assign score (1-5) │
│ 4. Self-verify with questions │
│ 5. Adjust if needed │
└─────────────────────────────────────────┘
│
Phase 4: Validation & Report
Verify scores in valid range (1-5)
Check justification has evidence
Confirm weighted total calculation
Present verdict with recommendationsUsage
When to Use
Scoring Interpretation
Score Range
Verdict
Recommendation
Quality Enhancement Techniques
Technique
Benefit
Theoretical Foundation
Last updated