·write-judge-prompt
</>

write-judge-prompt

Design LLM-as-Judge evaluators for subjective criteria that code-based checks cannot handle. Use when a failure mode requires interpretation (tone, faithfulness, relevance, completeness). Do NOT use when the failure mode can be checked with code (regex, schema validation, execution tests). Do NOT use when you need to validate or calibrate the judge — use validate-evaluator instead.

76Installs·3Trend·@hamelsmu

Installation

$npx skills add https://github.com/hamelsmu/evals-skills --skill write-judge-prompt

How to Install write-judge-prompt

Quickly install write-judge-prompt AI skill to your development environment via command line

  1. Open Terminal: Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.)
  2. Run Installation Command: Copy and run this command: npx skills add https://github.com/hamelsmu/evals-skills --skill write-judge-prompt
  3. Verify Installation: Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Source: hamelsmu/evals-skills.

SKILL.md

View raw

Design a binary Pass/Fail LLM-as-Judge evaluator for one specific failure mode. Each judge checks exactly one thing.

State what the judge evaluates. One failure mode per judge.

Not: "Evaluate whether the email is good" or "Rate the email quality from 1-5."

Design LLM-as-Judge evaluators for subjective criteria that code-based checks cannot handle. Use when a failure mode requires interpretation (tone, faithfulness, relevance, completeness). Do NOT use when the failure mode can be checked with code (regex, schema validation, execution tests). Do NOT use when you need to validate or calibrate the judge — use validate-evaluator instead. Source: hamelsmu/evals-skills.

Facts (cite-ready)

Stable fields and commands for AI/search citations.

Install command
npx skills add https://github.com/hamelsmu/evals-skills --skill write-judge-prompt
Category
</>Dev Tools
Verified
First Seen
2026-03-04
Updated
2026-03-10

Browse more skills from hamelsmu/evals-skills

Quick answers

What is write-judge-prompt?

Design LLM-as-Judge evaluators for subjective criteria that code-based checks cannot handle. Use when a failure mode requires interpretation (tone, faithfulness, relevance, completeness). Do NOT use when the failure mode can be checked with code (regex, schema validation, execution tests). Do NOT use when you need to validate or calibrate the judge — use validate-evaluator instead. Source: hamelsmu/evals-skills.

How do I install write-judge-prompt?

Open your terminal or command line tool (Terminal, iTerm, Windows Terminal, etc.) Copy and run this command: npx skills add https://github.com/hamelsmu/evals-skills --skill write-judge-prompt Once installed, the skill will be automatically configured in your AI coding environment and ready to use in Claude Code, Cursor, or OpenClaw

Where is the source repository?

https://github.com/hamelsmu/evals-skills