This skill should be used when the user asks to "implement LLM-as-judge", "compare model outputs", "create evaluation rubrics", "mitigate evaluation bias", or mentions direct scoring, pairwise comparison, position bias, evaluation pipelines, or automated quality assessment.
Activate this skill when:
AI-powered evaluation of trust, security posture, quality signals, and fit for your use case. Grounded in the skill's actual data.
Detected signals: missing license.
• No license declared. Usage rights are ambiguous — contact the skill author before using commercially.
Compare this skill side-by-side with an alternative to find the best fit.
Compare with another skill →