LLM Response Grader

Grade model responses against custom weighted rubric rules and detect banned-term violations.

Response score

0.0/100 (grade F)

Base 0.0 minus penalties for banned terms.

Rubric checks

Add rubric rules to start grading.

About This Tool

LLM Response Grader applies weighted rubric checks to evaluate response quality consistently. It is useful for prompt iteration, regression checks, and human-in-the-loop QA.

Frequently Asked Questions

Can rules use regex?

Yes. Use syntax like /pattern/i in the third column.

Is score objective?

It is objective relative to your rubric definition, not universal quality truth.

Is response data uploaded?

No. Grading runs locally in your browser.