Skip to content

Prompt Quality Evaluator

Run the Prompt Quality Evaluator Fullscreen

This MicroSim lets you practice evaluating AI prompts and responses against four quality dimensions:

  1. Relevance — Does the response address what was actually asked?
  2. Accuracy — Is the information factually correct?
  3. Completeness — Does the response cover the topic adequately?
  4. Conciseness — Is the response appropriately brief without unnecessary padding?

How to Use

  1. Read the sample prompt and its AI-generated response
  2. Use the four sliders to rate each quality dimension from 1 to 10
  3. Click Check My Ratings to see how your scores compare to the expert ratings
  4. Green means you're close (within 2 points), yellow means you're in the ballpark (3-4 points off), and red means you're far from the expert (5+ points off)
  5. Click Next Prompt to move to the next example

Lesson Plan

This activity supports Bloom's Evaluate level by asking students to make judgments about prompt and response quality. The calibration exercise — rating then comparing to experts — helps students internalize the four quality dimensions through active practice rather than passive reading.

Suggested classroom use:

  • Have students work through all 5 examples individually
  • Discuss as a class which dimensions were hardest to rate accurately
  • Ask students to write their own prompt/response pairs that score low on a specific dimension