M

Micro Edu Evals

A web tool for comparing AI model responses on educational micro-tasks. Built for educators, developers, and policymakers to evaluate different AI models on real-world education scenarios.

For Educators

Test pedagogical reasoning, tone adaptation, and content simplification.

  • Create a rubric for Grade 5 History
  • Differentiate text for dyslexia
  • Generate misconception-based quiz questions

For Developers

Test structured data extraction, code generation, and API reliability.

  • Extract JSON from a syllabus PDF
  • Write a Python unit test for grading logic
  • Convert natural language to SQL queries

For Policymakers

Test safety guardrails, bias detection, and compliance drafting.

  • Check PII handling in student data
  • Draft a parental consent form (COPPA)
  • Audit text for gender bias in career advice