Skip to content

Text-only playground for evaluating reasoning model outputs with mock accuracy, hallucination, and trust metrics — runs 100% locally.

License

Notifications You must be signed in to change notification settings

iltutishrak/eval-metrics-lab

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

1 Commit
 
 
 
 
 
 
 
 

Repository files navigation

eval-metrics-lab

A simple, text-only playground for evaluating reasoning model outputs.
Local, lightweight, and perfect for PMs exploring AI reliability.

Run it

python eval_metrics_lab.py

Example Output

Evaluating 3 mock responses...
Accuracy: 0.87
Hallucination: 0.11
Trust Score: 76.0
✅ Model reliability acceptable.

About

Text-only playground for evaluating reasoning model outputs with mock accuracy, hallucination, and trust metrics — runs 100% locally.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages