openai / evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
GitHub Repo 
42K
lines of main code
1.5K files
2.5K
lines of test code
50 files
5.8K
lines of other code
107 files
2y
age
783 days
<1%
main code touched
1 year (284 LOC)
<1%
new main code
1 year (31 LOC)
30K
py
8.2K
yaml
JSONL
1.9K
jsonl
1.3K
ipynb
0.3K
html
0.1K
js
0.06K
toml
IN
0.01K
in

0

47

636

0

19

436

2025 2024 2023

generated by sokrates.dev (configuration) on 2025-05-04