openai / simple-evals
File Size

The distribution of size of files (measured in lines of code).

Intro
Learn more...
File Size Overall
0% | 0% | 28% | 35% | 35%
Legend:
1001+
501-1000
201-500
101-200
1-100


explore: grouped by folders | grouped by size | sunburst | 3D view
File Size per Extension
1001+
501-1000
201-500
101-200
1-100
py0% | 0% | 28% | 35% | 35%
File Size per Logical Decomposition
primary
1001+
501-1000
201-500
101-200
1-100
ROOT0% | 0% | 34% | 42% | 22%
sampler0% | 0% | 0% | 0% | 100%
Longest Files (Top 15)
File# lines# units
224 1
220 20
common.py
in root
168 11
144 6
134 1
113 2
86 3
83 5
71 5
68 5
64 5
61 2
52 5
47 2
types.py
in root
22 2
Files With Most Units (Top 15)
File# lines# units
220 20
common.py
in root
168 11
144 6
83 5
68 5
52 5
71 5
64 5
86 3
types.py
in root
22 2
61 2
113 2
47 2
134 1
224 1
Files With Long Lines (Top 1)

There is only one file with lines longer than 120 characters. In total, there is only one long line.

File# lines# units# long lines
86 3 1
Correlations

File Size vs. Commits (all time): 15 points

sampler/responses_sampler.py x: 7 commits (all time) y: 71 lines of code simple_evals.py x: 22 commits (all time) y: 224 lines of code mgsm_eval.py x: 6 commits (all time) y: 144 lines of code browsecomp_eval.py x: 6 commits (all time) y: 83 lines of code simpleqa_eval.py x: 10 commits (all time) y: 86 lines of code run_multilingual_mmlu.py x: 7 commits (all time) y: 134 lines of code common.py x: 9 commits (all time) y: 168 lines of code drop_eval.py x: 3 commits (all time) y: 220 lines of code gpqa_eval.py x: 7 commits (all time) y: 61 lines of code math_eval.py x: 6 commits (all time) y: 47 lines of code mmlu_eval.py x: 5 commits (all time) y: 113 lines of code sampler/o_chat_completion_sampler.py x: 2 commits (all time) y: 52 lines of code sampler/chat_completion_sampler.py x: 1 commits (all time) y: 64 lines of code sampler/claude_sampler.py x: 1 commits (all time) y: 68 lines of code types.py x: 1 commits (all time) y: 22 lines of code
224.0
lines of code
  min: 22.0
  average: 103.8
  25th percentile: 61.0
  median: 83.0
  75th percentile: 144.0
  max: 224.0
0 22.0
commits (all time)
min: 1.0 | average: 6.2 | 25th percentile: 2.0 | median: 6.0 | 75th percentile: 7.0 | max: 22.0

File Size vs. Contributors (all time): 15 points

sampler/responses_sampler.py x: 2 contributors (all time) y: 71 lines of code simple_evals.py x: 6 contributors (all time) y: 224 lines of code mgsm_eval.py x: 5 contributors (all time) y: 144 lines of code browsecomp_eval.py x: 2 contributors (all time) y: 83 lines of code simpleqa_eval.py x: 5 contributors (all time) y: 86 lines of code run_multilingual_mmlu.py x: 3 contributors (all time) y: 134 lines of code common.py x: 5 contributors (all time) y: 168 lines of code drop_eval.py x: 3 contributors (all time) y: 220 lines of code gpqa_eval.py x: 6 contributors (all time) y: 61 lines of code math_eval.py x: 5 contributors (all time) y: 47 lines of code mmlu_eval.py x: 4 contributors (all time) y: 113 lines of code sampler/o_chat_completion_sampler.py x: 2 contributors (all time) y: 52 lines of code sampler/chat_completion_sampler.py x: 1 contributors (all time) y: 64 lines of code sampler/claude_sampler.py x: 1 contributors (all time) y: 68 lines of code types.py x: 1 contributors (all time) y: 22 lines of code
224.0
lines of code
  min: 22.0
  average: 103.8
  25th percentile: 61.0
  median: 83.0
  75th percentile: 144.0
  max: 224.0
0 6.0
contributors (all time)
min: 1.0 | average: 3.4 | 25th percentile: 2.0 | median: 3.0 | 75th percentile: 5.0 | max: 6.0

File Size vs. Commits (30 days): 5 points

sampler/responses_sampler.py x: 7 commits (30d) y: 71 lines of code simple_evals.py x: 8 commits (30d) y: 224 lines of code mgsm_eval.py x: 3 commits (30d) y: 144 lines of code browsecomp_eval.py x: 6 commits (30d) y: 83 lines of code simpleqa_eval.py x: 2 commits (30d) y: 86 lines of code
224.0
lines of code
  min: 71.0
  average: 121.6
  25th percentile: 77.0
  median: 86.0
  75th percentile: 184.0
  max: 224.0
0 8.0
commits (30d)
min: 2.0 | average: 5.2 | 25th percentile: 2.5 | median: 6.0 | 75th percentile: 7.5 | max: 8.0

File Size vs. Contributors (30 days): 5 points

sampler/responses_sampler.py x: 2 contributors (30d) y: 71 lines of code simple_evals.py x: 4 contributors (30d) y: 224 lines of code mgsm_eval.py x: 2 contributors (30d) y: 144 lines of code browsecomp_eval.py x: 2 contributors (30d) y: 83 lines of code simpleqa_eval.py x: 2 contributors (30d) y: 86 lines of code
224.0
lines of code
  min: 71.0
  average: 121.6
  25th percentile: 77.0
  median: 86.0
  75th percentile: 184.0
  max: 224.0
0 4.0
contributors (30d)
min: 2.0 | average: 2.4 | 25th percentile: 2.0 | median: 2.0 | 75th percentile: 3.0 | max: 4.0

File Size vs. Commits (90 days): 5 points

sampler/responses_sampler.py x: 7 commits (90d) y: 71 lines of code simple_evals.py x: 10 commits (90d) y: 224 lines of code mgsm_eval.py x: 3 commits (90d) y: 144 lines of code browsecomp_eval.py x: 6 commits (90d) y: 83 lines of code simpleqa_eval.py x: 2 commits (90d) y: 86 lines of code
224.0
lines of code
  min: 71.0
  average: 121.6
  25th percentile: 77.0
  median: 86.0
  75th percentile: 184.0
  max: 224.0
0 10.0
commits (90d)
min: 2.0 | average: 5.6 | 25th percentile: 2.5 | median: 6.0 | 75th percentile: 8.5 | max: 10.0

File Size vs. Contributors (90 days): 5 points

sampler/responses_sampler.py x: 2 contributors (90d) y: 71 lines of code simple_evals.py x: 5 contributors (90d) y: 224 lines of code mgsm_eval.py x: 2 contributors (90d) y: 144 lines of code browsecomp_eval.py x: 2 contributors (90d) y: 83 lines of code simpleqa_eval.py x: 2 contributors (90d) y: 86 lines of code
224.0
lines of code
  min: 71.0
  average: 121.6
  25th percentile: 77.0
  median: 86.0
  75th percentile: 184.0
  max: 224.0
0 5.0
contributors (90d)
min: 2.0 | average: 2.6 | 25th percentile: 2.0 | median: 2.0 | 75th percentile: 3.5 | max: 5.0