dataset
stringlengths
5
115
config
stringlengths
1
162
split
stringlengths
1
228
num_examples
int64
3
341M
column_name
stringlengths
0
77.9k
null_count
int64
0
62.9M
null_proportion
float64
0
1
min
int64
0
9.25M
max
int64
0
1.07B
mean
float64
0
90.4M
median
float64
0
80.1M
std
float64
0
130M
histogram
dict
partial
bool
2 classes
ceval/ceval-exam
college_economics
val
55
C
0
0
1
44
10.89091
9
8.17556
{ "bin_edges": [ 1, 6, 11, 16, 21, 26, 31, 36, 41, 44 ], "hist": [ 18, 13, 12, 4, 5, 2, 0, 0, 1 ] }
false
ceval/ceval-exam
college_economics
val
55
D
0
0
1
30
10.50909
8
7.20821
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 22, 25, 28, 30 ], "hist": [ 5, 16, 9, 7, 6, 5, 3, 0, 2, 2 ] }
false
ceval/ceval-exam
college_economics
val
55
question
0
0
10
168
39.29091
27
34.53108
{ "bin_edges": [ 10, 26, 42, 58, 74, 90, 106, 122, 138, 154, 168 ], "hist": [ 25, 17, 3, 4, 1, 1, 1, 1, 0, 2 ] }
false
ceval/ceval-exam
college_chemistry
test
224
A
0
0
1
107
12.625
8
14.43782
{ "bin_edges": [ 1, 12, 23, 34, 45, 56, 67, 78, 89, 100, 107 ], "hist": [ 139, 53, 21, 5, 3, 0, 0, 0, 1, 2 ] }
false
ceval/ceval-exam
college_chemistry
test
224
B
0
0
1
136
12.34375
8
13.86943
{ "bin_edges": [ 1, 15, 29, 43, 57, 71, 85, 99, 113, 127, 136 ], "hist": [ 162, 50, 6, 3, 1, 0, 1, 0, 0, 1 ] }
false
ceval/ceval-exam
college_chemistry
test
224
C
0
0
1
96
12.29464
9
12.32056
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 96 ], "hist": [ 128, 64, 17, 8, 2, 2, 1, 1, 0, 1 ] }
false
ceval/ceval-exam
college_chemistry
test
224
D
0
0
1
113
13.39286
8
14.52529
{ "bin_edges": [ 1, 13, 25, 37, 49, 61, 73, 85, 97, 109, 113 ], "hist": [ 144, 46, 21, 8, 2, 0, 0, 2, 0, 1 ] }
false
ceval/ceval-exam
college_chemistry
test
224
question
0
0
7
618
51.97768
29
63.57965
{ "bin_edges": [ 7, 69, 131, 193, 255, 317, 379, 441, 503, 565, 618 ], "hist": [ 183, 22, 11, 6, 1, 0, 0, 0, 0, 1 ] }
false
ceval/ceval-exam
college_programming
test
342
A
0
0
1
64
12.52632
8
12.25382
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 64 ], "hist": [ 161, 77, 45, 20, 19, 7, 5, 5, 2, 1 ] }
false
ceval/ceval-exam
college_programming
test
342
B
0
0
1
80
12.84795
8
12.63613
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 73, 80 ], "hist": [ 172, 69, 47, 25, 18, 2, 7, 1, 0, 1 ] }
false
ceval/ceval-exam
college_programming
test
342
C
0
0
1
67
12.85673
9
11.85271
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 67 ], "hist": [ 156, 63, 56, 32, 15, 12, 2, 4, 0, 2 ] }
false
ceval/ceval-exam
college_programming
test
342
D
0
0
1
70
12.83041
9
11.76731
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 70 ], "hist": [ 150, 83, 42, 28, 23, 9, 1, 4, 1, 1 ] }
false
ceval/ceval-exam
college_programming
test
342
question
0
0
11
570
60.11111
35
73.69593
{ "bin_edges": [ 11, 67, 123, 179, 235, 291, 347, 403, 459, 515, 570 ], "hist": [ 260, 41, 19, 10, 2, 5, 2, 1, 1, 1 ] }
false
ceval/ceval-exam
college_programming
val
37
A
0
0
1
57
12.67568
7
14.44925
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 57 ], "hist": [ 16, 10, 3, 4, 1, 0, 0, 0, 1, 2 ] }
false
ceval/ceval-exam
college_programming
val
37
B
0
0
1
55
12.54054
9
13.56178
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 55 ], "hist": [ 17, 8, 5, 3, 1, 0, 0, 0, 2, 1 ] }
false
ceval/ceval-exam
college_programming
val
37
C
0
0
1
55
12.02703
8
13.32393
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 55 ], "hist": [ 16, 9, 5, 4, 0, 0, 0, 1, 1, 1 ] }
false
ceval/ceval-exam
college_programming
val
37
D
0
0
1
55
11.89189
7
12.75579
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 55 ], "hist": [ 16, 8, 5, 4, 2, 0, 0, 0, 1, 1 ] }
false
ceval/ceval-exam
college_programming
val
37
question
0
0
12
376
67.89189
34
93.3416
{ "bin_edges": [ 12, 49, 86, 123, 160, 197, 234, 271, 308, 345, 376 ], "hist": [ 24, 8, 1, 0, 0, 1, 0, 0, 1, 2 ] }
false
ceval/ceval-exam
college_physics
test
176
A
0
0
1
76
13.80682
10
13.73201
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 73, 76 ], "hist": [ 77, 56, 16, 12, 5, 2, 1, 4, 2, 1 ] }
false
ceval/ceval-exam
college_physics
test
176
B
0
0
1
116
15.14205
11
15.8279
{ "bin_edges": [ 1, 13, 25, 37, 49, 61, 73, 85, 97, 109, 116 ], "hist": [ 106, 43, 14, 3, 5, 3, 1, 0, 0, 1 ] }
false
ceval/ceval-exam
college_physics
test
176
C
0
0
1
83
15.96023
11
15.7289
{ "bin_edges": [ 1, 10, 19, 28, 37, 46, 55, 64, 73, 82, 83 ], "hist": [ 81, 48, 18, 15, 3, 4, 0, 4, 2, 1 ] }
false
ceval/ceval-exam
college_physics
test
176
D
0
0
1
93
15.01705
10
14.36602
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 93 ], "hist": [ 93, 47, 12, 13, 4, 3, 2, 1, 0, 1 ] }
false
ceval/ceval-exam
college_physics
test
176
question
0
0
13
219
68.11932
65
32.88582
{ "bin_edges": [ 13, 34, 55, 76, 97, 118, 139, 160, 181, 202, 219 ], "hist": [ 24, 44, 48, 27, 22, 6, 3, 0, 1, 1 ] }
false
ceval/ceval-exam
computer_network
test
171
A
0
0
1
36
6.74854
5
6.31254
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 36 ], "hist": [ 81, 49, 20, 9, 4, 3, 1, 3, 1 ] }
false
ceval/ceval-exam
computer_network
test
171
B
0
0
1
37
7.17544
4
7.0747
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 37, 37 ], "hist": [ 87, 38, 19, 10, 8, 3, 1, 2, 2, 1 ] }
false
ceval/ceval-exam
computer_network
test
171
C
0
0
1
36
6.9883
5
6.16726
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 36 ], "hist": [ 83, 37, 25, 14, 6, 3, 0, 0, 3 ] }
false
ceval/ceval-exam
computer_network
test
171
D
0
0
1
36
7.36257
5
6.28421
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 36 ], "hist": [ 78, 42, 20, 17, 7, 4, 0, 0, 3 ] }
false
ceval/ceval-exam
computer_network
test
171
question
0
0
9
163
42.72515
31
30.39738
{ "bin_edges": [ 9, 25, 41, 57, 73, 89, 105, 121, 137, 153, 163 ], "hist": [ 48, 63, 23, 10, 11, 5, 4, 4, 2, 1 ] }
false
ceval/ceval-exam
computer_architecture
test
193
A
0
0
1
35
8.93264
6
7.06301
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 35 ], "hist": [ 67, 51, 28, 19, 10, 10, 2, 3, 3 ] }
false
ceval/ceval-exam
computer_architecture
test
193
B
0
0
1
34
8.86528
6
6.94941
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 34 ], "hist": [ 71, 48, 24, 21, 12, 10, 1, 4, 2 ] }
false
ceval/ceval-exam
computer_architecture
test
193
C
0
0
1
35
9.3057
6
7.41724
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 33, 35 ], "hist": [ 64, 54, 29, 11, 14, 12, 3, 4, 2 ] }
false
ceval/ceval-exam
computer_architecture
test
193
D
0
0
1
32
8.94301
6
6.82
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 32 ], "hist": [ 65, 57, 22, 15, 17, 11, 4, 2 ] }
false
ceval/ceval-exam
computer_architecture
test
193
question
0
0
9
127
32.72021
26
20.25191
{ "bin_edges": [ 9, 21, 33, 45, 57, 69, 81, 93, 105, 117, 127 ], "hist": [ 49, 84, 29, 9, 10, 4, 3, 1, 2, 2 ] }
false
ceval/ceval-exam
education_science
test
270
A
0
0
2
52
6.54815
4.5
5.63607
{ "bin_edges": [ 2, 8, 14, 20, 26, 32, 38, 44, 50, 52 ], "hist": [ 207, 36, 18, 7, 0, 0, 1, 0, 1 ] }
false
ceval/ceval-exam
education_science
test
270
B
0
0
2
38
6.68519
5
5.37039
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 38 ], "hist": [ 160, 60, 25, 8, 10, 2, 3, 1, 0, 1 ] }
false
ceval/ceval-exam
education_science
test
270
C
0
0
2
38
6.93333
5
5.86743
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 38 ], "hist": [ 163, 50, 29, 9, 11, 4, 0, 0, 3, 1 ] }
false
ceval/ceval-exam
education_science
test
270
D
0
0
2
38
6.95185
4.5
6.16814
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 38 ], "hist": [ 162, 57, 22, 12, 5, 5, 3, 0, 3, 1 ] }
false
ceval/ceval-exam
education_science
test
270
question
0
0
8
267
34.2037
28
23.82946
{ "bin_edges": [ 8, 34, 60, 86, 112, 138, 164, 190, 216, 242, 267 ], "hist": [ 168, 80, 16, 2, 2, 0, 1, 0, 0, 1 ] }
false
ceval/ceval-exam
discrete_mathematics
test
153
A
0
0
1
112
19.80392
9
23.661
{ "bin_edges": [ 1, 13, 25, 37, 49, 61, 73, 85, 97, 109, 112 ], "hist": [ 88, 19, 21, 8, 4, 6, 2, 1, 3, 1 ] }
false
ceval/ceval-exam
discrete_mathematics
test
153
B
0
0
1
126
21.30065
12
26.00306
{ "bin_edges": [ 1, 14, 27, 40, 53, 66, 79, 92, 105, 118, 126 ], "hist": [ 83, 29, 16, 10, 3, 5, 2, 0, 2, 3 ] }
false
ceval/ceval-exam
discrete_mathematics
test
153
C
0
0
1
147
22.04575
12
26.95399
{ "bin_edges": [ 1, 16, 31, 46, 61, 76, 91, 106, 121, 136, 147 ], "hist": [ 90, 26, 14, 11, 3, 3, 2, 2, 1, 1 ] }
false
ceval/ceval-exam
discrete_mathematics
test
153
D
0
0
1
197
22.15686
12
28.69603
{ "bin_edges": [ 1, 21, 41, 61, 81, 101, 121, 141, 161, 181, 197 ], "hist": [ 100, 28, 13, 3, 5, 2, 1, 0, 0, 1 ] }
false
ceval/ceval-exam
discrete_mathematics
test
153
question
0
0
10
291
47.29412
36
43.5281
{ "bin_edges": [ 10, 39, 68, 97, 126, 155, 184, 213, 242, 271, 291 ], "hist": [ 85, 37, 17, 6, 3, 2, 0, 2, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
test
339
A
0
0
1
61
7.64012
5
8.53536
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 61 ], "hist": [ 236, 68, 12, 9, 6, 4, 2, 1, 1 ] }
false
ceval/ceval-exam
electrical_engineer
test
339
B
0
0
1
55
7.71386
5
7.9932
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 55 ], "hist": [ 212, 82, 19, 7, 7, 6, 4, 1, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
test
339
C
0
0
1
61
7.75516
5
7.85382
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 61 ], "hist": [ 235, 63, 18, 11, 7, 3, 1, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
test
339
D
0
0
1
59
7.9233
5
7.89989
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 59 ], "hist": [ 199, 88, 26, 11, 5, 4, 2, 3, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
test
339
question
0
0
13
137
44.87021
41
21.49008
{ "bin_edges": [ 13, 26, 39, 52, 65, 78, 91, 104, 117, 130, 137 ], "hist": [ 54, 97, 89, 47, 23, 14, 9, 2, 3, 1 ] }
false
ceval/ceval-exam
electrical_engineer
val
37
A
0
0
1
25
6.2973
5
5.17937
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 22, 25, 25 ], "hist": [ 10, 16, 5, 2, 2, 0, 0, 1, 1 ] }
false
ceval/ceval-exam
electrical_engineer
val
37
B
0
0
1
48
7.27027
5
8.2045
{ "bin_edges": [ 1, 6, 11, 16, 21, 26, 31, 36, 41, 46, 48 ], "hist": [ 22, 10, 2, 1, 1, 0, 0, 0, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
val
37
C
0
0
1
52
8.18919
5
10.06876
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 52 ], "hist": [ 22, 11, 1, 0, 1, 0, 1, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
val
37
D
0
0
1
53
8.32432
5
10.1326
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 53 ], "hist": [ 23, 10, 1, 0, 1, 0, 1, 0, 1 ] }
false
ceval/ceval-exam
electrical_engineer
val
37
question
0
0
14
92
44.72973
36
22.21692
{ "bin_edges": [ 14, 22, 30, 38, 46, 54, 62, 70, 78, 86, 92 ], "hist": [ 3, 10, 6, 3, 2, 3, 4, 2, 2, 2 ] }
false
ceval/ceval-exam
fire_engineer
test
282
A
0
0
1
66
11.60638
5
13.62948
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 66 ], "hist": [ 170, 28, 30, 22, 8, 10, 5, 6, 2, 1 ] }
false
ceval/ceval-exam
fire_engineer
test
282
B
0
0
1
65
11.53546
5
13.11408
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 65 ], "hist": [ 170, 32, 23, 18, 13, 16, 6, 3, 0, 1 ] }
false
ceval/ceval-exam
fire_engineer
test
282
C
0
0
1
77
12.10284
5
14.68095
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 73, 77 ], "hist": [ 177, 36, 16, 22, 15, 5, 7, 1, 2, 1 ] }
false
ceval/ceval-exam
fire_engineer
test
282
D
0
0
1
66
11.6383
5
13.17783
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 66 ], "hist": [ 169, 31, 28, 17, 15, 13, 4, 1, 3, 1 ] }
false
ceval/ceval-exam
fire_engineer
test
282
question
0
0
16
127
50.85816
46.5
23.89213
{ "bin_edges": [ 16, 28, 40, 52, 64, 76, 88, 100, 112, 124, 127 ], "hist": [ 56, 52, 48, 42, 40, 22, 11, 7, 3, 1 ] }
false
ceval/ceval-exam
fire_engineer
val
31
A
0
0
1
71
12
6
14.30618
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 71 ], "hist": [ 17, 5, 7, 0, 1, 0, 0, 0, 1 ] }
false
ceval/ceval-exam
fire_engineer
val
31
B
0
0
1
66
15.96774
8
18.96397
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 66 ], "hist": [ 15, 6, 2, 2, 0, 2, 2, 0, 0, 2 ] }
false
ceval/ceval-exam
fire_engineer
val
31
C
0
0
1
64
15.32258
7
17.66803
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 64 ], "hist": [ 16, 4, 2, 4, 1, 1, 0, 1, 1, 1 ] }
false
ceval/ceval-exam
fire_engineer
val
31
D
0
0
1
74
15.09677
9
17.15101
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 73, 74 ], "hist": [ 15, 7, 1, 4, 2, 1, 0, 0, 0, 1 ] }
false
ceval/ceval-exam
fire_engineer
val
31
question
0
0
23
105
48.41935
41
24.033
{ "bin_edges": [ 23, 32, 41, 50, 59, 68, 77, 86, 95, 104, 105 ], "hist": [ 10, 5, 6, 1, 2, 2, 2, 1, 1, 1 ] }
false
ceval/ceval-exam
high_school_chinese
test
178
A
0
0
2
110
40.69101
42
22.61043
{ "bin_edges": [ 2, 13, 24, 35, 46, 57, 68, 79, 90, 101, 110 ], "hist": [ 28, 14, 19, 43, 29, 26, 12, 3, 3, 1 ] }
false
ceval/ceval-exam
high_school_chinese
test
178
B
0
0
2
93
39.5618
41
21.04231
{ "bin_edges": [ 2, 12, 22, 32, 42, 52, 62, 72, 82, 92, 93 ], "hist": [ 28, 10, 16, 37, 30, 29, 22, 4, 1, 1 ] }
false
ceval/ceval-exam
high_school_chinese
test
178
C
0
0
2
89
39.27528
40
21.50568
{ "bin_edges": [ 2, 11, 20, 29, 38, 47, 56, 65, 74, 83, 89 ], "hist": [ 26, 11, 16, 26, 19, 35, 24, 13, 5, 3 ] }
false
ceval/ceval-exam
high_school_chinese
test
178
D
0
0
2
93
39.92697
43.5
21.388
{ "bin_edges": [ 2, 12, 22, 32, 42, 52, 62, 72, 82, 92, 93 ], "hist": [ 25, 13, 24, 23, 33, 36, 14, 7, 2, 1 ] }
false
ceval/ceval-exam
high_school_chinese
test
178
question
0
0
13
281
50.34831
19
56.91883
{ "bin_edges": [ 13, 40, 67, 94, 121, 148, 175, 202, 229, 256, 281 ], "hist": [ 125, 6, 16, 9, 4, 9, 4, 2, 1, 2 ] }
false
ceval/ceval-exam
high_school_biology
test
175
A
0
0
2
49
13.22286
10
9.79746
{ "bin_edges": [ 2, 7, 12, 17, 22, 27, 32, 37, 42, 47, 49 ], "hist": [ 60, 31, 27, 17, 20, 8, 9, 2, 0, 1 ] }
false
ceval/ceval-exam
high_school_biology
test
175
B
0
0
2
40
13.96
11
10.20613
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 40 ], "hist": [ 49, 30, 20, 14, 18, 13, 15, 9, 4, 3 ] }
false
ceval/ceval-exam
high_school_biology
test
175
C
0
0
2
69
14.18286
11
11.18603
{ "bin_edges": [ 2, 9, 16, 23, 30, 37, 44, 51, 58, 65, 69 ], "hist": [ 72, 36, 22, 27, 14, 2, 1, 0, 0, 1 ] }
false
ceval/ceval-exam
high_school_biology
test
175
D
0
0
1
41
14.06857
11
10.49279
{ "bin_edges": [ 1, 6, 11, 16, 21, 26, 31, 36, 41, 41 ], "hist": [ 50, 37, 15, 27, 13, 20, 7, 4, 2 ] }
false
ceval/ceval-exam
high_school_biology
test
175
question
0
0
11
215
49.73714
34
39.91629
{ "bin_edges": [ 11, 32, 53, 74, 95, 116, 137, 158, 179, 200, 215 ], "hist": [ 83, 38, 17, 15, 9, 4, 3, 3, 1, 2 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
test
281
A
0
0
1
55
9.879
6
9.6551
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 55 ], "hist": [ 150, 62, 34, 13, 3, 7, 8, 3, 0, 1 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
test
281
B
0
0
1
95
10.77936
6
12.12028
{ "bin_edges": [ 1, 11, 21, 31, 41, 51, 61, 71, 81, 91, 95 ], "hist": [ 192, 49, 20, 11, 4, 2, 1, 1, 0, 1 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
test
281
C
0
0
1
53
10.84342
7
10.41275
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 53 ], "hist": [ 130, 72, 34, 18, 8, 8, 4, 2, 5 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
test
281
D
0
0
1
124
11.29181
7
12.92313
{ "bin_edges": [ 1, 14, 27, 40, 53, 66, 79, 92, 105, 118, 124 ], "hist": [ 204, 54, 14, 4, 2, 2, 0, 0, 0, 1 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
test
281
question
0
0
13
150
54.00356
51
24.50925
{ "bin_edges": [ 13, 27, 41, 55, 69, 83, 97, 111, 125, 139, 150 ], "hist": [ 43, 40, 74, 51, 44, 13, 9, 4, 2, 1 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
val
31
A
0
0
1
30
8.32258
6
7.70016
{ "bin_edges": [ 1, 4, 7, 10, 13, 16, 19, 22, 25, 28, 30 ], "hist": [ 8, 10, 6, 0, 3, 1, 0, 0, 1, 2 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
val
31
B
0
0
1
31
8.83871
7
7.54143
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 31 ], "hist": [ 10, 8, 8, 2, 0, 1, 0, 2 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
val
31
C
0
0
1
31
9.3871
7
8.47615
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 31 ], "hist": [ 12, 8, 3, 2, 3, 0, 1, 2 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
val
31
D
0
0
1
31
9.16129
6
8.43839
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 31 ], "hist": [ 11, 9, 4, 2, 1, 1, 2, 1 ] }
false
ceval/ceval-exam
environmental_impact_assessment_engineer
val
31
question
0
0
23
108
58.25806
58
23.5074
{ "bin_edges": [ 23, 32, 41, 50, 59, 68, 77, 86, 95, 104, 108 ], "hist": [ 4, 3, 4, 5, 8, 0, 2, 2, 2, 1 ] }
false
ceval/ceval-exam
high_school_geography
test
178
A
0
0
1
31
8.91011
7
6.43725
{ "bin_edges": [ 1, 5, 9, 13, 17, 21, 25, 29, 31 ], "hist": [ 54, 48, 31, 22, 15, 2, 1, 5 ] }
false
ceval/ceval-exam
high_school_geography
test
178
B
0
0
2
38
9.58427
8
6.78118
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 38, 38 ], "hist": [ 65, 35, 32, 22, 12, 8, 2, 1, 0, 1 ] }
false
ceval/ceval-exam
high_school_geography
test
178
C
0
0
2
32
9.50562
7
6.84307
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 32 ], "hist": [ 61, 45, 30, 15, 13, 10, 1, 3 ] }
false
ceval/ceval-exam
high_school_geography
test
178
D
0
0
2
36
9.9382
7
7.45166
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 36 ], "hist": [ 59, 47, 26, 18, 11, 9, 4, 2, 2 ] }
false
ceval/ceval-exam
high_school_geography
test
178
question
0
0
11
167
32.29213
23
29.01517
{ "bin_edges": [ 11, 27, 43, 59, 75, 91, 107, 123, 139, 155, 167 ], "hist": [ 120, 35, 2, 4, 5, 5, 2, 0, 3, 2 ] }
false
ceval/ceval-exam
high_school_history
test
182
A
0
0
2
32
8.71429
9
4.49379
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 32 ], "hist": [ 50, 55, 57, 15, 3, 0, 1, 1 ] }
false
ceval/ceval-exam
high_school_history
test
182
B
0
0
2
34
8.81319
8
4.73826
{ "bin_edges": [ 2, 6, 10, 14, 18, 22, 26, 30, 34, 34 ], "hist": [ 47, 57, 54, 18, 4, 1, 0, 0, 1 ] }
false
ceval/ceval-exam
high_school_history
test
182
C
0
0
2
22
8.86813
9
4.36639
{ "bin_edges": [ 2, 5, 8, 11, 14, 17, 20, 22 ], "hist": [ 38, 30, 56, 34, 15, 5, 4 ] }
false
ceval/ceval-exam
high_school_history
test
182
D
0
0
2
27
9.19231
9
4.74561
{ "bin_edges": [ 2, 5, 8, 11, 14, 17, 20, 23, 26, 27 ], "hist": [ 38, 30, 45, 41, 15, 8, 3, 1, 1 ] }
false
ceval/ceval-exam
high_school_history
test
182
question
0
0
14
140
62.46154
60
33.81549
{ "bin_edges": [ 14, 27, 40, 53, 66, 79, 92, 105, 118, 131, 140 ], "hist": [ 32, 31, 19, 17, 16, 24, 24, 8, 5, 6 ] }
false
ceval/ceval-exam
high_school_chemistry
test
172
A
0
0
1
62
14.15698
11
12.16475
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 62 ], "hist": [ 63, 40, 33, 19, 5, 3, 5, 2, 2 ] }
false
ceval/ceval-exam
high_school_chemistry
test
172
B
0
0
1
72
15.5
13
12.97524
{ "bin_edges": [ 1, 9, 17, 25, 33, 41, 49, 57, 65, 72 ], "hist": [ 64, 44, 29, 14, 13, 5, 0, 1, 2 ] }
false
ceval/ceval-exam
high_school_chemistry
test
172
C
0
0
1
65
16.15698
13
13.44922
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 64, 65 ], "hist": [ 62, 34, 24, 24, 11, 10, 3, 1, 2, 1 ] }
false
ceval/ceval-exam
high_school_chemistry
test
172
D
0
0
1
110
16.68023
13
15.55549
{ "bin_edges": [ 1, 12, 23, 34, 45, 56, 67, 78, 89, 100, 110 ], "hist": [ 81, 44, 31, 8, 3, 2, 2, 0, 0, 1 ] }
false
ceval/ceval-exam
high_school_chemistry
test
172
question
0
0
11
267
46.6686
28.5
42.76565
{ "bin_edges": [ 11, 37, 63, 89, 115, 141, 167, 193, 219, 245, 267 ], "hist": [ 105, 26, 16, 11, 7, 3, 1, 0, 2, 1 ] }
false
ceval/ceval-exam
high_school_mathematics
test
166
A
0
0
1
60
11.83735
8
11.75884
{ "bin_edges": [ 1, 7, 13, 19, 25, 31, 37, 43, 49, 55, 60 ], "hist": [ 72, 27, 27, 19, 9, 5, 3, 0, 2, 2 ] }
false
ceval/ceval-exam
high_school_mathematics
test
166
B
0
0
1
61
13.66867
10
12.73144
{ "bin_edges": [ 1, 8, 15, 22, 29, 36, 43, 50, 57, 61 ], "hist": [ 71, 34, 25, 17, 8, 5, 1, 4, 1 ] }
false