ror HF Staff commited on
Commit
c03347d
·
1 Parent(s): 6f570d6

Sample data overhaul

Browse files
Files changed (4) hide show
  1. data.py +26 -4
  2. sample_amd.json +1709 -0
  3. sample_data.csv +0 -22
  4. sample_nvidia.json +1776 -0
data.py CHANGED
@@ -77,10 +77,32 @@ def get_distant_data() -> pd.DataFrame:
77
 
78
 
79
  def get_sample_data() -> pd.DataFrame:
80
- path = os.path.join(os.path.dirname(__file__), "sample_data.csv")
81
- df = pd.read_csv(path)
82
- df = df.set_index("model_name")
83
- return df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
84
 
85
  def extract_model_data(row: pd.Series) -> tuple[dict[str, int], dict[str, int], int, int, int, int]:
86
  """Extract and process model data from DataFrame row."""
 
77
 
78
 
79
  def get_sample_data() -> pd.DataFrame:
80
+ # Retrieve sample dataframes
81
+ df_amd = read_one_dataframe("sample_amd.json", "amd")
82
+ df_nvidia = read_one_dataframe("sample_nvidia.json", "nvidia")
83
+ # Join both dataframes
84
+ joined = df_amd.join(df_nvidia, rsuffix="_nvidia", lsuffix="_amd", how="outer")
85
+ joined = joined[
86
+ [
87
+ "success_amd",
88
+ "success_nvidia",
89
+ "failed_multi_no_amd",
90
+ "failed_multi_no_nvidia",
91
+ "failed_single_no_amd",
92
+ "failed_single_no_nvidia",
93
+ "failures_amd",
94
+ "failures_nvidia",
95
+ "job_link_amd",
96
+ "job_link_nvidia",
97
+ ]
98
+ ]
99
+ joined.index = joined.index.str.replace("^models_", "", regex=True)
100
+ # Fitler out all but important models
101
+ important_models_lower = [model.lower() for model in IMPORTANT_MODELS]
102
+ filtered_joined = joined[joined.index.str.lower().isin(important_models_lower)]
103
+ # Prefix all model names with "sample_"
104
+ filtered_joined.index = "sample_" + filtered_joined.index
105
+ return filtered_joined
106
 
107
  def extract_model_data(row: pd.Series) -> tuple[dict[str, int], dict[str, int], int, int, int, int]:
108
  """Extract and process model data from DataFrame row."""
sample_amd.json ADDED
@@ -0,0 +1,1709 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_auto": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 0
8
+ },
9
+ "TensorFlow": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Flax": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Tokenizers": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "Pipelines": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Trainer": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "ONNX": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Auto": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ },
44
+ "Quantization": {
45
+ "unclassified": 0,
46
+ "single": 0,
47
+ "multi": 0
48
+ },
49
+ "Unclassified": {
50
+ "unclassified": 0,
51
+ "single": 0,
52
+ "multi": 0
53
+ }
54
+ },
55
+ "success": 80,
56
+ "time_spent": "5.47, 1.74, ",
57
+ "failures": {},
58
+ "job_link": {
59
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409911",
60
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410689"
61
+ }
62
+ },
63
+ "models_bert": {
64
+ "failed": {
65
+ "PyTorch": {
66
+ "unclassified": 0,
67
+ "single": 2,
68
+ "multi": 2
69
+ },
70
+ "TensorFlow": {
71
+ "unclassified": 0,
72
+ "single": 0,
73
+ "multi": 0
74
+ },
75
+ "Flax": {
76
+ "unclassified": 0,
77
+ "single": 0,
78
+ "multi": 0
79
+ },
80
+ "Tokenizers": {
81
+ "unclassified": 0,
82
+ "single": 0,
83
+ "multi": 0
84
+ },
85
+ "Pipelines": {
86
+ "unclassified": 0,
87
+ "single": 0,
88
+ "multi": 0
89
+ },
90
+ "Trainer": {
91
+ "unclassified": 0,
92
+ "single": 0,
93
+ "multi": 0
94
+ },
95
+ "ONNX": {
96
+ "unclassified": 0,
97
+ "single": 0,
98
+ "multi": 0
99
+ },
100
+ "Auto": {
101
+ "unclassified": 0,
102
+ "single": 0,
103
+ "multi": 0
104
+ },
105
+ "Quantization": {
106
+ "unclassified": 0,
107
+ "single": 0,
108
+ "multi": 0
109
+ },
110
+ "Unclassified": {
111
+ "unclassified": 0,
112
+ "single": 0,
113
+ "multi": 0
114
+ }
115
+ },
116
+ "success": 239,
117
+ "time_spent": "0:02:22, 0:02:16, ",
118
+ "failures": {
119
+ "single": [
120
+ {
121
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids",
122
+ "trace": "(line 4140) KeyError: 'eager'"
123
+ },
124
+ {
125
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
126
+ "trace": "(line 4216) AssertionError: Tensor-likes are not equal!"
127
+ }
128
+ ],
129
+ "multi": [
130
+ {
131
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids",
132
+ "trace": "(line 4140) KeyError: 'eager'"
133
+ },
134
+ {
135
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
136
+ "trace": "(line 4216) AssertionError: Tensor-likes are not equal!"
137
+ }
138
+ ]
139
+ },
140
+ "job_link": {
141
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409938",
142
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410716"
143
+ }
144
+ },
145
+ "models_clip": {
146
+ "failed": {
147
+ "PyTorch": {
148
+ "unclassified": 0,
149
+ "single": 0,
150
+ "multi": 0
151
+ },
152
+ "TensorFlow": {
153
+ "unclassified": 0,
154
+ "single": 0,
155
+ "multi": 0
156
+ },
157
+ "Flax": {
158
+ "unclassified": 0,
159
+ "single": 0,
160
+ "multi": 0
161
+ },
162
+ "Tokenizers": {
163
+ "unclassified": 0,
164
+ "single": 0,
165
+ "multi": 0
166
+ },
167
+ "Pipelines": {
168
+ "unclassified": 0,
169
+ "single": 0,
170
+ "multi": 0
171
+ },
172
+ "Trainer": {
173
+ "unclassified": 0,
174
+ "single": 0,
175
+ "multi": 0
176
+ },
177
+ "ONNX": {
178
+ "unclassified": 0,
179
+ "single": 0,
180
+ "multi": 0
181
+ },
182
+ "Auto": {
183
+ "unclassified": 0,
184
+ "single": 0,
185
+ "multi": 0
186
+ },
187
+ "Quantization": {
188
+ "unclassified": 0,
189
+ "single": 0,
190
+ "multi": 0
191
+ },
192
+ "Unclassified": {
193
+ "unclassified": 0,
194
+ "single": 0,
195
+ "multi": 0
196
+ }
197
+ },
198
+ "success": 288,
199
+ "time_spent": "0:03:29, 0:03:30, ",
200
+ "failures": {},
201
+ "job_link": {
202
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410734",
203
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409965"
204
+ }
205
+ },
206
+ "models_detr": {
207
+ "failed": {
208
+ "PyTorch": {
209
+ "unclassified": 0,
210
+ "single": 0,
211
+ "multi": 0
212
+ },
213
+ "TensorFlow": {
214
+ "unclassified": 0,
215
+ "single": 0,
216
+ "multi": 0
217
+ },
218
+ "Flax": {
219
+ "unclassified": 0,
220
+ "single": 0,
221
+ "multi": 0
222
+ },
223
+ "Tokenizers": {
224
+ "unclassified": 0,
225
+ "single": 0,
226
+ "multi": 0
227
+ },
228
+ "Pipelines": {
229
+ "unclassified": 0,
230
+ "single": 0,
231
+ "multi": 0
232
+ },
233
+ "Trainer": {
234
+ "unclassified": 0,
235
+ "single": 0,
236
+ "multi": 0
237
+ },
238
+ "ONNX": {
239
+ "unclassified": 0,
240
+ "single": 0,
241
+ "multi": 0
242
+ },
243
+ "Auto": {
244
+ "unclassified": 0,
245
+ "single": 0,
246
+ "multi": 0
247
+ },
248
+ "Quantization": {
249
+ "unclassified": 0,
250
+ "single": 0,
251
+ "multi": 0
252
+ },
253
+ "Unclassified": {
254
+ "unclassified": 0,
255
+ "single": 0,
256
+ "multi": 0
257
+ }
258
+ },
259
+ "success": 77,
260
+ "time_spent": "0:01:13, 0:01:49, ",
261
+ "failures": {},
262
+ "job_link": {
263
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410020",
264
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410868"
265
+ }
266
+ },
267
+ "models_gemma3": {
268
+ "failed": {
269
+ "PyTorch": {
270
+ "unclassified": 0,
271
+ "single": 12,
272
+ "multi": 13
273
+ },
274
+ "TensorFlow": {
275
+ "unclassified": 0,
276
+ "single": 0,
277
+ "multi": 0
278
+ },
279
+ "Flax": {
280
+ "unclassified": 0,
281
+ "single": 0,
282
+ "multi": 0
283
+ },
284
+ "Tokenizers": {
285
+ "unclassified": 0,
286
+ "single": 0,
287
+ "multi": 0
288
+ },
289
+ "Pipelines": {
290
+ "unclassified": 0,
291
+ "single": 0,
292
+ "multi": 0
293
+ },
294
+ "Trainer": {
295
+ "unclassified": 0,
296
+ "single": 0,
297
+ "multi": 0
298
+ },
299
+ "ONNX": {
300
+ "unclassified": 0,
301
+ "single": 0,
302
+ "multi": 0
303
+ },
304
+ "Auto": {
305
+ "unclassified": 0,
306
+ "single": 0,
307
+ "multi": 0
308
+ },
309
+ "Quantization": {
310
+ "unclassified": 0,
311
+ "single": 0,
312
+ "multi": 0
313
+ },
314
+ "Unclassified": {
315
+ "unclassified": 0,
316
+ "single": 0,
317
+ "multi": 0
318
+ }
319
+ },
320
+ "success": 341,
321
+ "time_spent": "0:07:52, 0:09:43, ",
322
+ "failures": {
323
+ "single": [
324
+ {
325
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
326
+ "trace": "(line 4140) KeyError: 'eager'"
327
+ },
328
+ {
329
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
330
+ "trace": "(line 4219) AssertionError: Tensor-likes are not close!"
331
+ },
332
+ {
333
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids",
334
+ "trace": "(line 4140) KeyError: 'eager'"
335
+ },
336
+ {
337
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only_with_hybrid_cache",
338
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., size=(1, 4, 1, 256), grad_fn=<AddBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>)), **{'attn_mask': FakeTensor(..., size=(1, 1, 1, 512), dtype=torch.bool), 'dropout_p': 0.0, 'scale': 0.0625, 'is_causal': False}): got RuntimeError('Attempting to broadcast a dimension of length 512 at -1! Mismatching argument at index 1 had torch.Size([1, 1, 1, 512]); but expected shape should be broadcastable to [1, 4, 1, 4096]')"
339
+ },
340
+ {
341
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa",
342
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (4826) must match the existing size (4807) at non-singleton dimension 3. Target sizes: [2, 4, 4807, 4826]. Tensor sizes: [2, 1, 4807, 4807]"
343
+ },
344
+ {
345
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager",
346
+ "trace": "(line 265) RuntimeError: The size of tensor a (4826) must match the size of tensor b (4807) at non-singleton dimension 3"
347
+ },
348
+ {
349
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_1b_text_only",
350
+ "trace": "(line 715) AssertionError: Lists differ: ['Wri[57 chars]s, a silent stream,\\nInto the neural net, a wa[42 chars],\\n'] != ['Wri[57 chars]s, a river deep,\\nWith patterns hidden, secret[46 chars]ing']"
351
+ },
352
+ {
353
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch",
354
+ "trace": "(line 715) AssertionError: Lists differ: ['use[114 chars]rown cow standing on a sandy beach with clear [264 chars]cow\"] != ['use[114 chars]rown and white cow standing on a sandy beach n[272 chars]ach']"
355
+ },
356
+ {
357
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops",
358
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (1646) must match the existing size (1617) at non-singleton dimension 3. Target sizes: [2, 8, 1617, 1646]. Tensor sizes: [2, 1, 1617, 1617]"
359
+ },
360
+ {
361
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_bf16",
362
+ "trace": "(line 715) AssertionError: Lists differ: ['use[114 chars]rown cow standing on a sandy beach with clear [55 chars]ike'] != ['use[114 chars]rown and white cow standing on a sandy beach w[68 chars]oks']"
363
+ },
364
+ {
365
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_crops",
366
+ "trace": "(line 715) AssertionError: Lists differ: [\"use[251 chars]. There's a blue sky with some white clouds in the background\"] != [\"use[251 chars]. There's a bright blue sky with some white clouds in the\"]"
367
+ },
368
+ {
369
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_multiimage",
370
+ "trace": "(line 715) AssertionError: Lists differ: [\"use[122 chars]n\\n**Main Features:**\\n\\n* **Chinese Archway[19 chars]ent\"] != [\"use[122 chars]n\\n**Overall Scene:**\\n\\nIt looks like a stree[18 chars]nt,\"]"
371
+ }
372
+ ],
373
+ "multi": [
374
+ {
375
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
376
+ "trace": "(line 4140) KeyError: 'eager'"
377
+ },
378
+ {
379
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
380
+ "trace": "(line 4219) AssertionError: Tensor-likes are not close!"
381
+ },
382
+ {
383
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids",
384
+ "trace": "(line 4140) KeyError: 'eager'"
385
+ },
386
+ {
387
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_model_parallelism",
388
+ "trace": "(line 925) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0!"
389
+ },
390
+ {
391
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only_with_hybrid_cache",
392
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., size=(1, 4, 1, 256), grad_fn=<AddBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>)), **{'attn_mask': FakeTensor(..., size=(1, 1, 1, 512), dtype=torch.bool), 'dropout_p': 0.0, 'scale': 0.0625, 'is_causal': False}): got RuntimeError('Attempting to broadcast a dimension of length 512 at -1! Mismatching argument at index 1 had torch.Size([1, 1, 1, 512]); but expected shape should be broadcastable to [1, 4, 1, 4096]')"
393
+ },
394
+ {
395
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa",
396
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (4826) must match the existing size (4807) at non-singleton dimension 3. Target sizes: [2, 4, 4807, 4826]. Tensor sizes: [2, 1, 4807, 4807]"
397
+ },
398
+ {
399
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager",
400
+ "trace": "(line 265) RuntimeError: The size of tensor a (4826) must match the size of tensor b (4807) at non-singleton dimension 3"
401
+ },
402
+ {
403
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_1b_text_only",
404
+ "trace": "(line 715) AssertionError: Lists differ: ['Wri[57 chars]s, a silent stream,\\nInto the neural net, a wa[42 chars],\\n'] != ['Wri[57 chars]s, a river deep,\\nWith patterns hidden, secret[46 chars]ing']"
405
+ },
406
+ {
407
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch",
408
+ "trace": "(line 715) AssertionError: Lists differ: ['use[114 chars]rown cow standing on a sandy beach with clear [264 chars]cow\"] != ['use[114 chars]rown and white cow standing on a sandy beach n[272 chars]ach']"
409
+ },
410
+ {
411
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops",
412
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (1646) must match the existing size (1617) at non-singleton dimension 3. Target sizes: [2, 8, 1617, 1646]. Tensor sizes: [2, 1, 1617, 1617]"
413
+ },
414
+ {
415
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_bf16",
416
+ "trace": "(line 715) AssertionError: Lists differ: ['use[114 chars]rown cow standing on a sandy beach with clear [55 chars]ike'] != ['use[114 chars]rown and white cow standing on a sandy beach w[68 chars]oks']"
417
+ },
418
+ {
419
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_crops",
420
+ "trace": "(line 715) AssertionError: Lists differ: [\"use[251 chars]. There's a blue sky with some white clouds in the background\"] != [\"use[251 chars]. There's a bright blue sky with some white clouds in the\"]"
421
+ },
422
+ {
423
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_multiimage",
424
+ "trace": "(line 715) AssertionError: Lists differ: [\"use[122 chars]n\\n**Main Features:**\\n\\n* **Chinese Archway[19 chars]ent\"] != [\"use[122 chars]n\\n**Overall Scene:**\\n\\nIt looks like a stree[18 chars]nt,\"]"
425
+ }
426
+ ]
427
+ },
428
+ "job_link": {
429
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410076",
430
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410943"
431
+ }
432
+ },
433
+ "models_gemma3n": {
434
+ "failed": {
435
+ "PyTorch": {
436
+ "unclassified": 0,
437
+ "single": 0,
438
+ "multi": 0
439
+ },
440
+ "TensorFlow": {
441
+ "unclassified": 0,
442
+ "single": 0,
443
+ "multi": 0
444
+ },
445
+ "Flax": {
446
+ "unclassified": 0,
447
+ "single": 0,
448
+ "multi": 0
449
+ },
450
+ "Tokenizers": {
451
+ "unclassified": 0,
452
+ "single": 0,
453
+ "multi": 0
454
+ },
455
+ "Pipelines": {
456
+ "unclassified": 0,
457
+ "single": 0,
458
+ "multi": 0
459
+ },
460
+ "Trainer": {
461
+ "unclassified": 0,
462
+ "single": 0,
463
+ "multi": 0
464
+ },
465
+ "ONNX": {
466
+ "unclassified": 0,
467
+ "single": 0,
468
+ "multi": 0
469
+ },
470
+ "Auto": {
471
+ "unclassified": 0,
472
+ "single": 0,
473
+ "multi": 0
474
+ },
475
+ "Quantization": {
476
+ "unclassified": 0,
477
+ "single": 0,
478
+ "multi": 0
479
+ },
480
+ "Unclassified": {
481
+ "unclassified": 0,
482
+ "single": 0,
483
+ "multi": 0
484
+ }
485
+ },
486
+ "success": 0,
487
+ "time_spent": ".56, .97, ",
488
+ "failures": {},
489
+ "job_link": {
490
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410944",
491
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410122"
492
+ }
493
+ },
494
+ "models_got_ocr2": {
495
+ "failed": {
496
+ "PyTorch": {
497
+ "unclassified": 0,
498
+ "single": 1,
499
+ "multi": 2
500
+ },
501
+ "TensorFlow": {
502
+ "unclassified": 0,
503
+ "single": 0,
504
+ "multi": 0
505
+ },
506
+ "Flax": {
507
+ "unclassified": 0,
508
+ "single": 0,
509
+ "multi": 0
510
+ },
511
+ "Tokenizers": {
512
+ "unclassified": 0,
513
+ "single": 0,
514
+ "multi": 0
515
+ },
516
+ "Pipelines": {
517
+ "unclassified": 0,
518
+ "single": 0,
519
+ "multi": 0
520
+ },
521
+ "Trainer": {
522
+ "unclassified": 0,
523
+ "single": 0,
524
+ "multi": 0
525
+ },
526
+ "ONNX": {
527
+ "unclassified": 0,
528
+ "single": 0,
529
+ "multi": 0
530
+ },
531
+ "Auto": {
532
+ "unclassified": 0,
533
+ "single": 0,
534
+ "multi": 0
535
+ },
536
+ "Quantization": {
537
+ "unclassified": 0,
538
+ "single": 0,
539
+ "multi": 0
540
+ },
541
+ "Unclassified": {
542
+ "unclassified": 0,
543
+ "single": 0,
544
+ "multi": 0
545
+ }
546
+ },
547
+ "success": 146,
548
+ "time_spent": "0:01:56, 0:01:39, ",
549
+ "failures": {
550
+ "multi": [
551
+ {
552
+ "line": "tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
553
+ "trace": "(line 4140) KeyError: 'eager'"
554
+ },
555
+ {
556
+ "line": "tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_multi_gpu_data_parallel_forward",
557
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
558
+ }
559
+ ],
560
+ "single": [
561
+ {
562
+ "line": "tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
563
+ "trace": "(line 4140) KeyError: 'eager'"
564
+ }
565
+ ]
566
+ },
567
+ "job_link": {
568
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410969",
569
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410123"
570
+ }
571
+ },
572
+ "models_gpt2": {
573
+ "failed": {
574
+ "PyTorch": {
575
+ "unclassified": 0,
576
+ "single": 1,
577
+ "multi": 1
578
+ },
579
+ "TensorFlow": {
580
+ "unclassified": 0,
581
+ "single": 0,
582
+ "multi": 0
583
+ },
584
+ "Flax": {
585
+ "unclassified": 0,
586
+ "single": 0,
587
+ "multi": 0
588
+ },
589
+ "Tokenizers": {
590
+ "unclassified": 0,
591
+ "single": 0,
592
+ "multi": 0
593
+ },
594
+ "Pipelines": {
595
+ "unclassified": 0,
596
+ "single": 0,
597
+ "multi": 0
598
+ },
599
+ "Trainer": {
600
+ "unclassified": 0,
601
+ "single": 0,
602
+ "multi": 0
603
+ },
604
+ "ONNX": {
605
+ "unclassified": 0,
606
+ "single": 0,
607
+ "multi": 0
608
+ },
609
+ "Auto": {
610
+ "unclassified": 0,
611
+ "single": 0,
612
+ "multi": 0
613
+ },
614
+ "Quantization": {
615
+ "unclassified": 0,
616
+ "single": 0,
617
+ "multi": 0
618
+ },
619
+ "Unclassified": {
620
+ "unclassified": 0,
621
+ "single": 0,
622
+ "multi": 0
623
+ }
624
+ },
625
+ "success": 249,
626
+ "time_spent": "0:04:53, 0:02:05, ",
627
+ "failures": {
628
+ "multi": [
629
+ {
630
+ "line": "tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
631
+ "trace": "(line 4140) KeyError: 'eager'"
632
+ }
633
+ ],
634
+ "single": [
635
+ {
636
+ "line": "tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
637
+ "trace": "(line 4140) KeyError: 'eager'"
638
+ }
639
+ ]
640
+ },
641
+ "job_link": {
642
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410990",
643
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410088"
644
+ }
645
+ },
646
+ "models_internvl": {
647
+ "failed": {
648
+ "PyTorch": {
649
+ "unclassified": 0,
650
+ "single": 2,
651
+ "multi": 3
652
+ },
653
+ "TensorFlow": {
654
+ "unclassified": 0,
655
+ "single": 0,
656
+ "multi": 0
657
+ },
658
+ "Flax": {
659
+ "unclassified": 0,
660
+ "single": 0,
661
+ "multi": 0
662
+ },
663
+ "Tokenizers": {
664
+ "unclassified": 0,
665
+ "single": 0,
666
+ "multi": 0
667
+ },
668
+ "Pipelines": {
669
+ "unclassified": 0,
670
+ "single": 0,
671
+ "multi": 0
672
+ },
673
+ "Trainer": {
674
+ "unclassified": 0,
675
+ "single": 0,
676
+ "multi": 0
677
+ },
678
+ "ONNX": {
679
+ "unclassified": 0,
680
+ "single": 0,
681
+ "multi": 0
682
+ },
683
+ "Auto": {
684
+ "unclassified": 0,
685
+ "single": 0,
686
+ "multi": 0
687
+ },
688
+ "Quantization": {
689
+ "unclassified": 0,
690
+ "single": 0,
691
+ "multi": 0
692
+ },
693
+ "Unclassified": {
694
+ "unclassified": 0,
695
+ "single": 0,
696
+ "multi": 0
697
+ }
698
+ },
699
+ "success": 252,
700
+ "time_spent": "0:02:54, 0:02:55, ",
701
+ "failures": {
702
+ "multi": [
703
+ {
704
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids",
705
+ "trace": "(line 4140) KeyError: 'eager'"
706
+ },
707
+ {
708
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_multi_gpu_data_parallel_forward",
709
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
710
+ },
711
+ {
712
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_forward",
713
+ "trace": "(line 727) AssertionError: False is not true : Actual logits: tensor([ -9.8750, -0.4885, 1.4668, -10.3359, -10.3359], dtype=torch.float16)"
714
+ }
715
+ ],
716
+ "single": [
717
+ {
718
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids",
719
+ "trace": "(line 4140) KeyError: 'eager'"
720
+ },
721
+ {
722
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_forward",
723
+ "trace": "(line 727) AssertionError: False is not true : Actual logits: tensor([ -9.8750, -0.4885, 1.4668, -10.3359, -10.3359], dtype=torch.float16)"
724
+ }
725
+ ]
726
+ },
727
+ "job_link": {
728
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527411014",
729
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410165"
730
+ }
731
+ },
732
+ "models_llama": {
733
+ "failed": {
734
+ "PyTorch": {
735
+ "unclassified": 0,
736
+ "single": 2,
737
+ "multi": 3
738
+ },
739
+ "TensorFlow": {
740
+ "unclassified": 0,
741
+ "single": 0,
742
+ "multi": 0
743
+ },
744
+ "Flax": {
745
+ "unclassified": 0,
746
+ "single": 0,
747
+ "multi": 0
748
+ },
749
+ "Tokenizers": {
750
+ "unclassified": 0,
751
+ "single": 0,
752
+ "multi": 0
753
+ },
754
+ "Pipelines": {
755
+ "unclassified": 0,
756
+ "single": 0,
757
+ "multi": 0
758
+ },
759
+ "Trainer": {
760
+ "unclassified": 0,
761
+ "single": 0,
762
+ "multi": 0
763
+ },
764
+ "ONNX": {
765
+ "unclassified": 0,
766
+ "single": 0,
767
+ "multi": 0
768
+ },
769
+ "Auto": {
770
+ "unclassified": 0,
771
+ "single": 0,
772
+ "multi": 0
773
+ },
774
+ "Quantization": {
775
+ "unclassified": 0,
776
+ "single": 0,
777
+ "multi": 0
778
+ },
779
+ "Unclassified": {
780
+ "unclassified": 0,
781
+ "single": 0,
782
+ "multi": 0
783
+ }
784
+ },
785
+ "success": 232,
786
+ "time_spent": "0:10:51, 0:23:47, ",
787
+ "failures": {
788
+ "multi": [
789
+ {
790
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids",
791
+ "trace": "(line 4140) KeyError: 'eager'"
792
+ },
793
+ {
794
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_multi_gpu_data_parallel_forward",
795
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
796
+ },
797
+ {
798
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaIntegrationTest::test_model_7b_logits_bf16",
799
+ "trace": "(line 727) AssertionError: False is not true"
800
+ }
801
+ ],
802
+ "single": [
803
+ {
804
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids",
805
+ "trace": "(line 4140) KeyError: 'eager'"
806
+ },
807
+ {
808
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaIntegrationTest::test_model_7b_logits_bf16",
809
+ "trace": "(line 727) AssertionError: False is not true"
810
+ }
811
+ ]
812
+ },
813
+ "job_link": {
814
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527411041",
815
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410199"
816
+ }
817
+ },
818
+ "models_llava": {
819
+ "failed": {
820
+ "PyTorch": {
821
+ "unclassified": 0,
822
+ "single": 4,
823
+ "multi": 5
824
+ },
825
+ "TensorFlow": {
826
+ "unclassified": 0,
827
+ "single": 0,
828
+ "multi": 0
829
+ },
830
+ "Flax": {
831
+ "unclassified": 0,
832
+ "single": 0,
833
+ "multi": 0
834
+ },
835
+ "Tokenizers": {
836
+ "unclassified": 0,
837
+ "single": 0,
838
+ "multi": 0
839
+ },
840
+ "Pipelines": {
841
+ "unclassified": 0,
842
+ "single": 0,
843
+ "multi": 0
844
+ },
845
+ "Trainer": {
846
+ "unclassified": 0,
847
+ "single": 0,
848
+ "multi": 0
849
+ },
850
+ "ONNX": {
851
+ "unclassified": 0,
852
+ "single": 0,
853
+ "multi": 0
854
+ },
855
+ "Auto": {
856
+ "unclassified": 0,
857
+ "single": 0,
858
+ "multi": 0
859
+ },
860
+ "Quantization": {
861
+ "unclassified": 0,
862
+ "single": 0,
863
+ "multi": 0
864
+ },
865
+ "Unclassified": {
866
+ "unclassified": 0,
867
+ "single": 0,
868
+ "multi": 0
869
+ }
870
+ },
871
+ "success": 202,
872
+ "time_spent": "0:02:38, 0:02:51, ",
873
+ "failures": {
874
+ "multi": [
875
+ {
876
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
877
+ "trace": "(line 4140) KeyError: 'eager'"
878
+ },
879
+ {
880
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads",
881
+ "trace": "(line 727) AssertionError: False is not true"
882
+ },
883
+ {
884
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
885
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
886
+ },
887
+ {
888
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
889
+ "trace": "(line 4197) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1"
890
+ },
891
+ {
892
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation",
893
+ "trace": "(line 399) importlib.metadata.PackageNotFoundError: No package metadata was found for bitsandbytes"
894
+ }
895
+ ],
896
+ "single": [
897
+ {
898
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
899
+ "trace": "(line 4140) KeyError: 'eager'"
900
+ },
901
+ {
902
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads",
903
+ "trace": "(line 727) AssertionError: False is not true"
904
+ },
905
+ {
906
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
907
+ "trace": "(line 4197) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1"
908
+ },
909
+ {
910
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation",
911
+ "trace": "(line 399) importlib.metadata.PackageNotFoundError: No package metadata was found for bitsandbytes"
912
+ }
913
+ ]
914
+ },
915
+ "job_link": {
916
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527411134",
917
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410218"
918
+ }
919
+ },
920
+ "models_mistral3": {
921
+ "failed": {
922
+ "PyTorch": {
923
+ "unclassified": 0,
924
+ "single": 2,
925
+ "multi": 3
926
+ },
927
+ "TensorFlow": {
928
+ "unclassified": 0,
929
+ "single": 0,
930
+ "multi": 0
931
+ },
932
+ "Flax": {
933
+ "unclassified": 0,
934
+ "single": 0,
935
+ "multi": 0
936
+ },
937
+ "Tokenizers": {
938
+ "unclassified": 0,
939
+ "single": 0,
940
+ "multi": 0
941
+ },
942
+ "Pipelines": {
943
+ "unclassified": 0,
944
+ "single": 0,
945
+ "multi": 0
946
+ },
947
+ "Trainer": {
948
+ "unclassified": 0,
949
+ "single": 0,
950
+ "multi": 0
951
+ },
952
+ "ONNX": {
953
+ "unclassified": 0,
954
+ "single": 0,
955
+ "multi": 0
956
+ },
957
+ "Auto": {
958
+ "unclassified": 0,
959
+ "single": 0,
960
+ "multi": 0
961
+ },
962
+ "Quantization": {
963
+ "unclassified": 0,
964
+ "single": 0,
965
+ "multi": 0
966
+ },
967
+ "Unclassified": {
968
+ "unclassified": 0,
969
+ "single": 0,
970
+ "multi": 0
971
+ }
972
+ },
973
+ "success": 198,
974
+ "time_spent": "0:14:37, 0:05:43, ",
975
+ "failures": {
976
+ "single": [
977
+ {
978
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
979
+ "trace": "(line 4140) KeyError: 'eager'"
980
+ },
981
+ {
982
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3IntegrationTest::test_mistral3_integration_generate",
983
+ "trace": "(line 715) AssertionError: 'The [14 chars] two cats lying on a pink surface, which appea[21 chars] bed' != 'The [14 chars] two tabby cats lying on a pink surface, which[23 chars]n or'"
984
+ }
985
+ ],
986
+ "multi": [
987
+ {
988
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
989
+ "trace": "(line 4140) KeyError: 'eager'"
990
+ },
991
+ {
992
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_multi_gpu_data_parallel_forward",
993
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
994
+ },
995
+ {
996
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3IntegrationTest::test_mistral3_integration_generate",
997
+ "trace": "(line 715) AssertionError: 'The [14 chars] two cats lying on a pink surface, which appea[21 chars] bed' != 'The [14 chars] two tabby cats lying on a pink surface, which[23 chars]n or'"
998
+ }
999
+ ]
1000
+ },
1001
+ "job_link": {
1002
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409417",
1003
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410265"
1004
+ }
1005
+ },
1006
+ "models_modernbert": {
1007
+ "failed": {
1008
+ "PyTorch": {
1009
+ "unclassified": 0,
1010
+ "single": 5,
1011
+ "multi": 5
1012
+ },
1013
+ "TensorFlow": {
1014
+ "unclassified": 0,
1015
+ "single": 0,
1016
+ "multi": 0
1017
+ },
1018
+ "Flax": {
1019
+ "unclassified": 0,
1020
+ "single": 0,
1021
+ "multi": 0
1022
+ },
1023
+ "Tokenizers": {
1024
+ "unclassified": 0,
1025
+ "single": 0,
1026
+ "multi": 0
1027
+ },
1028
+ "Pipelines": {
1029
+ "unclassified": 0,
1030
+ "single": 0,
1031
+ "multi": 0
1032
+ },
1033
+ "Trainer": {
1034
+ "unclassified": 0,
1035
+ "single": 0,
1036
+ "multi": 0
1037
+ },
1038
+ "ONNX": {
1039
+ "unclassified": 0,
1040
+ "single": 0,
1041
+ "multi": 0
1042
+ },
1043
+ "Auto": {
1044
+ "unclassified": 0,
1045
+ "single": 0,
1046
+ "multi": 0
1047
+ },
1048
+ "Quantization": {
1049
+ "unclassified": 0,
1050
+ "single": 0,
1051
+ "multi": 0
1052
+ },
1053
+ "Unclassified": {
1054
+ "unclassified": 0,
1055
+ "single": 0,
1056
+ "multi": 0
1057
+ }
1058
+ },
1059
+ "success": 132,
1060
+ "time_spent": "0:02:22, 0:01:49, ",
1061
+ "failures": {
1062
+ "multi": [
1063
+ {
1064
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export",
1065
+ "trace": "(line 715) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']"
1066
+ },
1067
+ {
1068
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm",
1069
+ "trace": "(line 401) AssertionError: Tensor-likes are not close!"
1070
+ },
1071
+ {
1072
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head",
1073
+ "trace": "(line 423) AssertionError: Tensor-likes are not close!"
1074
+ },
1075
+ {
1076
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification",
1077
+ "trace": "(line 469) AssertionError: Tensor-likes are not close!"
1078
+ },
1079
+ {
1080
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification",
1081
+ "trace": "(line 446) AssertionError: Tensor-likes are not close!"
1082
+ }
1083
+ ],
1084
+ "single": [
1085
+ {
1086
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export",
1087
+ "trace": "(line 715) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']"
1088
+ },
1089
+ {
1090
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm",
1091
+ "trace": "(line 401) AssertionError: Tensor-likes are not close!"
1092
+ },
1093
+ {
1094
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head",
1095
+ "trace": "(line 423) AssertionError: Tensor-likes are not close!"
1096
+ },
1097
+ {
1098
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification",
1099
+ "trace": "(line 469) AssertionError: Tensor-likes are not close!"
1100
+ },
1101
+ {
1102
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification",
1103
+ "trace": "(line 446) AssertionError: Tensor-likes are not close!"
1104
+ }
1105
+ ]
1106
+ },
1107
+ "job_link": {
1108
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410294",
1109
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409446"
1110
+ }
1111
+ },
1112
+ "models_qwen2": {
1113
+ "failed": {
1114
+ "PyTorch": {
1115
+ "unclassified": 0,
1116
+ "single": 2,
1117
+ "multi": 3
1118
+ },
1119
+ "TensorFlow": {
1120
+ "unclassified": 0,
1121
+ "single": 0,
1122
+ "multi": 0
1123
+ },
1124
+ "Flax": {
1125
+ "unclassified": 0,
1126
+ "single": 0,
1127
+ "multi": 0
1128
+ },
1129
+ "Tokenizers": {
1130
+ "unclassified": 0,
1131
+ "single": 0,
1132
+ "multi": 0
1133
+ },
1134
+ "Pipelines": {
1135
+ "unclassified": 0,
1136
+ "single": 0,
1137
+ "multi": 0
1138
+ },
1139
+ "Trainer": {
1140
+ "unclassified": 0,
1141
+ "single": 0,
1142
+ "multi": 0
1143
+ },
1144
+ "ONNX": {
1145
+ "unclassified": 0,
1146
+ "single": 0,
1147
+ "multi": 0
1148
+ },
1149
+ "Auto": {
1150
+ "unclassified": 0,
1151
+ "single": 0,
1152
+ "multi": 0
1153
+ },
1154
+ "Quantization": {
1155
+ "unclassified": 0,
1156
+ "single": 0,
1157
+ "multi": 0
1158
+ },
1159
+ "Unclassified": {
1160
+ "unclassified": 0,
1161
+ "single": 0,
1162
+ "multi": 0
1163
+ }
1164
+ },
1165
+ "success": 214,
1166
+ "time_spent": "0:02:23, 0:02:39, ",
1167
+ "failures": {
1168
+ "multi": [
1169
+ {
1170
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1171
+ "trace": "(line 4140) KeyError: 'eager'"
1172
+ },
1173
+ {
1174
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_multi_gpu_data_parallel_forward",
1175
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
1176
+ },
1177
+ {
1178
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache",
1179
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,"
1180
+ }
1181
+ ],
1182
+ "single": [
1183
+ {
1184
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1185
+ "trace": "(line 4140) KeyError: 'eager'"
1186
+ },
1187
+ {
1188
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache",
1189
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,"
1190
+ }
1191
+ ]
1192
+ },
1193
+ "job_link": {
1194
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410392",
1195
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409572"
1196
+ }
1197
+ },
1198
+ "models_qwen2_5_omni": {
1199
+ "failed": {
1200
+ "PyTorch": {
1201
+ "unclassified": 0,
1202
+ "single": 1,
1203
+ "multi": 3
1204
+ },
1205
+ "TensorFlow": {
1206
+ "unclassified": 0,
1207
+ "single": 0,
1208
+ "multi": 0
1209
+ },
1210
+ "Flax": {
1211
+ "unclassified": 0,
1212
+ "single": 0,
1213
+ "multi": 0
1214
+ },
1215
+ "Tokenizers": {
1216
+ "unclassified": 0,
1217
+ "single": 0,
1218
+ "multi": 0
1219
+ },
1220
+ "Pipelines": {
1221
+ "unclassified": 0,
1222
+ "single": 0,
1223
+ "multi": 0
1224
+ },
1225
+ "Trainer": {
1226
+ "unclassified": 0,
1227
+ "single": 0,
1228
+ "multi": 0
1229
+ },
1230
+ "ONNX": {
1231
+ "unclassified": 0,
1232
+ "single": 0,
1233
+ "multi": 0
1234
+ },
1235
+ "Auto": {
1236
+ "unclassified": 0,
1237
+ "single": 0,
1238
+ "multi": 0
1239
+ },
1240
+ "Quantization": {
1241
+ "unclassified": 0,
1242
+ "single": 0,
1243
+ "multi": 0
1244
+ },
1245
+ "Unclassified": {
1246
+ "unclassified": 0,
1247
+ "single": 0,
1248
+ "multi": 0
1249
+ }
1250
+ },
1251
+ "success": 167,
1252
+ "time_spent": "0:06:59, 0:02:55, ",
1253
+ "failures": {
1254
+ "multi": [
1255
+ {
1256
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_model_parallelism",
1257
+ "trace": "(line 715) AssertionError: Items in the second set but not the first:"
1258
+ },
1259
+ {
1260
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
1261
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
1262
+ },
1263
+ {
1264
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
1265
+ "trace": "(line 715) AssertionError: Lists differ: [\"sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.\"] != [\"sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.\"]"
1266
+ }
1267
+ ],
1268
+ "single": [
1269
+ {
1270
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
1271
+ "trace": "(line 715) AssertionError: Lists differ: [\"sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.\"] != [\"sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.\"]"
1272
+ }
1273
+ ]
1274
+ },
1275
+ "job_link": {
1276
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410407",
1277
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409568"
1278
+ }
1279
+ },
1280
+ "models_qwen2_5_vl": {
1281
+ "failed": {
1282
+ "PyTorch": {
1283
+ "unclassified": 0,
1284
+ "single": 1,
1285
+ "multi": 2
1286
+ },
1287
+ "TensorFlow": {
1288
+ "unclassified": 0,
1289
+ "single": 0,
1290
+ "multi": 0
1291
+ },
1292
+ "Flax": {
1293
+ "unclassified": 0,
1294
+ "single": 0,
1295
+ "multi": 0
1296
+ },
1297
+ "Tokenizers": {
1298
+ "unclassified": 0,
1299
+ "single": 0,
1300
+ "multi": 0
1301
+ },
1302
+ "Pipelines": {
1303
+ "unclassified": 0,
1304
+ "single": 0,
1305
+ "multi": 0
1306
+ },
1307
+ "Trainer": {
1308
+ "unclassified": 0,
1309
+ "single": 0,
1310
+ "multi": 0
1311
+ },
1312
+ "ONNX": {
1313
+ "unclassified": 0,
1314
+ "single": 0,
1315
+ "multi": 0
1316
+ },
1317
+ "Auto": {
1318
+ "unclassified": 0,
1319
+ "single": 0,
1320
+ "multi": 0
1321
+ },
1322
+ "Quantization": {
1323
+ "unclassified": 0,
1324
+ "single": 0,
1325
+ "multi": 0
1326
+ },
1327
+ "Unclassified": {
1328
+ "unclassified": 0,
1329
+ "single": 0,
1330
+ "multi": 0
1331
+ }
1332
+ },
1333
+ "success": 204,
1334
+ "time_spent": "0:03:59, 0:03:58, ",
1335
+ "failures": {
1336
+ "multi": [
1337
+ {
1338
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
1339
+ "trace": "(line 406) AssertionError: Tensor-likes are not equal!"
1340
+ },
1341
+ {
1342
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
1343
+ "trace": "(line 715) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']"
1344
+ }
1345
+ ],
1346
+ "single": [
1347
+ {
1348
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
1349
+ "trace": "(line 715) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']"
1350
+ }
1351
+ ]
1352
+ },
1353
+ "job_link": {
1354
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410397",
1355
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409587"
1356
+ }
1357
+ },
1358
+ "models_smolvlm": {
1359
+ "failed": {
1360
+ "PyTorch": {
1361
+ "unclassified": 0,
1362
+ "single": 1,
1363
+ "multi": 1
1364
+ },
1365
+ "TensorFlow": {
1366
+ "unclassified": 0,
1367
+ "single": 0,
1368
+ "multi": 0
1369
+ },
1370
+ "Flax": {
1371
+ "unclassified": 0,
1372
+ "single": 0,
1373
+ "multi": 0
1374
+ },
1375
+ "Tokenizers": {
1376
+ "unclassified": 0,
1377
+ "single": 0,
1378
+ "multi": 0
1379
+ },
1380
+ "Pipelines": {
1381
+ "unclassified": 0,
1382
+ "single": 0,
1383
+ "multi": 0
1384
+ },
1385
+ "Trainer": {
1386
+ "unclassified": 0,
1387
+ "single": 0,
1388
+ "multi": 0
1389
+ },
1390
+ "ONNX": {
1391
+ "unclassified": 0,
1392
+ "single": 0,
1393
+ "multi": 0
1394
+ },
1395
+ "Auto": {
1396
+ "unclassified": 0,
1397
+ "single": 0,
1398
+ "multi": 0
1399
+ },
1400
+ "Quantization": {
1401
+ "unclassified": 0,
1402
+ "single": 0,
1403
+ "multi": 0
1404
+ },
1405
+ "Unclassified": {
1406
+ "unclassified": 0,
1407
+ "single": 0,
1408
+ "multi": 0
1409
+ }
1410
+ },
1411
+ "success": 323,
1412
+ "time_spent": "0:02:49, 0:02:35, ",
1413
+ "failures": {
1414
+ "single": [
1415
+ {
1416
+ "line": "tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1417
+ "trace": "(line 4140) KeyError: 'eager'"
1418
+ }
1419
+ ],
1420
+ "multi": [
1421
+ {
1422
+ "line": "tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1423
+ "trace": "(line 4140) KeyError: 'eager'"
1424
+ }
1425
+ ]
1426
+ },
1427
+ "job_link": {
1428
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409653",
1429
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410495"
1430
+ }
1431
+ },
1432
+ "models_t5": {
1433
+ "failed": {
1434
+ "PyTorch": {
1435
+ "unclassified": 0,
1436
+ "single": 3,
1437
+ "multi": 4
1438
+ },
1439
+ "TensorFlow": {
1440
+ "unclassified": 0,
1441
+ "single": 0,
1442
+ "multi": 0
1443
+ },
1444
+ "Flax": {
1445
+ "unclassified": 0,
1446
+ "single": 0,
1447
+ "multi": 0
1448
+ },
1449
+ "Tokenizers": {
1450
+ "unclassified": 0,
1451
+ "single": 0,
1452
+ "multi": 0
1453
+ },
1454
+ "Pipelines": {
1455
+ "unclassified": 0,
1456
+ "single": 0,
1457
+ "multi": 0
1458
+ },
1459
+ "Trainer": {
1460
+ "unclassified": 0,
1461
+ "single": 0,
1462
+ "multi": 0
1463
+ },
1464
+ "ONNX": {
1465
+ "unclassified": 0,
1466
+ "single": 0,
1467
+ "multi": 0
1468
+ },
1469
+ "Auto": {
1470
+ "unclassified": 0,
1471
+ "single": 0,
1472
+ "multi": 0
1473
+ },
1474
+ "Quantization": {
1475
+ "unclassified": 0,
1476
+ "single": 0,
1477
+ "multi": 0
1478
+ },
1479
+ "Unclassified": {
1480
+ "unclassified": 0,
1481
+ "single": 0,
1482
+ "multi": 0
1483
+ }
1484
+ },
1485
+ "success": 254,
1486
+ "time_spent": "0:05:05, 0:03:30, ",
1487
+ "failures": {
1488
+ "multi": [
1489
+ {
1490
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1491
+ "trace": "(line 4140) KeyError: 'eager'"
1492
+ },
1493
+ {
1494
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelTest::test_multi_gpu_data_parallel_forward",
1495
+ "trace": "(line 131) TypeError: EncoderDecoderCache.__init__() missing 1 required positional argument: 'cross_attention_cache'"
1496
+ },
1497
+ {
1498
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization",
1499
+ "trace": "(line 687) AttributeError: 'dict' object has no attribute 'batch_size'"
1500
+ },
1501
+ {
1502
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_small_integration_test",
1503
+ "trace": "(line 727) AssertionError: False is not true"
1504
+ }
1505
+ ],
1506
+ "single": [
1507
+ {
1508
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1509
+ "trace": "(line 4140) KeyError: 'eager'"
1510
+ },
1511
+ {
1512
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization",
1513
+ "trace": "(line 687) AttributeError: 'dict' object has no attribute 'batch_size'"
1514
+ },
1515
+ {
1516
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_small_integration_test",
1517
+ "trace": "(line 727) AssertionError: False is not true"
1518
+ }
1519
+ ]
1520
+ },
1521
+ "job_link": {
1522
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410524",
1523
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409705"
1524
+ }
1525
+ },
1526
+ "models_vit": {
1527
+ "failed": {
1528
+ "PyTorch": {
1529
+ "unclassified": 0,
1530
+ "single": 0,
1531
+ "multi": 0
1532
+ },
1533
+ "TensorFlow": {
1534
+ "unclassified": 0,
1535
+ "single": 0,
1536
+ "multi": 0
1537
+ },
1538
+ "Flax": {
1539
+ "unclassified": 0,
1540
+ "single": 0,
1541
+ "multi": 0
1542
+ },
1543
+ "Tokenizers": {
1544
+ "unclassified": 0,
1545
+ "single": 0,
1546
+ "multi": 0
1547
+ },
1548
+ "Pipelines": {
1549
+ "unclassified": 0,
1550
+ "single": 0,
1551
+ "multi": 0
1552
+ },
1553
+ "Trainer": {
1554
+ "unclassified": 0,
1555
+ "single": 0,
1556
+ "multi": 0
1557
+ },
1558
+ "ONNX": {
1559
+ "unclassified": 0,
1560
+ "single": 0,
1561
+ "multi": 0
1562
+ },
1563
+ "Auto": {
1564
+ "unclassified": 0,
1565
+ "single": 0,
1566
+ "multi": 0
1567
+ },
1568
+ "Quantization": {
1569
+ "unclassified": 0,
1570
+ "single": 0,
1571
+ "multi": 0
1572
+ },
1573
+ "Unclassified": {
1574
+ "unclassified": 0,
1575
+ "single": 0,
1576
+ "multi": 0
1577
+ }
1578
+ },
1579
+ "success": 135,
1580
+ "time_spent": "0:02:19, 0:01:21, ",
1581
+ "failures": {},
1582
+ "job_link": {
1583
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410589",
1584
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409755"
1585
+ }
1586
+ },
1587
+ "models_wav2vec2": {
1588
+ "failed": {
1589
+ "PyTorch": {
1590
+ "unclassified": 0,
1591
+ "single": 0,
1592
+ "multi": 0
1593
+ },
1594
+ "TensorFlow": {
1595
+ "unclassified": 0,
1596
+ "single": 0,
1597
+ "multi": 0
1598
+ },
1599
+ "Flax": {
1600
+ "unclassified": 0,
1601
+ "single": 0,
1602
+ "multi": 0
1603
+ },
1604
+ "Tokenizers": {
1605
+ "unclassified": 0,
1606
+ "single": 0,
1607
+ "multi": 0
1608
+ },
1609
+ "Pipelines": {
1610
+ "unclassified": 0,
1611
+ "single": 0,
1612
+ "multi": 0
1613
+ },
1614
+ "Trainer": {
1615
+ "unclassified": 0,
1616
+ "single": 0,
1617
+ "multi": 0
1618
+ },
1619
+ "ONNX": {
1620
+ "unclassified": 0,
1621
+ "single": 0,
1622
+ "multi": 0
1623
+ },
1624
+ "Auto": {
1625
+ "unclassified": 0,
1626
+ "single": 0,
1627
+ "multi": 0
1628
+ },
1629
+ "Quantization": {
1630
+ "unclassified": 0,
1631
+ "single": 0,
1632
+ "multi": 0
1633
+ },
1634
+ "Unclassified": {
1635
+ "unclassified": 0,
1636
+ "single": 0,
1637
+ "multi": 0
1638
+ }
1639
+ },
1640
+ "success": 0,
1641
+ "time_spent": "0.96, .03, ",
1642
+ "failures": {},
1643
+ "job_link": {
1644
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410594",
1645
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409797"
1646
+ }
1647
+ },
1648
+ "models_whisper": {
1649
+ "failed": {
1650
+ "PyTorch": {
1651
+ "unclassified": 0,
1652
+ "single": 0,
1653
+ "multi": 0
1654
+ },
1655
+ "TensorFlow": {
1656
+ "unclassified": 0,
1657
+ "single": 0,
1658
+ "multi": 0
1659
+ },
1660
+ "Flax": {
1661
+ "unclassified": 0,
1662
+ "single": 0,
1663
+ "multi": 0
1664
+ },
1665
+ "Tokenizers": {
1666
+ "unclassified": 0,
1667
+ "single": 0,
1668
+ "multi": 0
1669
+ },
1670
+ "Pipelines": {
1671
+ "unclassified": 0,
1672
+ "single": 0,
1673
+ "multi": 0
1674
+ },
1675
+ "Trainer": {
1676
+ "unclassified": 0,
1677
+ "single": 0,
1678
+ "multi": 0
1679
+ },
1680
+ "ONNX": {
1681
+ "unclassified": 0,
1682
+ "single": 0,
1683
+ "multi": 0
1684
+ },
1685
+ "Auto": {
1686
+ "unclassified": 0,
1687
+ "single": 0,
1688
+ "multi": 0
1689
+ },
1690
+ "Quantization": {
1691
+ "unclassified": 0,
1692
+ "single": 0,
1693
+ "multi": 0
1694
+ },
1695
+ "Unclassified": {
1696
+ "unclassified": 0,
1697
+ "single": 0,
1698
+ "multi": 0
1699
+ }
1700
+ },
1701
+ "success": 0,
1702
+ "time_spent": ".19, .20, ",
1703
+ "failures": {},
1704
+ "job_link": {
1705
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527409794",
1706
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460430974/job/46527410606"
1707
+ }
1708
+ }
1709
+ }
sample_data.csv DELETED
@@ -1,22 +0,0 @@
1
- model_name,success_amd,success_nvidia,failed_multi_no_amd,failed_multi_no_nvidia,failed_single_no_amd,failed_single_no_nvidia,failures_amd,failures_nvidia,job_link_amd,job_link_nvidia
2
- sample_auto,80,226,0,0,0,0,{},{},"{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501262', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500785'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561673', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561472'}"
3
- sample_bert,239,527,2,2,2,2,"{'multi': [{'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4201) AssertionError: Tensor-likes are not equal!'}], 'single': [{'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4201) AssertionError: Tensor-likes are not equal!'}]}","{'single': [{'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4216) AssertionError: Tensor-likes are not equal!'}], 'multi': [{'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4216) AssertionError: Tensor-likes are not equal!'}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501282', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500788'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561709', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561482'}"
4
- sample_clip,288,660,0,0,0,0,{},{},"{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500866', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501323'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561994', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562125'}"
5
- sample_detr,69,177,4,0,4,0,"{'multi': [{'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTestsTimmBackbone::test_inference_no_head', 'trace': '(line 595) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTestsTimmBackbone::test_inference_object_detection_head', 'trace': '(line 619) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTestsTimmBackbone::test_inference_panoptic_segmentation_head', 'trace': '(line 667) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTests::test_inference_no_head', 'trace': '(line 741) AssertionError: Tensor-likes are not close!'}], 'single': [{'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTestsTimmBackbone::test_inference_no_head', 'trace': '(line 595) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTestsTimmBackbone::test_inference_object_detection_head', 'trace': '(line 619) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTestsTimmBackbone::test_inference_panoptic_segmentation_head', 'trace': '(line 667) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/detr/test_modeling_detr.py::DetrModelIntegrationTests::test_inference_no_head', 'trace': '(line 741) AssertionError: Tensor-likes are not close!'}]}",{},"{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501397', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500969'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562517', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562397'}"
6
- sample_gemma3,349,499,8,8,7,7,"{'single': [{'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch', 'trace': ""(line 675) AssertionError: Lists differ: ['use[374 chars]t scenes:\\n\\n* **Image 1** shows a cow on a beach.\\n'] != ['use[374 chars]t scenes. \\n\\n* **Image 1** shows a cow standing on a beach']""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops', 'trace': ""(line 675) AssertionError: Lists differ: ['use[251 chars]. The sky is blue with some white clouds. It’s[405 chars]h a'] != ['use[251 chars]. There are clouds in the blue sky above.', 'u[398 chars]h a']""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_bf16', 'trace': ""(line 675) AssertionError: Lists differ: ['use[154 chars]each next to a turquoise ocean. There are some[16 chars]lue'] != ['use[154 chars]each with turquoise water and a distant coastl[28 chars]oks']""}], 'multi': [{'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4204) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch', 'trace': ""(line 675) AssertionError: Lists differ: ['use[374 chars]t scenes:\\n\\n* **Image 1** shows a cow on a beach.\\n'] != ['use[374 chars]t scenes. \\n\\n* **Image 1** shows a cow standing on a beach']""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops', 'trace': ""(line 675) AssertionError: Lists differ: ['use[251 chars]. The sky is blue with some white clouds. It’s[405 chars]h a'] != ['use[251 chars]. There are clouds in the blue sky above.', 'u[398 chars]h a']""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_bf16', 'trace': ""(line 675) AssertionError: Lists differ: ['use[154 chars]each next to a turquoise ocean. There are some[16 chars]lue'] != ['use[154 chars]each with turquoise water and a distant coastl[28 chars]oks']""}]}","{'single': [{'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4216) AssertionError: Tensor-likes are not equal!'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only_with_hybrid_cache', 'trace': ""(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., size=(1, 4, 1, 256), grad_fn=<AddBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>)), **{'attn_mask': FakeTensor(..., size=(1, 1, 1, 512), dtype=torch.bool), 'dropout_p': 0.0, 'scale': 0.0625, 'is_causal': False}): got RuntimeError('Attempting to broadcast a dimension of length 512 at -1! Mismatching argument at index 1 had torch.Size([1, 1, 1, 512]); but expected shape should be broadcastable to [1, 4, 1, 4096]')""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa', 'trace': '(line 81) RuntimeError: The expanded size of the tensor (4826) must match the existing size (4807) at non-singleton dimension 3. Target sizes: [2, 4, 4807, 4826]. Tensor sizes: [2, 1, 4807, 4807]'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager', 'trace': '(line 265) RuntimeError: The size of tensor a (4826) must match the size of tensor b (4807) at non-singleton dimension 3'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops', 'trace': '(line 81) RuntimeError: The expanded size of the tensor (1646) must match the existing size (1617) at non-singleton dimension 3. Target sizes: [2, 8, 1617, 1646]. Tensor sizes: [2, 1, 1617, 1617]'}], 'multi': [{'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4219) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_model_parallelism', 'trace': '(line 925) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0!'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only_with_hybrid_cache', 'trace': ""(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., size=(1, 4, 1, 256), grad_fn=<AddBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>)), **{'attn_mask': FakeTensor(..., size=(1, 1, 1, 512), dtype=torch.bool), 'dropout_p': 0.0, 'scale': 0.0625, 'is_causal': False}): got RuntimeError('Attempting to broadcast a dimension of length 512 at -1! Mismatching argument at index 1 had torch.Size([1, 1, 1, 512]); but expected shape should be broadcastable to [1, 4, 1, 4096]')""}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa', 'trace': '(line 81) RuntimeError: The expanded size of the tensor (4826) must match the existing size (4807) at non-singleton dimension 3. Target sizes: [2, 4, 4807, 4826]. Tensor sizes: [2, 1, 4807, 4807]'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager', 'trace': '(line 265) RuntimeError: The size of tensor a (4826) must match the size of tensor b (4807) at non-singleton dimension 3'}, {'line': 'tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops', 'trace': '(line 81) RuntimeError: The expanded size of the tensor (1646) must match the existing size (1617) at non-singleton dimension 3. Target sizes: [2, 8, 1617, 1646]. Tensor sizes: [2, 1, 1617, 1617]'}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501046', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501545'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563053', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562857'}"
7
- sample_gemma3n,0,286,0,2,0,1,{},"{'multi': [{'line': 'tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}], 'single': [{'line': 'tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501047', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501538'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562955', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563061'}"
8
- sample_got_ocr2,145,254,2,2,2,1,"{'multi': [{'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}], 'single': [{'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}]}","{'multi': [{'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}], 'single': [{'line': 'tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501556', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501063'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562995', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563212'}"
9
- sample_gpt2,249,487,1,1,1,1,"{'single': [{'line': 'tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}], 'multi': [{'line': 'tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}]}","{'multi': [{'line': 'tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}], 'single': [{'line': 'tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501087', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501566'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563001', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563255'}"
10
- sample_internvl,249,356,4,3,4,2,"{'single': [{'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_forward', 'trace': '(line 687) AssertionError: False is not true : Actual logits: tensor([ -9.8828, -0.5005, 1.4697, -10.3438, -10.3438], dtype=torch.float16)'}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_interleaved_images_videos', 'trace': ""(line 675) AssertionError: 'user[118 chars]nse. Upon closer inspection, the differences b[31 chars]. **' != 'user[118 chars]nse. After re-examining the images, I can see [13 chars]e no'""}], 'multi': [{'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_forward', 'trace': '(line 687) AssertionError: False is not true : Actual logits: tensor([ -9.8828, -0.5005, 1.4697, -10.3438, -10.3438], dtype=torch.float16)'}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLLlamaIntegrationTest::test_llama_small_model_integration_interleaved_images_videos', 'trace': ""(line 675) AssertionError: 'user[118 chars]nse. Upon closer inspection, the differences b[31 chars]. **' != 'user[118 chars]nse. After re-examining the images, I can see [13 chars]e no'""}]}","{'multi': [{'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_flex_attention_with_grads', 'trace': '(line 439) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfResources: out of resource: shared memory, Required: 106496, Hardware limit: 101376. Reducing block sizes or `num_stages` may help.'}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}], 'single': [{'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_flex_attention_with_grads', 'trace': '(line 439) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfResources: out of resource: shared memory, Required: 106496, Hardware limit: 101376. Reducing block sizes or `num_stages` may help.'}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501143', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501636'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563553', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563712'}"
11
- sample_llama,229,478,4,2,4,1,"{'multi': [{'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_torch_compile_for_training', 'trace': '(line 951) AssertionError: expected size 2==2, stride 20==64 at dim=0; expected size 2==2, stride 10==32 at dim=1; expected size 10==32, stride 1==1 at dim=2'}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaIntegrationTest::test_model_7b_logits_bf16', 'trace': '(line 687) AssertionError: False is not true'}], 'single': [{'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_torch_compile_for_training', 'trace': '(line 951) AssertionError: expected size 2==2, stride 20==64 at dim=0; expected size 2==2, stride 10==32 at dim=1; expected size 10==32, stride 1==1 at dim=2'}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaIntegrationTest::test_model_7b_logits_bf16', 'trace': '(line 687) AssertionError: False is not true'}]}","{'multi': [{'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}], 'single': [{'line': 'tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501675', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501165'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563871', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526564103'}"
12
- sample_llava,201,346,5,4,4,3,"{'single': [{'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads', 'trace': '(line 687) AssertionError: False is not true'}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation', 'trace': '(line 548) importlib.metadata.PackageNotFoundError: No package metadata was found for bitsandbytes'}], 'multi': [{'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads', 'trace': '(line 687) AssertionError: False is not true'}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4182) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1'}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation', 'trace': '(line 548) importlib.metadata.PackageNotFoundError: No package metadata was found for bitsandbytes'}]}","{'multi': [{'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads', 'trace': '(line 687) AssertionError: False is not true'}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4197) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1'}], 'single': [{'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads', 'trace': '(line 687) AssertionError: False is not true'}, {'line': 'tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids', 'trace': '(line 4197) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1'}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501186', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447501727'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526564002', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526564108'}"
13
- sample_mistral3,197,286,3,2,3,1,"{'multi': [{'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3IntegrationTest::test_mistral3_integration_batched_generate', 'trace': '(line 675) AssertionError: \'Calm waters reflect\\nWooden path to distant shore\\nSilence in the scene\' != ""Wooden path to calm,\\nReflections whisper secrets,\\nNature\'s peace unfolds.""'}], 'single': [{'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3IntegrationTest::test_mistral3_integration_batched_generate', 'trace': '(line 675) AssertionError: \'Calm waters reflect\\nWooden path to distant shore\\nSilence in the scene\' != ""Wooden path to calm,\\nReflections whisper secrets,\\nNature\'s peace unfolds.""'}]}","{'single': [{'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}], 'multi': [{'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500305', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447499780'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561480', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561618'}"
14
- sample_modernbert,132,164,5,5,5,5,"{'single': [{'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export', 'trace': ""(line 675) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']""}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm', 'trace': '(line 401) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head', 'trace': '(line 423) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification', 'trace': '(line 469) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification', 'trace': '(line 446) AssertionError: Tensor-likes are not close!'}], 'multi': [{'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export', 'trace': ""(line 675) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']""}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm', 'trace': '(line 401) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head', 'trace': '(line 423) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification', 'trace': '(line 469) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification', 'trace': '(line 446) AssertionError: Tensor-likes are not close!'}]}","{'multi': [{'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export', 'trace': ""(line 675) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']""}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm', 'trace': '(line 401) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head', 'trace': '(line 423) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification', 'trace': '(line 469) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification', 'trace': '(line 446) AssertionError: Tensor-likes are not close!'}], 'single': [{'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export', 'trace': ""(line 675) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']""}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm', 'trace': '(line 401) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head', 'trace': '(line 423) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification', 'trace': '(line 469) AssertionError: Tensor-likes are not close!'}, {'line': 'tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification', 'trace': '(line 446) AssertionError: Tensor-likes are not close!'}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447499811', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500326'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561668', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561515'}"
15
- sample_qwen2,213,438,3,3,3,2,"{'multi': [{'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache', 'trace': ""(line 1638) torch._dynamo.exc.TorchRuntimeError: Failed running call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,""}], 'single': [{'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_generate_compilation_all_outputs', 'trace': ""(line 317) torch._dynamo.exc.Unsupported: isinstance(NestedUserFunctionVariable(), TorchInGraphFunctionVariable(<class 'torch.nn.parameter.Parameter'>)): can't determine type of NestedUserFunctionVariable()""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache', 'trace': ""(line 1638) torch._dynamo.exc.TorchRuntimeError: Failed running call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,""}]}","{'multi': [{'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache', 'trace': ""(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,""}], 'single': [{'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache', 'trace': ""(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500458', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447499989'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562376', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562270'}"
16
- sample_qwen2_5_omni,168,277,2,5,1,1,"{'single': [{'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch', 'trace': '(line 675) AssertionError: Lists differ: [""sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.""] != [""sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.""]'}], 'multi': [{'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_model_parallelism', 'trace': '(line 675) AssertionError: Items in the second set but not the first:'}, {'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch', 'trace': '(line 675) AssertionError: Lists differ: [""sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.""] != [""sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.""]'}]}","{'multi': [{'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_model_parallelism', 'trace': '(line 675) AssertionError: Items in the second set but not the first:'}, {'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}, {'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch', 'trace': '(line 675) AssertionError: Lists differ: [""sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.""] != [""sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.""]'}, {'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_multiturn', 'trace': '(line 849) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU 1 has a total capacity of 22.18 GiB of which 6.50 MiB is free. Process 51940 has 22.17 GiB memory in use. Of the allocated memory 21.74 GiB is allocated by PyTorch, and 27.83 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)'}, {'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_w_audio', 'trace': '(line 1000) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU 1 has a total capacity of 22.18 GiB of which 8.50 MiB is free. Process 51940 has 22.17 GiB memory in use. Of the allocated memory 21.75 GiB is allocated by PyTorch, and 17.78 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)'}], 'single': [{'line': 'tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch', 'trace': '(line 675) AssertionError: Lists differ: [""sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.""] != [""sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.""]'}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447499993', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500491'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562375', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562289'}"
17
- sample_qwen2_5_vl,204,311,1,1,2,1,"{'single': [{'line': 'tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test', 'trace': ""(line 700) requests.exceptions.ConnectionError: HTTPSConnectionPool(host='qianwen-res.oss-accelerate-overseas.aliyuncs.com', port=443): Max retries exceeded with url: /Qwen2-VL/demo_small.jpg (Caused by NewConnectionError('<urllib3.connection.HTTPSConnection object at 0x7b289312aad0>: Failed to establish a new connection: [Errno -2] Name or service not known'))""}, {'line': 'tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions', 'trace': ""(line 675) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']""}], 'multi': [{'line': 'tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions', 'trace': ""(line 675) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']""}]}","{'multi': [{'line': 'tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions', 'trace': ""(line 675) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']""}], 'single': [{'line': 'tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions', 'trace': ""(line 675) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']""}]}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447499984', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500447'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562382', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562290'}"
18
- sample_smolvlm,323,499,1,1,1,1,"{'multi': [{'line': 'tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}], 'single': [{'line': 'tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}]}","{'single': [{'line': 'tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}], 'multi': [{'line': 'tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500533', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500052'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562675', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562798'}"
19
- sample_t5,254,592,4,3,3,2,"{'multi': [{'line': 'tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 130) TypeError: EncoderDecoderCache.__init__() missing 1 required positional argument: 'cross_attention_cache'""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization', 'trace': ""(line 885) torch._dynamo.exc.TorchRuntimeError: Failed running call_function <built-in function add>(*(FakeTensor(..., size=(1, 8, 1, 1234)), FakeTensor(..., device='cuda:1', size=(1, 1, 1, 1234))), **{}):""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_small_integration_test', 'trace': '(line 687) AssertionError: False is not true'}], 'single': [{'line': 'tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4125) KeyError: 'eager'""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization', 'trace': ""(line 885) torch._dynamo.exc.TorchRuntimeError: Failed running call_function <built-in function add>(*(FakeTensor(..., size=(1, 8, 1, 1234)), FakeTensor(..., device='cuda:0', size=(1, 1, 1, 1234))), **{}):""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_small_integration_test', 'trace': '(line 687) AssertionError: False is not true'}]}","{'multi': [{'line': 'tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 131) TypeError: EncoderDecoderCache.__init__() missing 1 required positional argument: 'cross_attention_cache'""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization', 'trace': ""(line 687) AttributeError: 'dict' object has no attribute 'batch_size'""}], 'single': [{'line': 'tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization', 'trace': ""(line 687) AttributeError: 'dict' object has no attribute 'batch_size'""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500560', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500103'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563047', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562939'}"
20
- sample_vit,135,217,0,0,0,0,{},{},"{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500654', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500177'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563537', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563397'}"
21
- sample_wav2vec2,0,672,0,4,0,4,{},"{'multi': [{'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_inference_mms_1b_all', 'trace': '(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py'}, {'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm', 'trace': '(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py'}, {'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_invalid_pool', 'trace': '(line 675) AssertionError: Traceback (most recent call last):'}, {'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_pool', 'trace': '(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py'}], 'single': [{'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_inference_mms_1b_all', 'trace': '(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py'}, {'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm', 'trace': '(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py'}, {'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_invalid_pool', 'trace': '(line 675) AssertionError: Traceback (most recent call last):'}, {'line': 'tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_pool', 'trace': '(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py'}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500676', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500194'}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563711', 'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563582'}"
22
- sample_whisper,0,1010,0,11,0,8,{},"{'single': [{'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_large_batched_generation_multilingual', 'trace': '(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_small_longform_timestamps_generation', 'trace': '(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_tiny_longform_timestamps_generation', 'trace': '(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard', 'trace': '(line 675) AssertionError: Lists differ: ["" Fo[272 chars]ting of classics, Sicilian, nade door variatio[8147 chars]le!\'] != ["" Fo[272 chars]ting a classic Sicilian, nade door variation o[8150 chars]le!\']'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard_prev_cond', 'trace': '(line 675) AssertionError: Lists differ: ["" Fo[422 chars]to a fisher shows in lip-nitsky attack that cu[7903 chars]le!""] != ["" Fo[422 chars]to a Fisher shows in lip-nitsky attack that cu[7918 chars]le.""]'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_shortform_single_batch_prev_cond', 'trace': '(line 675) AssertionError: Lists differ: ["" Fo[268 chars]ating, so soft, it would make JD power and her[196 chars]ke.""] != ["" Fo[268 chars]ating so soft, it would make JD power and her [195 chars]ke.""]'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperStandaloneDecoderModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}], 'multi': [{'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 131) TypeError: EncoderDecoderCache.__init__() missing 1 required positional argument: 'cross_attention_cache'""}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_generate_with_forced_decoder_ids', 'trace': '(line 713) requests.exceptions.ReadTimeout: (ReadTimeoutError(""HTTPSConnectionPool(host=\'huggingface.co\', port=443): Read timed out. (read timeout=10)""), \'(Request ID: 13cb0b08-c261-4ca3-a58f-91a2f3e327ed)\')'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_large_batched_generation_multilingual', 'trace': '(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_small_longform_timestamps_generation', 'trace': '(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_tiny_longform_timestamps_generation', 'trace': '(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard', 'trace': '(line 675) AssertionError: Lists differ: ["" Fo[272 chars]ting of classics, Sicilian, nade door variatio[8147 chars]le!\'] != ["" Fo[272 chars]ting a classic Sicilian, nade door variation o[8150 chars]le!\']'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard_prev_cond', 'trace': '(line 675) AssertionError: Lists differ: ["" Fo[422 chars]to a fisher shows in lip-nitsky attack that cu[7903 chars]le!""] != ["" Fo[422 chars]to a Fisher shows in lip-nitsky attack that cu[7918 chars]le.""]'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_shortform_single_batch_prev_cond', 'trace': '(line 675) AssertionError: Lists differ: ["" Fo[268 chars]ating, so soft, it would make JD power and her[196 chars]ke.""] != ["" Fo[268 chars]ating so soft, it would make JD power and her [195 chars]ke.""]'}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperStandaloneDecoderModelTest::test_eager_padding_matches_padding_free_with_position_ids', 'trace': ""(line 4140) KeyError: 'eager'""}, {'line': 'tests/models/whisper/test_modeling_whisper.py::WhisperStandaloneDecoderModelTest::test_multi_gpu_data_parallel_forward', 'trace': ""(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'""}]}","{'multi': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500690', 'single': 'https://github.com/huggingface/transformers/actions/runs/16433423306/job/46447500204'}","{'single': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563737', 'multi': 'https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563862'}"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
sample_nvidia.json ADDED
@@ -0,0 +1,1776 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_auto": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 0
8
+ },
9
+ "TensorFlow": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Flax": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Tokenizers": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "Pipelines": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Trainer": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "ONNX": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Auto": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ },
44
+ "Quantization": {
45
+ "unclassified": 0,
46
+ "single": 0,
47
+ "multi": 0
48
+ },
49
+ "Unclassified": {
50
+ "unclassified": 0,
51
+ "single": 0,
52
+ "multi": 0
53
+ }
54
+ },
55
+ "success": 226,
56
+ "time_spent": "4.66, 6.10, ",
57
+ "failures": {},
58
+ "job_link": {
59
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561673",
60
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561472"
61
+ }
62
+ },
63
+ "models_bert": {
64
+ "failed": {
65
+ "PyTorch": {
66
+ "unclassified": 0,
67
+ "single": 2,
68
+ "multi": 2
69
+ },
70
+ "TensorFlow": {
71
+ "unclassified": 0,
72
+ "single": 0,
73
+ "multi": 0
74
+ },
75
+ "Flax": {
76
+ "unclassified": 0,
77
+ "single": 0,
78
+ "multi": 0
79
+ },
80
+ "Tokenizers": {
81
+ "unclassified": 0,
82
+ "single": 0,
83
+ "multi": 0
84
+ },
85
+ "Pipelines": {
86
+ "unclassified": 0,
87
+ "single": 0,
88
+ "multi": 0
89
+ },
90
+ "Trainer": {
91
+ "unclassified": 0,
92
+ "single": 0,
93
+ "multi": 0
94
+ },
95
+ "ONNX": {
96
+ "unclassified": 0,
97
+ "single": 0,
98
+ "multi": 0
99
+ },
100
+ "Auto": {
101
+ "unclassified": 0,
102
+ "single": 0,
103
+ "multi": 0
104
+ },
105
+ "Quantization": {
106
+ "unclassified": 0,
107
+ "single": 0,
108
+ "multi": 0
109
+ },
110
+ "Unclassified": {
111
+ "unclassified": 0,
112
+ "single": 0,
113
+ "multi": 0
114
+ }
115
+ },
116
+ "success": 527,
117
+ "time_spent": "0:01:58, 0:02:00, ",
118
+ "failures": {
119
+ "single": [
120
+ {
121
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids",
122
+ "trace": "(line 4140) KeyError: 'eager'"
123
+ },
124
+ {
125
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
126
+ "trace": "(line 4216) AssertionError: Tensor-likes are not equal!"
127
+ }
128
+ ],
129
+ "multi": [
130
+ {
131
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_eager_padding_matches_padding_free_with_position_ids",
132
+ "trace": "(line 4140) KeyError: 'eager'"
133
+ },
134
+ {
135
+ "line": "tests/models/bert/test_modeling_bert.py::BertModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
136
+ "trace": "(line 4216) AssertionError: Tensor-likes are not equal!"
137
+ }
138
+ ]
139
+ },
140
+ "job_link": {
141
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561709",
142
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561482"
143
+ }
144
+ },
145
+ "models_clip": {
146
+ "failed": {
147
+ "PyTorch": {
148
+ "unclassified": 0,
149
+ "single": 0,
150
+ "multi": 0
151
+ },
152
+ "TensorFlow": {
153
+ "unclassified": 0,
154
+ "single": 0,
155
+ "multi": 0
156
+ },
157
+ "Flax": {
158
+ "unclassified": 0,
159
+ "single": 0,
160
+ "multi": 0
161
+ },
162
+ "Tokenizers": {
163
+ "unclassified": 0,
164
+ "single": 0,
165
+ "multi": 0
166
+ },
167
+ "Pipelines": {
168
+ "unclassified": 0,
169
+ "single": 0,
170
+ "multi": 0
171
+ },
172
+ "Trainer": {
173
+ "unclassified": 0,
174
+ "single": 0,
175
+ "multi": 0
176
+ },
177
+ "ONNX": {
178
+ "unclassified": 0,
179
+ "single": 0,
180
+ "multi": 0
181
+ },
182
+ "Auto": {
183
+ "unclassified": 0,
184
+ "single": 0,
185
+ "multi": 0
186
+ },
187
+ "Quantization": {
188
+ "unclassified": 0,
189
+ "single": 0,
190
+ "multi": 0
191
+ },
192
+ "Unclassified": {
193
+ "unclassified": 0,
194
+ "single": 0,
195
+ "multi": 0
196
+ }
197
+ },
198
+ "success": 660,
199
+ "time_spent": "0:02:24, 0:02:20, ",
200
+ "failures": {},
201
+ "job_link": {
202
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561994",
203
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562125"
204
+ }
205
+ },
206
+ "models_detr": {
207
+ "failed": {
208
+ "PyTorch": {
209
+ "unclassified": 0,
210
+ "single": 0,
211
+ "multi": 0
212
+ },
213
+ "TensorFlow": {
214
+ "unclassified": 0,
215
+ "single": 0,
216
+ "multi": 0
217
+ },
218
+ "Flax": {
219
+ "unclassified": 0,
220
+ "single": 0,
221
+ "multi": 0
222
+ },
223
+ "Tokenizers": {
224
+ "unclassified": 0,
225
+ "single": 0,
226
+ "multi": 0
227
+ },
228
+ "Pipelines": {
229
+ "unclassified": 0,
230
+ "single": 0,
231
+ "multi": 0
232
+ },
233
+ "Trainer": {
234
+ "unclassified": 0,
235
+ "single": 0,
236
+ "multi": 0
237
+ },
238
+ "ONNX": {
239
+ "unclassified": 0,
240
+ "single": 0,
241
+ "multi": 0
242
+ },
243
+ "Auto": {
244
+ "unclassified": 0,
245
+ "single": 0,
246
+ "multi": 0
247
+ },
248
+ "Quantization": {
249
+ "unclassified": 0,
250
+ "single": 0,
251
+ "multi": 0
252
+ },
253
+ "Unclassified": {
254
+ "unclassified": 0,
255
+ "single": 0,
256
+ "multi": 0
257
+ }
258
+ },
259
+ "success": 177,
260
+ "time_spent": "0:01:14, 0:01:19, ",
261
+ "failures": {},
262
+ "job_link": {
263
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562517",
264
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562397"
265
+ }
266
+ },
267
+ "models_gemma3": {
268
+ "failed": {
269
+ "PyTorch": {
270
+ "unclassified": 0,
271
+ "single": 7,
272
+ "multi": 8
273
+ },
274
+ "TensorFlow": {
275
+ "unclassified": 0,
276
+ "single": 0,
277
+ "multi": 0
278
+ },
279
+ "Flax": {
280
+ "unclassified": 0,
281
+ "single": 0,
282
+ "multi": 0
283
+ },
284
+ "Tokenizers": {
285
+ "unclassified": 0,
286
+ "single": 0,
287
+ "multi": 0
288
+ },
289
+ "Pipelines": {
290
+ "unclassified": 0,
291
+ "single": 0,
292
+ "multi": 0
293
+ },
294
+ "Trainer": {
295
+ "unclassified": 0,
296
+ "single": 0,
297
+ "multi": 0
298
+ },
299
+ "ONNX": {
300
+ "unclassified": 0,
301
+ "single": 0,
302
+ "multi": 0
303
+ },
304
+ "Auto": {
305
+ "unclassified": 0,
306
+ "single": 0,
307
+ "multi": 0
308
+ },
309
+ "Quantization": {
310
+ "unclassified": 0,
311
+ "single": 0,
312
+ "multi": 0
313
+ },
314
+ "Unclassified": {
315
+ "unclassified": 0,
316
+ "single": 0,
317
+ "multi": 0
318
+ }
319
+ },
320
+ "success": 499,
321
+ "time_spent": "0:07:50, 0:07:52, ",
322
+ "failures": {
323
+ "single": [
324
+ {
325
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
326
+ "trace": "(line 4140) KeyError: 'eager'"
327
+ },
328
+ {
329
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
330
+ "trace": "(line 4216) AssertionError: Tensor-likes are not equal!"
331
+ },
332
+ {
333
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids",
334
+ "trace": "(line 4140) KeyError: 'eager'"
335
+ },
336
+ {
337
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only_with_hybrid_cache",
338
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., size=(1, 4, 1, 256), grad_fn=<AddBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>)), **{'attn_mask': FakeTensor(..., size=(1, 1, 1, 512), dtype=torch.bool), 'dropout_p': 0.0, 'scale': 0.0625, 'is_causal': False}): got RuntimeError('Attempting to broadcast a dimension of length 512 at -1! Mismatching argument at index 1 had torch.Size([1, 1, 1, 512]); but expected shape should be broadcastable to [1, 4, 1, 4096]')"
339
+ },
340
+ {
341
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa",
342
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (4826) must match the existing size (4807) at non-singleton dimension 3. Target sizes: [2, 4, 4807, 4826]. Tensor sizes: [2, 1, 4807, 4807]"
343
+ },
344
+ {
345
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager",
346
+ "trace": "(line 265) RuntimeError: The size of tensor a (4826) must match the size of tensor b (4807) at non-singleton dimension 3"
347
+ },
348
+ {
349
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops",
350
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (1646) must match the existing size (1617) at non-singleton dimension 3. Target sizes: [2, 8, 1617, 1646]. Tensor sizes: [2, 1, 1617, 1617]"
351
+ }
352
+ ],
353
+ "multi": [
354
+ {
355
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
356
+ "trace": "(line 4140) KeyError: 'eager'"
357
+ },
358
+ {
359
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3ModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
360
+ "trace": "(line 4219) AssertionError: Tensor-likes are not close!"
361
+ },
362
+ {
363
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_eager_padding_matches_padding_free_with_position_ids",
364
+ "trace": "(line 4140) KeyError: 'eager'"
365
+ },
366
+ {
367
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3Vision2TextModelTest::test_model_parallelism",
368
+ "trace": "(line 925) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0!"
369
+ },
370
+ {
371
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_export_text_only_with_hybrid_cache",
372
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., size=(1, 4, 1, 256), grad_fn=<AddBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>), FakeTensor(..., size=(1, 4, 4096, 256), grad_fn=<CloneBackward0>)), **{'attn_mask': FakeTensor(..., size=(1, 1, 1, 512), dtype=torch.bool), 'dropout_p': 0.0, 'scale': 0.0625, 'is_causal': False}): got RuntimeError('Attempting to broadcast a dimension of length 512 at -1! Mismatching argument at index 1 had torch.Size([1, 1, 1, 512]); but expected shape should be broadcastable to [1, 4, 1, 4096]')"
373
+ },
374
+ {
375
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_1_sdpa",
376
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (4826) must match the existing size (4807) at non-singleton dimension 3. Target sizes: [2, 4, 4807, 4826]. Tensor sizes: [2, 1, 4807, 4807]"
377
+ },
378
+ {
379
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_generation_beyond_sliding_window_2_eager",
380
+ "trace": "(line 265) RuntimeError: The size of tensor a (4826) must match the size of tensor b (4807) at non-singleton dimension 3"
381
+ },
382
+ {
383
+ "line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_model_4b_batch_crops",
384
+ "trace": "(line 81) RuntimeError: The expanded size of the tensor (1646) must match the existing size (1617) at non-singleton dimension 3. Target sizes: [2, 8, 1617, 1646]. Tensor sizes: [2, 1, 1617, 1617]"
385
+ }
386
+ ]
387
+ },
388
+ "job_link": {
389
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563053",
390
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562857"
391
+ }
392
+ },
393
+ "models_gemma3n": {
394
+ "failed": {
395
+ "PyTorch": {
396
+ "unclassified": 0,
397
+ "single": 1,
398
+ "multi": 2
399
+ },
400
+ "TensorFlow": {
401
+ "unclassified": 0,
402
+ "single": 0,
403
+ "multi": 0
404
+ },
405
+ "Flax": {
406
+ "unclassified": 0,
407
+ "single": 0,
408
+ "multi": 0
409
+ },
410
+ "Tokenizers": {
411
+ "unclassified": 0,
412
+ "single": 0,
413
+ "multi": 0
414
+ },
415
+ "Pipelines": {
416
+ "unclassified": 0,
417
+ "single": 0,
418
+ "multi": 0
419
+ },
420
+ "Trainer": {
421
+ "unclassified": 0,
422
+ "single": 0,
423
+ "multi": 0
424
+ },
425
+ "ONNX": {
426
+ "unclassified": 0,
427
+ "single": 0,
428
+ "multi": 0
429
+ },
430
+ "Auto": {
431
+ "unclassified": 0,
432
+ "single": 0,
433
+ "multi": 0
434
+ },
435
+ "Quantization": {
436
+ "unclassified": 0,
437
+ "single": 0,
438
+ "multi": 0
439
+ },
440
+ "Unclassified": {
441
+ "unclassified": 0,
442
+ "single": 0,
443
+ "multi": 0
444
+ }
445
+ },
446
+ "success": 286,
447
+ "time_spent": "0:02:29, 0:02:32, ",
448
+ "failures": {
449
+ "multi": [
450
+ {
451
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_eager_padding_matches_padding_free_with_position_ids",
452
+ "trace": "(line 4140) KeyError: 'eager'"
453
+ },
454
+ {
455
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_multi_gpu_data_parallel_forward",
456
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
457
+ }
458
+ ],
459
+ "single": [
460
+ {
461
+ "line": "tests/models/gemma3n/test_modeling_gemma3n.py::Gemma3nTextModelTest::test_eager_padding_matches_padding_free_with_position_ids",
462
+ "trace": "(line 4140) KeyError: 'eager'"
463
+ }
464
+ ]
465
+ },
466
+ "job_link": {
467
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562955",
468
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563061"
469
+ }
470
+ },
471
+ "models_got_ocr2": {
472
+ "failed": {
473
+ "PyTorch": {
474
+ "unclassified": 0,
475
+ "single": 1,
476
+ "multi": 2
477
+ },
478
+ "TensorFlow": {
479
+ "unclassified": 0,
480
+ "single": 0,
481
+ "multi": 0
482
+ },
483
+ "Flax": {
484
+ "unclassified": 0,
485
+ "single": 0,
486
+ "multi": 0
487
+ },
488
+ "Tokenizers": {
489
+ "unclassified": 0,
490
+ "single": 0,
491
+ "multi": 0
492
+ },
493
+ "Pipelines": {
494
+ "unclassified": 0,
495
+ "single": 0,
496
+ "multi": 0
497
+ },
498
+ "Trainer": {
499
+ "unclassified": 0,
500
+ "single": 0,
501
+ "multi": 0
502
+ },
503
+ "ONNX": {
504
+ "unclassified": 0,
505
+ "single": 0,
506
+ "multi": 0
507
+ },
508
+ "Auto": {
509
+ "unclassified": 0,
510
+ "single": 0,
511
+ "multi": 0
512
+ },
513
+ "Quantization": {
514
+ "unclassified": 0,
515
+ "single": 0,
516
+ "multi": 0
517
+ },
518
+ "Unclassified": {
519
+ "unclassified": 0,
520
+ "single": 0,
521
+ "multi": 0
522
+ }
523
+ },
524
+ "success": 254,
525
+ "time_spent": "0:02:02, 0:02:15, ",
526
+ "failures": {
527
+ "multi": [
528
+ {
529
+ "line": "tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
530
+ "trace": "(line 4140) KeyError: 'eager'"
531
+ },
532
+ {
533
+ "line": "tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_multi_gpu_data_parallel_forward",
534
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
535
+ }
536
+ ],
537
+ "single": [
538
+ {
539
+ "line": "tests/models/got_ocr2/test_modeling_got_ocr2.py::GotOcr2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
540
+ "trace": "(line 4140) KeyError: 'eager'"
541
+ }
542
+ ]
543
+ },
544
+ "job_link": {
545
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562995",
546
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563212"
547
+ }
548
+ },
549
+ "models_gpt2": {
550
+ "failed": {
551
+ "PyTorch": {
552
+ "unclassified": 0,
553
+ "single": 1,
554
+ "multi": 1
555
+ },
556
+ "TensorFlow": {
557
+ "unclassified": 0,
558
+ "single": 0,
559
+ "multi": 0
560
+ },
561
+ "Flax": {
562
+ "unclassified": 0,
563
+ "single": 0,
564
+ "multi": 0
565
+ },
566
+ "Tokenizers": {
567
+ "unclassified": 0,
568
+ "single": 0,
569
+ "multi": 0
570
+ },
571
+ "Pipelines": {
572
+ "unclassified": 0,
573
+ "single": 0,
574
+ "multi": 0
575
+ },
576
+ "Trainer": {
577
+ "unclassified": 0,
578
+ "single": 0,
579
+ "multi": 0
580
+ },
581
+ "ONNX": {
582
+ "unclassified": 0,
583
+ "single": 0,
584
+ "multi": 0
585
+ },
586
+ "Auto": {
587
+ "unclassified": 0,
588
+ "single": 0,
589
+ "multi": 0
590
+ },
591
+ "Quantization": {
592
+ "unclassified": 0,
593
+ "single": 0,
594
+ "multi": 0
595
+ },
596
+ "Unclassified": {
597
+ "unclassified": 0,
598
+ "single": 0,
599
+ "multi": 0
600
+ }
601
+ },
602
+ "success": 487,
603
+ "time_spent": "0:02:23, 0:02:38, ",
604
+ "failures": {
605
+ "multi": [
606
+ {
607
+ "line": "tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
608
+ "trace": "(line 4140) KeyError: 'eager'"
609
+ }
610
+ ],
611
+ "single": [
612
+ {
613
+ "line": "tests/models/gpt2/test_modeling_gpt2.py::GPT2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
614
+ "trace": "(line 4140) KeyError: 'eager'"
615
+ }
616
+ ]
617
+ },
618
+ "job_link": {
619
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563001",
620
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563255"
621
+ }
622
+ },
623
+ "models_internvl": {
624
+ "failed": {
625
+ "PyTorch": {
626
+ "unclassified": 0,
627
+ "single": 2,
628
+ "multi": 3
629
+ },
630
+ "TensorFlow": {
631
+ "unclassified": 0,
632
+ "single": 0,
633
+ "multi": 0
634
+ },
635
+ "Flax": {
636
+ "unclassified": 0,
637
+ "single": 0,
638
+ "multi": 0
639
+ },
640
+ "Tokenizers": {
641
+ "unclassified": 0,
642
+ "single": 0,
643
+ "multi": 0
644
+ },
645
+ "Pipelines": {
646
+ "unclassified": 0,
647
+ "single": 0,
648
+ "multi": 0
649
+ },
650
+ "Trainer": {
651
+ "unclassified": 0,
652
+ "single": 0,
653
+ "multi": 0
654
+ },
655
+ "ONNX": {
656
+ "unclassified": 0,
657
+ "single": 0,
658
+ "multi": 0
659
+ },
660
+ "Auto": {
661
+ "unclassified": 0,
662
+ "single": 0,
663
+ "multi": 0
664
+ },
665
+ "Quantization": {
666
+ "unclassified": 0,
667
+ "single": 0,
668
+ "multi": 0
669
+ },
670
+ "Unclassified": {
671
+ "unclassified": 0,
672
+ "single": 0,
673
+ "multi": 0
674
+ }
675
+ },
676
+ "success": 356,
677
+ "time_spent": "0:05:48, 0:04:49, ",
678
+ "failures": {
679
+ "multi": [
680
+ {
681
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids",
682
+ "trace": "(line 4140) KeyError: 'eager'"
683
+ },
684
+ {
685
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_flex_attention_with_grads",
686
+ "trace": "(line 439) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfResources: out of resource: shared memory, Required: 106496, Hardware limit: 101376. Reducing block sizes or `num_stages` may help."
687
+ },
688
+ {
689
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_multi_gpu_data_parallel_forward",
690
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
691
+ }
692
+ ],
693
+ "single": [
694
+ {
695
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_eager_padding_matches_padding_free_with_position_ids",
696
+ "trace": "(line 4140) KeyError: 'eager'"
697
+ },
698
+ {
699
+ "line": "tests/models/internvl/test_modeling_internvl.py::InternVLModelTest::test_flex_attention_with_grads",
700
+ "trace": "(line 439) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfResources: out of resource: shared memory, Required: 106496, Hardware limit: 101376. Reducing block sizes or `num_stages` may help."
701
+ }
702
+ ]
703
+ },
704
+ "job_link": {
705
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563553",
706
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563712"
707
+ }
708
+ },
709
+ "models_llama": {
710
+ "failed": {
711
+ "PyTorch": {
712
+ "unclassified": 0,
713
+ "single": 1,
714
+ "multi": 2
715
+ },
716
+ "TensorFlow": {
717
+ "unclassified": 0,
718
+ "single": 0,
719
+ "multi": 0
720
+ },
721
+ "Flax": {
722
+ "unclassified": 0,
723
+ "single": 0,
724
+ "multi": 0
725
+ },
726
+ "Tokenizers": {
727
+ "unclassified": 0,
728
+ "single": 0,
729
+ "multi": 0
730
+ },
731
+ "Pipelines": {
732
+ "unclassified": 0,
733
+ "single": 0,
734
+ "multi": 0
735
+ },
736
+ "Trainer": {
737
+ "unclassified": 0,
738
+ "single": 0,
739
+ "multi": 0
740
+ },
741
+ "ONNX": {
742
+ "unclassified": 0,
743
+ "single": 0,
744
+ "multi": 0
745
+ },
746
+ "Auto": {
747
+ "unclassified": 0,
748
+ "single": 0,
749
+ "multi": 0
750
+ },
751
+ "Quantization": {
752
+ "unclassified": 0,
753
+ "single": 0,
754
+ "multi": 0
755
+ },
756
+ "Unclassified": {
757
+ "unclassified": 0,
758
+ "single": 0,
759
+ "multi": 0
760
+ }
761
+ },
762
+ "success": 478,
763
+ "time_spent": "0:04:05, 0:03:53, ",
764
+ "failures": {
765
+ "multi": [
766
+ {
767
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids",
768
+ "trace": "(line 4140) KeyError: 'eager'"
769
+ },
770
+ {
771
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_multi_gpu_data_parallel_forward",
772
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
773
+ }
774
+ ],
775
+ "single": [
776
+ {
777
+ "line": "tests/models/llama/test_modeling_llama.py::LlamaModelTest::test_eager_padding_matches_padding_free_with_position_ids",
778
+ "trace": "(line 4140) KeyError: 'eager'"
779
+ }
780
+ ]
781
+ },
782
+ "job_link": {
783
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563871",
784
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526564103"
785
+ }
786
+ },
787
+ "models_llava": {
788
+ "failed": {
789
+ "PyTorch": {
790
+ "unclassified": 0,
791
+ "single": 3,
792
+ "multi": 4
793
+ },
794
+ "TensorFlow": {
795
+ "unclassified": 0,
796
+ "single": 0,
797
+ "multi": 0
798
+ },
799
+ "Flax": {
800
+ "unclassified": 0,
801
+ "single": 0,
802
+ "multi": 0
803
+ },
804
+ "Tokenizers": {
805
+ "unclassified": 0,
806
+ "single": 0,
807
+ "multi": 0
808
+ },
809
+ "Pipelines": {
810
+ "unclassified": 0,
811
+ "single": 0,
812
+ "multi": 0
813
+ },
814
+ "Trainer": {
815
+ "unclassified": 0,
816
+ "single": 0,
817
+ "multi": 0
818
+ },
819
+ "ONNX": {
820
+ "unclassified": 0,
821
+ "single": 0,
822
+ "multi": 0
823
+ },
824
+ "Auto": {
825
+ "unclassified": 0,
826
+ "single": 0,
827
+ "multi": 0
828
+ },
829
+ "Quantization": {
830
+ "unclassified": 0,
831
+ "single": 0,
832
+ "multi": 0
833
+ },
834
+ "Unclassified": {
835
+ "unclassified": 0,
836
+ "single": 0,
837
+ "multi": 0
838
+ }
839
+ },
840
+ "success": 346,
841
+ "time_spent": "0:10:11, 0:09:28, ",
842
+ "failures": {
843
+ "multi": [
844
+ {
845
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
846
+ "trace": "(line 4140) KeyError: 'eager'"
847
+ },
848
+ {
849
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads",
850
+ "trace": "(line 687) AssertionError: False is not true"
851
+ },
852
+ {
853
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
854
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
855
+ },
856
+ {
857
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
858
+ "trace": "(line 4197) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1"
859
+ }
860
+ ],
861
+ "single": [
862
+ {
863
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
864
+ "trace": "(line 4140) KeyError: 'eager'"
865
+ },
866
+ {
867
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_flex_attention_with_grads",
868
+ "trace": "(line 687) AssertionError: False is not true"
869
+ },
870
+ {
871
+ "line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_sdpa_padding_matches_padding_free_with_position_ids",
872
+ "trace": "(line 4197) IndexError: The shape of the mask [3, 23] at index 1 does not match the shape of the indexed tensor [3, 3, 8, 8] at index 1"
873
+ }
874
+ ]
875
+ },
876
+ "job_link": {
877
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526564002",
878
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526564108"
879
+ }
880
+ },
881
+ "models_mistral3": {
882
+ "failed": {
883
+ "PyTorch": {
884
+ "unclassified": 0,
885
+ "single": 1,
886
+ "multi": 2
887
+ },
888
+ "TensorFlow": {
889
+ "unclassified": 0,
890
+ "single": 0,
891
+ "multi": 0
892
+ },
893
+ "Flax": {
894
+ "unclassified": 0,
895
+ "single": 0,
896
+ "multi": 0
897
+ },
898
+ "Tokenizers": {
899
+ "unclassified": 0,
900
+ "single": 0,
901
+ "multi": 0
902
+ },
903
+ "Pipelines": {
904
+ "unclassified": 0,
905
+ "single": 0,
906
+ "multi": 0
907
+ },
908
+ "Trainer": {
909
+ "unclassified": 0,
910
+ "single": 0,
911
+ "multi": 0
912
+ },
913
+ "ONNX": {
914
+ "unclassified": 0,
915
+ "single": 0,
916
+ "multi": 0
917
+ },
918
+ "Auto": {
919
+ "unclassified": 0,
920
+ "single": 0,
921
+ "multi": 0
922
+ },
923
+ "Quantization": {
924
+ "unclassified": 0,
925
+ "single": 0,
926
+ "multi": 0
927
+ },
928
+ "Unclassified": {
929
+ "unclassified": 0,
930
+ "single": 0,
931
+ "multi": 0
932
+ }
933
+ },
934
+ "success": 286,
935
+ "time_spent": "0:10:06, 0:09:57, ",
936
+ "failures": {
937
+ "single": [
938
+ {
939
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
940
+ "trace": "(line 4140) KeyError: 'eager'"
941
+ }
942
+ ],
943
+ "multi": [
944
+ {
945
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
946
+ "trace": "(line 4140) KeyError: 'eager'"
947
+ },
948
+ {
949
+ "line": "tests/models/mistral3/test_modeling_mistral3.py::Mistral3ModelTest::test_multi_gpu_data_parallel_forward",
950
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
951
+ }
952
+ ]
953
+ },
954
+ "job_link": {
955
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561480",
956
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561618"
957
+ }
958
+ },
959
+ "models_modernbert": {
960
+ "failed": {
961
+ "PyTorch": {
962
+ "unclassified": 0,
963
+ "single": 5,
964
+ "multi": 5
965
+ },
966
+ "TensorFlow": {
967
+ "unclassified": 0,
968
+ "single": 0,
969
+ "multi": 0
970
+ },
971
+ "Flax": {
972
+ "unclassified": 0,
973
+ "single": 0,
974
+ "multi": 0
975
+ },
976
+ "Tokenizers": {
977
+ "unclassified": 0,
978
+ "single": 0,
979
+ "multi": 0
980
+ },
981
+ "Pipelines": {
982
+ "unclassified": 0,
983
+ "single": 0,
984
+ "multi": 0
985
+ },
986
+ "Trainer": {
987
+ "unclassified": 0,
988
+ "single": 0,
989
+ "multi": 0
990
+ },
991
+ "ONNX": {
992
+ "unclassified": 0,
993
+ "single": 0,
994
+ "multi": 0
995
+ },
996
+ "Auto": {
997
+ "unclassified": 0,
998
+ "single": 0,
999
+ "multi": 0
1000
+ },
1001
+ "Quantization": {
1002
+ "unclassified": 0,
1003
+ "single": 0,
1004
+ "multi": 0
1005
+ },
1006
+ "Unclassified": {
1007
+ "unclassified": 0,
1008
+ "single": 0,
1009
+ "multi": 0
1010
+ }
1011
+ },
1012
+ "success": 164,
1013
+ "time_spent": "0:01:29, 0:01:27, ",
1014
+ "failures": {
1015
+ "multi": [
1016
+ {
1017
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export",
1018
+ "trace": "(line 675) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']"
1019
+ },
1020
+ {
1021
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm",
1022
+ "trace": "(line 401) AssertionError: Tensor-likes are not close!"
1023
+ },
1024
+ {
1025
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head",
1026
+ "trace": "(line 423) AssertionError: Tensor-likes are not close!"
1027
+ },
1028
+ {
1029
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification",
1030
+ "trace": "(line 469) AssertionError: Tensor-likes are not close!"
1031
+ },
1032
+ {
1033
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification",
1034
+ "trace": "(line 446) AssertionError: Tensor-likes are not close!"
1035
+ }
1036
+ ],
1037
+ "single": [
1038
+ {
1039
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_export",
1040
+ "trace": "(line 675) AssertionError: Lists differ: ['mechanic', 'lawyer', 'teacher', 'waiter', 'doctor'] != ['lawyer', 'mechanic', 'teacher', 'doctor', 'waiter']"
1041
+ },
1042
+ {
1043
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_masked_lm",
1044
+ "trace": "(line 401) AssertionError: Tensor-likes are not close!"
1045
+ },
1046
+ {
1047
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_no_head",
1048
+ "trace": "(line 423) AssertionError: Tensor-likes are not close!"
1049
+ },
1050
+ {
1051
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_sequence_classification",
1052
+ "trace": "(line 469) AssertionError: Tensor-likes are not close!"
1053
+ },
1054
+ {
1055
+ "line": "tests/models/modernbert/test_modeling_modernbert.py::ModernBertModelIntegrationTest::test_inference_token_classification",
1056
+ "trace": "(line 446) AssertionError: Tensor-likes are not close!"
1057
+ }
1058
+ ]
1059
+ },
1060
+ "job_link": {
1061
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561668",
1062
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526561515"
1063
+ }
1064
+ },
1065
+ "models_qwen2": {
1066
+ "failed": {
1067
+ "PyTorch": {
1068
+ "unclassified": 0,
1069
+ "single": 2,
1070
+ "multi": 3
1071
+ },
1072
+ "TensorFlow": {
1073
+ "unclassified": 0,
1074
+ "single": 0,
1075
+ "multi": 0
1076
+ },
1077
+ "Flax": {
1078
+ "unclassified": 0,
1079
+ "single": 0,
1080
+ "multi": 0
1081
+ },
1082
+ "Tokenizers": {
1083
+ "unclassified": 0,
1084
+ "single": 0,
1085
+ "multi": 0
1086
+ },
1087
+ "Pipelines": {
1088
+ "unclassified": 0,
1089
+ "single": 0,
1090
+ "multi": 0
1091
+ },
1092
+ "Trainer": {
1093
+ "unclassified": 0,
1094
+ "single": 0,
1095
+ "multi": 0
1096
+ },
1097
+ "ONNX": {
1098
+ "unclassified": 0,
1099
+ "single": 0,
1100
+ "multi": 0
1101
+ },
1102
+ "Auto": {
1103
+ "unclassified": 0,
1104
+ "single": 0,
1105
+ "multi": 0
1106
+ },
1107
+ "Quantization": {
1108
+ "unclassified": 0,
1109
+ "single": 0,
1110
+ "multi": 0
1111
+ },
1112
+ "Unclassified": {
1113
+ "unclassified": 0,
1114
+ "single": 0,
1115
+ "multi": 0
1116
+ }
1117
+ },
1118
+ "success": 438,
1119
+ "time_spent": "0:02:17, 0:02:18, ",
1120
+ "failures": {
1121
+ "multi": [
1122
+ {
1123
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1124
+ "trace": "(line 4140) KeyError: 'eager'"
1125
+ },
1126
+ {
1127
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_multi_gpu_data_parallel_forward",
1128
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
1129
+ },
1130
+ {
1131
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache",
1132
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,"
1133
+ }
1134
+ ],
1135
+ "single": [
1136
+ {
1137
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1138
+ "trace": "(line 4140) KeyError: 'eager'"
1139
+ },
1140
+ {
1141
+ "line": "tests/models/qwen2/test_modeling_qwen2.py::Qwen2IntegrationTest::test_export_static_cache",
1142
+ "trace": "(line 1642) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_method index_copy_(*(FakeTensor(..., size=(1, 2, 26, 64), dtype=torch.bfloat16), 2, FakeTensor(..., device='cuda:0', size=(1,), dtype=torch.int64), FakeTensor(..., device='cuda:0', size=(1, 2, 1, 64), dtype=torch.bfloat16,"
1143
+ }
1144
+ ]
1145
+ },
1146
+ "job_link": {
1147
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562376",
1148
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562270"
1149
+ }
1150
+ },
1151
+ "models_qwen2_5_omni": {
1152
+ "failed": {
1153
+ "PyTorch": {
1154
+ "unclassified": 0,
1155
+ "single": 1,
1156
+ "multi": 5
1157
+ },
1158
+ "TensorFlow": {
1159
+ "unclassified": 0,
1160
+ "single": 0,
1161
+ "multi": 0
1162
+ },
1163
+ "Flax": {
1164
+ "unclassified": 0,
1165
+ "single": 0,
1166
+ "multi": 0
1167
+ },
1168
+ "Tokenizers": {
1169
+ "unclassified": 0,
1170
+ "single": 0,
1171
+ "multi": 0
1172
+ },
1173
+ "Pipelines": {
1174
+ "unclassified": 0,
1175
+ "single": 0,
1176
+ "multi": 0
1177
+ },
1178
+ "Trainer": {
1179
+ "unclassified": 0,
1180
+ "single": 0,
1181
+ "multi": 0
1182
+ },
1183
+ "ONNX": {
1184
+ "unclassified": 0,
1185
+ "single": 0,
1186
+ "multi": 0
1187
+ },
1188
+ "Auto": {
1189
+ "unclassified": 0,
1190
+ "single": 0,
1191
+ "multi": 0
1192
+ },
1193
+ "Quantization": {
1194
+ "unclassified": 0,
1195
+ "single": 0,
1196
+ "multi": 0
1197
+ },
1198
+ "Unclassified": {
1199
+ "unclassified": 0,
1200
+ "single": 0,
1201
+ "multi": 0
1202
+ }
1203
+ },
1204
+ "success": 277,
1205
+ "time_spent": "0:03:01, 0:03:21, ",
1206
+ "failures": {
1207
+ "multi": [
1208
+ {
1209
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_model_parallelism",
1210
+ "trace": "(line 675) AssertionError: Items in the second set but not the first:"
1211
+ },
1212
+ {
1213
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniThinkerForConditionalGenerationModelTest::test_multi_gpu_data_parallel_forward",
1214
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
1215
+ },
1216
+ {
1217
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
1218
+ "trace": "(line 675) AssertionError: Lists differ: [\"sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.\"] != [\"sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.\"]"
1219
+ },
1220
+ {
1221
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_multiturn",
1222
+ "trace": "(line 849) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU 1 has a total capacity of 22.18 GiB of which 6.50 MiB is free. Process 51940 has 22.17 GiB memory in use. Of the allocated memory 21.74 GiB is allocated by PyTorch, and 27.83 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1223
+ },
1224
+ {
1225
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_w_audio",
1226
+ "trace": "(line 1000) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 20.00 MiB. GPU 1 has a total capacity of 22.18 GiB of which 8.50 MiB is free. Process 51940 has 22.17 GiB memory in use. Of the allocated memory 21.75 GiB is allocated by PyTorch, and 17.78 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
1227
+ }
1228
+ ],
1229
+ "single": [
1230
+ {
1231
+ "line": "tests/models/qwen2_5_omni/test_modeling_qwen2_5_omni.py::Qwen2_5OmniModelIntegrationTest::test_small_model_integration_test_batch",
1232
+ "trace": "(line 675) AssertionError: Lists differ: [\"sys[96 chars]ant\\nsystem\\nYou are a helpful assistant.\\nuse[129 chars]er.\"] != [\"sys[96 chars]ant\\nThe sound is glass shattering, and the do[198 chars]er.\"]"
1233
+ }
1234
+ ]
1235
+ },
1236
+ "job_link": {
1237
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562375",
1238
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562289"
1239
+ }
1240
+ },
1241
+ "models_qwen2_5_vl": {
1242
+ "failed": {
1243
+ "PyTorch": {
1244
+ "unclassified": 0,
1245
+ "single": 1,
1246
+ "multi": 1
1247
+ },
1248
+ "TensorFlow": {
1249
+ "unclassified": 0,
1250
+ "single": 0,
1251
+ "multi": 0
1252
+ },
1253
+ "Flax": {
1254
+ "unclassified": 0,
1255
+ "single": 0,
1256
+ "multi": 0
1257
+ },
1258
+ "Tokenizers": {
1259
+ "unclassified": 0,
1260
+ "single": 0,
1261
+ "multi": 0
1262
+ },
1263
+ "Pipelines": {
1264
+ "unclassified": 0,
1265
+ "single": 0,
1266
+ "multi": 0
1267
+ },
1268
+ "Trainer": {
1269
+ "unclassified": 0,
1270
+ "single": 0,
1271
+ "multi": 0
1272
+ },
1273
+ "ONNX": {
1274
+ "unclassified": 0,
1275
+ "single": 0,
1276
+ "multi": 0
1277
+ },
1278
+ "Auto": {
1279
+ "unclassified": 0,
1280
+ "single": 0,
1281
+ "multi": 0
1282
+ },
1283
+ "Quantization": {
1284
+ "unclassified": 0,
1285
+ "single": 0,
1286
+ "multi": 0
1287
+ },
1288
+ "Unclassified": {
1289
+ "unclassified": 0,
1290
+ "single": 0,
1291
+ "multi": 0
1292
+ }
1293
+ },
1294
+ "success": 311,
1295
+ "time_spent": "0:03:25, 0:03:29, ",
1296
+ "failures": {
1297
+ "multi": [
1298
+ {
1299
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
1300
+ "trace": "(line 675) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']"
1301
+ }
1302
+ ],
1303
+ "single": [
1304
+ {
1305
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
1306
+ "trace": "(line 675) AssertionError: Lists differ: ['sys[314 chars]ion\\n addCriterion\\n\\n addCriterion\\n\\n addCri[75 chars]n\\n'] != ['sys[314 chars]ion\\nThe dog in the picture appears to be a La[81 chars] is']"
1307
+ }
1308
+ ]
1309
+ },
1310
+ "job_link": {
1311
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562382",
1312
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562290"
1313
+ }
1314
+ },
1315
+ "models_smolvlm": {
1316
+ "failed": {
1317
+ "PyTorch": {
1318
+ "unclassified": 0,
1319
+ "single": 1,
1320
+ "multi": 1
1321
+ },
1322
+ "TensorFlow": {
1323
+ "unclassified": 0,
1324
+ "single": 0,
1325
+ "multi": 0
1326
+ },
1327
+ "Flax": {
1328
+ "unclassified": 0,
1329
+ "single": 0,
1330
+ "multi": 0
1331
+ },
1332
+ "Tokenizers": {
1333
+ "unclassified": 0,
1334
+ "single": 0,
1335
+ "multi": 0
1336
+ },
1337
+ "Pipelines": {
1338
+ "unclassified": 0,
1339
+ "single": 0,
1340
+ "multi": 0
1341
+ },
1342
+ "Trainer": {
1343
+ "unclassified": 0,
1344
+ "single": 0,
1345
+ "multi": 0
1346
+ },
1347
+ "ONNX": {
1348
+ "unclassified": 0,
1349
+ "single": 0,
1350
+ "multi": 0
1351
+ },
1352
+ "Auto": {
1353
+ "unclassified": 0,
1354
+ "single": 0,
1355
+ "multi": 0
1356
+ },
1357
+ "Quantization": {
1358
+ "unclassified": 0,
1359
+ "single": 0,
1360
+ "multi": 0
1361
+ },
1362
+ "Unclassified": {
1363
+ "unclassified": 0,
1364
+ "single": 0,
1365
+ "multi": 0
1366
+ }
1367
+ },
1368
+ "success": 499,
1369
+ "time_spent": "0:01:55, 0:01:47, ",
1370
+ "failures": {
1371
+ "single": [
1372
+ {
1373
+ "line": "tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1374
+ "trace": "(line 4140) KeyError: 'eager'"
1375
+ }
1376
+ ],
1377
+ "multi": [
1378
+ {
1379
+ "line": "tests/models/smolvlm/test_modeling_smolvlm.py::SmolVLMForConditionalGenerationModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1380
+ "trace": "(line 4140) KeyError: 'eager'"
1381
+ }
1382
+ ]
1383
+ },
1384
+ "job_link": {
1385
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562675",
1386
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562798"
1387
+ }
1388
+ },
1389
+ "models_t5": {
1390
+ "failed": {
1391
+ "PyTorch": {
1392
+ "unclassified": 0,
1393
+ "single": 2,
1394
+ "multi": 3
1395
+ },
1396
+ "TensorFlow": {
1397
+ "unclassified": 0,
1398
+ "single": 0,
1399
+ "multi": 0
1400
+ },
1401
+ "Flax": {
1402
+ "unclassified": 0,
1403
+ "single": 0,
1404
+ "multi": 0
1405
+ },
1406
+ "Tokenizers": {
1407
+ "unclassified": 0,
1408
+ "single": 0,
1409
+ "multi": 0
1410
+ },
1411
+ "Pipelines": {
1412
+ "unclassified": 0,
1413
+ "single": 0,
1414
+ "multi": 0
1415
+ },
1416
+ "Trainer": {
1417
+ "unclassified": 0,
1418
+ "single": 0,
1419
+ "multi": 0
1420
+ },
1421
+ "ONNX": {
1422
+ "unclassified": 0,
1423
+ "single": 0,
1424
+ "multi": 0
1425
+ },
1426
+ "Auto": {
1427
+ "unclassified": 0,
1428
+ "single": 0,
1429
+ "multi": 0
1430
+ },
1431
+ "Quantization": {
1432
+ "unclassified": 0,
1433
+ "single": 0,
1434
+ "multi": 0
1435
+ },
1436
+ "Unclassified": {
1437
+ "unclassified": 0,
1438
+ "single": 0,
1439
+ "multi": 0
1440
+ }
1441
+ },
1442
+ "success": 592,
1443
+ "time_spent": "0:03:34, 0:03:41, ",
1444
+ "failures": {
1445
+ "multi": [
1446
+ {
1447
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1448
+ "trace": "(line 4140) KeyError: 'eager'"
1449
+ },
1450
+ {
1451
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelTest::test_multi_gpu_data_parallel_forward",
1452
+ "trace": "(line 131) TypeError: EncoderDecoderCache.__init__() missing 1 required positional argument: 'cross_attention_cache'"
1453
+ },
1454
+ {
1455
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization",
1456
+ "trace": "(line 687) AttributeError: 'dict' object has no attribute 'batch_size'"
1457
+ }
1458
+ ],
1459
+ "single": [
1460
+ {
1461
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1462
+ "trace": "(line 4140) KeyError: 'eager'"
1463
+ },
1464
+ {
1465
+ "line": "tests/models/t5/test_modeling_t5.py::T5ModelIntegrationTests::test_export_t5_summarization",
1466
+ "trace": "(line 687) AttributeError: 'dict' object has no attribute 'batch_size'"
1467
+ }
1468
+ ]
1469
+ },
1470
+ "job_link": {
1471
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563047",
1472
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526562939"
1473
+ }
1474
+ },
1475
+ "models_vit": {
1476
+ "failed": {
1477
+ "PyTorch": {
1478
+ "unclassified": 0,
1479
+ "single": 0,
1480
+ "multi": 0
1481
+ },
1482
+ "TensorFlow": {
1483
+ "unclassified": 0,
1484
+ "single": 0,
1485
+ "multi": 0
1486
+ },
1487
+ "Flax": {
1488
+ "unclassified": 0,
1489
+ "single": 0,
1490
+ "multi": 0
1491
+ },
1492
+ "Tokenizers": {
1493
+ "unclassified": 0,
1494
+ "single": 0,
1495
+ "multi": 0
1496
+ },
1497
+ "Pipelines": {
1498
+ "unclassified": 0,
1499
+ "single": 0,
1500
+ "multi": 0
1501
+ },
1502
+ "Trainer": {
1503
+ "unclassified": 0,
1504
+ "single": 0,
1505
+ "multi": 0
1506
+ },
1507
+ "ONNX": {
1508
+ "unclassified": 0,
1509
+ "single": 0,
1510
+ "multi": 0
1511
+ },
1512
+ "Auto": {
1513
+ "unclassified": 0,
1514
+ "single": 0,
1515
+ "multi": 0
1516
+ },
1517
+ "Quantization": {
1518
+ "unclassified": 0,
1519
+ "single": 0,
1520
+ "multi": 0
1521
+ },
1522
+ "Unclassified": {
1523
+ "unclassified": 0,
1524
+ "single": 0,
1525
+ "multi": 0
1526
+ }
1527
+ },
1528
+ "success": 217,
1529
+ "time_spent": "7.34, 0:01:09, ",
1530
+ "failures": {},
1531
+ "job_link": {
1532
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563537",
1533
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563397"
1534
+ }
1535
+ },
1536
+ "models_wav2vec2": {
1537
+ "failed": {
1538
+ "PyTorch": {
1539
+ "unclassified": 0,
1540
+ "single": 4,
1541
+ "multi": 4
1542
+ },
1543
+ "TensorFlow": {
1544
+ "unclassified": 0,
1545
+ "single": 0,
1546
+ "multi": 0
1547
+ },
1548
+ "Flax": {
1549
+ "unclassified": 0,
1550
+ "single": 0,
1551
+ "multi": 0
1552
+ },
1553
+ "Tokenizers": {
1554
+ "unclassified": 0,
1555
+ "single": 0,
1556
+ "multi": 0
1557
+ },
1558
+ "Pipelines": {
1559
+ "unclassified": 0,
1560
+ "single": 0,
1561
+ "multi": 0
1562
+ },
1563
+ "Trainer": {
1564
+ "unclassified": 0,
1565
+ "single": 0,
1566
+ "multi": 0
1567
+ },
1568
+ "ONNX": {
1569
+ "unclassified": 0,
1570
+ "single": 0,
1571
+ "multi": 0
1572
+ },
1573
+ "Auto": {
1574
+ "unclassified": 0,
1575
+ "single": 0,
1576
+ "multi": 0
1577
+ },
1578
+ "Quantization": {
1579
+ "unclassified": 0,
1580
+ "single": 0,
1581
+ "multi": 0
1582
+ },
1583
+ "Unclassified": {
1584
+ "unclassified": 0,
1585
+ "single": 0,
1586
+ "multi": 0
1587
+ }
1588
+ },
1589
+ "success": 672,
1590
+ "time_spent": "0:04:46, 0:04:23, ",
1591
+ "failures": {
1592
+ "multi": [
1593
+ {
1594
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_inference_mms_1b_all",
1595
+ "trace": "(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py"
1596
+ },
1597
+ {
1598
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm",
1599
+ "trace": "(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py"
1600
+ },
1601
+ {
1602
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_invalid_pool",
1603
+ "trace": "(line 675) AssertionError: Traceback (most recent call last):"
1604
+ },
1605
+ {
1606
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_pool",
1607
+ "trace": "(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py"
1608
+ }
1609
+ ],
1610
+ "single": [
1611
+ {
1612
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_inference_mms_1b_all",
1613
+ "trace": "(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py"
1614
+ },
1615
+ {
1616
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm",
1617
+ "trace": "(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py"
1618
+ },
1619
+ {
1620
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_invalid_pool",
1621
+ "trace": "(line 675) AssertionError: Traceback (most recent call last):"
1622
+ },
1623
+ {
1624
+ "line": "tests/models/wav2vec2/test_modeling_wav2vec2.py::Wav2Vec2ModelIntegrationTest::test_wav2vec2_with_lm_pool",
1625
+ "trace": "(line 989) RuntimeError: Dataset scripts are no longer supported, but found common_voice_11_0.py"
1626
+ }
1627
+ ]
1628
+ },
1629
+ "job_link": {
1630
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563711",
1631
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563582"
1632
+ }
1633
+ },
1634
+ "models_whisper": {
1635
+ "failed": {
1636
+ "PyTorch": {
1637
+ "unclassified": 0,
1638
+ "single": 8,
1639
+ "multi": 11
1640
+ },
1641
+ "TensorFlow": {
1642
+ "unclassified": 0,
1643
+ "single": 0,
1644
+ "multi": 0
1645
+ },
1646
+ "Flax": {
1647
+ "unclassified": 0,
1648
+ "single": 0,
1649
+ "multi": 0
1650
+ },
1651
+ "Tokenizers": {
1652
+ "unclassified": 0,
1653
+ "single": 0,
1654
+ "multi": 0
1655
+ },
1656
+ "Pipelines": {
1657
+ "unclassified": 0,
1658
+ "single": 0,
1659
+ "multi": 0
1660
+ },
1661
+ "Trainer": {
1662
+ "unclassified": 0,
1663
+ "single": 0,
1664
+ "multi": 0
1665
+ },
1666
+ "ONNX": {
1667
+ "unclassified": 0,
1668
+ "single": 0,
1669
+ "multi": 0
1670
+ },
1671
+ "Auto": {
1672
+ "unclassified": 0,
1673
+ "single": 0,
1674
+ "multi": 0
1675
+ },
1676
+ "Quantization": {
1677
+ "unclassified": 0,
1678
+ "single": 0,
1679
+ "multi": 0
1680
+ },
1681
+ "Unclassified": {
1682
+ "unclassified": 0,
1683
+ "single": 0,
1684
+ "multi": 0
1685
+ }
1686
+ },
1687
+ "success": 1010,
1688
+ "time_spent": "0:12:29, 0:14:19, ",
1689
+ "failures": {
1690
+ "single": [
1691
+ {
1692
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1693
+ "trace": "(line 4140) KeyError: 'eager'"
1694
+ },
1695
+ {
1696
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_large_batched_generation_multilingual",
1697
+ "trace": "(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
1698
+ },
1699
+ {
1700
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_small_longform_timestamps_generation",
1701
+ "trace": "(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
1702
+ },
1703
+ {
1704
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_tiny_longform_timestamps_generation",
1705
+ "trace": "(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
1706
+ },
1707
+ {
1708
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard",
1709
+ "trace": "(line 675) AssertionError: Lists differ: [\" Fo[272 chars]ting of classics, Sicilian, nade door variatio[8147 chars]le!'] != [\" Fo[272 chars]ting a classic Sicilian, nade door variation o[8150 chars]le!']"
1710
+ },
1711
+ {
1712
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard_prev_cond",
1713
+ "trace": "(line 675) AssertionError: Lists differ: [\" Fo[422 chars]to a fisher shows in lip-nitsky attack that cu[7903 chars]le!\"] != [\" Fo[422 chars]to a Fisher shows in lip-nitsky attack that cu[7918 chars]le.\"]"
1714
+ },
1715
+ {
1716
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_shortform_single_batch_prev_cond",
1717
+ "trace": "(line 675) AssertionError: Lists differ: [\" Fo[268 chars]ating, so soft, it would make JD power and her[196 chars]ke.\"] != [\" Fo[268 chars]ating so soft, it would make JD power and her [195 chars]ke.\"]"
1718
+ },
1719
+ {
1720
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperStandaloneDecoderModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1721
+ "trace": "(line 4140) KeyError: 'eager'"
1722
+ }
1723
+ ],
1724
+ "multi": [
1725
+ {
1726
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1727
+ "trace": "(line 4140) KeyError: 'eager'"
1728
+ },
1729
+ {
1730
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelTest::test_multi_gpu_data_parallel_forward",
1731
+ "trace": "(line 131) TypeError: EncoderDecoderCache.__init__() missing 1 required positional argument: 'cross_attention_cache'"
1732
+ },
1733
+ {
1734
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_generate_with_forced_decoder_ids",
1735
+ "trace": "(line 713) requests.exceptions.ReadTimeout: (ReadTimeoutError(\"HTTPSConnectionPool(host='huggingface.co', port=443): Read timed out. (read timeout=10)\"), '(Request ID: 13cb0b08-c261-4ca3-a58f-91a2f3e327ed)')"
1736
+ },
1737
+ {
1738
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_large_batched_generation_multilingual",
1739
+ "trace": "(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
1740
+ },
1741
+ {
1742
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_small_longform_timestamps_generation",
1743
+ "trace": "(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
1744
+ },
1745
+ {
1746
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_tiny_longform_timestamps_generation",
1747
+ "trace": "(line 756) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
1748
+ },
1749
+ {
1750
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard",
1751
+ "trace": "(line 675) AssertionError: Lists differ: [\" Fo[272 chars]ting of classics, Sicilian, nade door variatio[8147 chars]le!'] != [\" Fo[272 chars]ting a classic Sicilian, nade door variation o[8150 chars]le!']"
1752
+ },
1753
+ {
1754
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_longform_multi_batch_hard_prev_cond",
1755
+ "trace": "(line 675) AssertionError: Lists differ: [\" Fo[422 chars]to a fisher shows in lip-nitsky attack that cu[7903 chars]le!\"] != [\" Fo[422 chars]to a Fisher shows in lip-nitsky attack that cu[7918 chars]le.\"]"
1756
+ },
1757
+ {
1758
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperModelIntegrationTests::test_whisper_shortform_single_batch_prev_cond",
1759
+ "trace": "(line 675) AssertionError: Lists differ: [\" Fo[268 chars]ating, so soft, it would make JD power and her[196 chars]ke.\"] != [\" Fo[268 chars]ating so soft, it would make JD power and her [195 chars]ke.\"]"
1760
+ },
1761
+ {
1762
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperStandaloneDecoderModelTest::test_eager_padding_matches_padding_free_with_position_ids",
1763
+ "trace": "(line 4140) KeyError: 'eager'"
1764
+ },
1765
+ {
1766
+ "line": "tests/models/whisper/test_modeling_whisper.py::WhisperStandaloneDecoderModelTest::test_multi_gpu_data_parallel_forward",
1767
+ "trace": "(line 1305) AttributeError: 'DynamicCache' object has no attribute 'layers'"
1768
+ }
1769
+ ]
1770
+ },
1771
+ "job_link": {
1772
+ "single": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563737",
1773
+ "multi": "https://github.com/huggingface/transformers/actions/runs/16460401119/job/46526563862"
1774
+ }
1775
+ }
1776
+ }