ArthurZ HF Staff commited on
Commit
6ec3fee
·
verified ·
1 Parent(s): 006c755

Update CircleCI artifacts for PR 41541 (0e9d3d27d89d)

Browse files
pr-41541/sha-0e9d3d27d89d/failure_summary.json ADDED
@@ -0,0 +1,291 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "failures": [
3
+ {
4
+ "job_name": "tests_torch",
5
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_batched_and_grouped_inference_1_fp32",
6
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
7
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
8
+ "model_name": "qwen3_omni_moe"
9
+ },
10
+ {
11
+ "job_name": "tests_torch",
12
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_08_fp32_pad_left_sdpa_kernels",
13
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
14
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
15
+ "model_name": "qwen3_omni_moe"
16
+ },
17
+ {
18
+ "job_name": "tests_torch",
19
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_09_fp32_pad_left",
20
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
21
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
22
+ "model_name": "qwen3_omni_moe"
23
+ },
24
+ {
25
+ "job_name": "tests_torch",
26
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_10_fp32_pad_left_no_attn_mask_sdpa_kernels",
27
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
28
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
29
+ "model_name": "qwen3_omni_moe"
30
+ },
31
+ {
32
+ "job_name": "tests_torch",
33
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_11_fp32_pad_left_no_attn_mask",
34
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
35
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
36
+ "model_name": "qwen3_omni_moe"
37
+ },
38
+ {
39
+ "job_name": "tests_torch",
40
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_12_fp32_pad_right_sdpa_kernels",
41
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
42
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
43
+ "model_name": "qwen3_omni_moe"
44
+ },
45
+ {
46
+ "job_name": "tests_torch",
47
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_13_fp32_pad_right",
48
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
49
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
50
+ "model_name": "qwen3_omni_moe"
51
+ },
52
+ {
53
+ "job_name": "tests_torch",
54
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_14_fp32_pad_right_no_attn_mask_sdpa_kernels",
55
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
56
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
57
+ "model_name": "qwen3_omni_moe"
58
+ },
59
+ {
60
+ "job_name": "tests_torch",
61
+ "test_name": "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_15_fp32_pad_right_no_attn_mask",
62
+ "short_error": "RuntimeError: expected data_ptr to be aligned to 16 bytes",
63
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes",
64
+ "model_name": "qwen3_omni_moe"
65
+ },
66
+ {
67
+ "job_name": "tests_processors",
68
+ "test_name": "tests/models/markuplm/test_processing_markuplm.py::MarkupLMProcessorTest::test_save_load_pretrained_default",
69
+ "short_error": "AssertionError: TokenizersBackend(name_or_path='/tmp/tmptemvc1sw', vocab_size=22, model_max_length=1000000000000000019884624838656, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': '<mask>'}, added_tokens_decoder={",
70
+ "error": "/usr/local/lib/python3.10/unittest/case.py:675: AssertionError: TokenizersBackend(name_or_path='/tmp/tmptemvc1sw', vocab_size=22, model_max_length=1000000000000000019884624838656, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': '<mask>'}, added_tokens_decoder={",
71
+ "model_name": "markuplm"
72
+ },
73
+ {
74
+ "job_name": "tests_generate",
75
+ "test_name": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_assisted_decoding_matches_greedy_search_1_same",
76
+ "short_error": "AssertionError: False is not true",
77
+ "error": "/usr/local/lib/python3.10/unittest/case.py:687: AssertionError: False is not true",
78
+ "model_name": "paligemma"
79
+ },
80
+ {
81
+ "job_name": "tests_exotic_models",
82
+ "test_name": "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_image_processor_defaults_preserved_by_image_kwargs",
83
+ "short_error": "TypeError: TextInputSequence must be str",
84
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str",
85
+ "model_name": "udop"
86
+ },
87
+ {
88
+ "job_name": "tests_exotic_models",
89
+ "test_name": "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_image_processor_kwargs",
90
+ "short_error": "TypeError: TextInputSequence must be str",
91
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str",
92
+ "model_name": "udop"
93
+ },
94
+ {
95
+ "job_name": "tests_exotic_models",
96
+ "test_name": "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_tokenizer_kwargs",
97
+ "short_error": "TypeError: TextInputSequence must be str",
98
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str",
99
+ "model_name": "udop"
100
+ },
101
+ {
102
+ "job_name": "tests_exotic_models",
103
+ "test_name": "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_model_input_names",
104
+ "short_error": "TypeError: TextInputSequence must be str",
105
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str",
106
+ "model_name": "udop"
107
+ },
108
+ {
109
+ "job_name": "tests_exotic_models",
110
+ "test_name": "tests/models/layoutlmv2/test_processing_layoutlmv2.py::LayoutLMv2ProcessorTest::test_model_input_names",
111
+ "short_error": "TypeError: TextInputSequence must be str",
112
+ "error": "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str",
113
+ "model_name": "layoutlmv2"
114
+ }
115
+ ],
116
+ "by_test": {
117
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_batched_and_grouped_inference_1_fp32": {
118
+ "count": 1,
119
+ "errors": {
120
+ "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes": 1
121
+ },
122
+ "jobs": [
123
+ "tests_torch"
124
+ ],
125
+ "variants": [
126
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_batched_and_grouped_inference_1_fp32"
127
+ ]
128
+ },
129
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference": {
130
+ "count": 8,
131
+ "errors": {
132
+ "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes": 8
133
+ },
134
+ "jobs": [
135
+ "tests_torch"
136
+ ],
137
+ "variants": [
138
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_08_fp32_pad_left_sdpa_kernels",
139
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_09_fp32_pad_left",
140
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_10_fp32_pad_left_no_attn_mask_sdpa_kernels",
141
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_11_fp32_pad_left_no_attn_mask",
142
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_12_fp32_pad_right_sdpa_kernels",
143
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_13_fp32_pad_right",
144
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_14_fp32_pad_right_no_attn_mask_sdpa_kernels",
145
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_15_fp32_pad_right_no_attn_mask"
146
+ ]
147
+ },
148
+ "tests/models/markuplm/test_processing_markuplm.py::MarkupLMProcessorTest::test_save_load_pretrained_default": {
149
+ "count": 1,
150
+ "errors": {
151
+ "/usr/local/lib/python3.10/unittest/case.py:675: AssertionError: TokenizersBackend(name_or_path='/tmp/tmptemvc1sw', vocab_size=22, model_max_length=1000000000000000019884624838656, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': '<mask>'}, added_tokens_decoder={": 1
152
+ },
153
+ "jobs": [
154
+ "tests_processors"
155
+ ],
156
+ "variants": [
157
+ "tests/models/markuplm/test_processing_markuplm.py::MarkupLMProcessorTest::test_save_load_pretrained_default"
158
+ ]
159
+ },
160
+ "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_assisted_decoding_matches_greedy_search_1_same": {
161
+ "count": 1,
162
+ "errors": {
163
+ "/usr/local/lib/python3.10/unittest/case.py:687: AssertionError: False is not true": 1
164
+ },
165
+ "jobs": [
166
+ "tests_generate"
167
+ ],
168
+ "variants": [
169
+ "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_assisted_decoding_matches_greedy_search_1_same"
170
+ ]
171
+ },
172
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_image_processor_defaults_preserved_by_image_kwargs": {
173
+ "count": 1,
174
+ "errors": {
175
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 1
176
+ },
177
+ "jobs": [
178
+ "tests_exotic_models"
179
+ ],
180
+ "variants": [
181
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_image_processor_defaults_preserved_by_image_kwargs"
182
+ ]
183
+ },
184
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_image_processor_kwargs": {
185
+ "count": 1,
186
+ "errors": {
187
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 1
188
+ },
189
+ "jobs": [
190
+ "tests_exotic_models"
191
+ ],
192
+ "variants": [
193
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_image_processor_kwargs"
194
+ ]
195
+ },
196
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_tokenizer_kwargs": {
197
+ "count": 1,
198
+ "errors": {
199
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 1
200
+ },
201
+ "jobs": [
202
+ "tests_exotic_models"
203
+ ],
204
+ "variants": [
205
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_tokenizer_kwargs"
206
+ ]
207
+ },
208
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_model_input_names": {
209
+ "count": 1,
210
+ "errors": {
211
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 1
212
+ },
213
+ "jobs": [
214
+ "tests_exotic_models"
215
+ ],
216
+ "variants": [
217
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_model_input_names"
218
+ ]
219
+ },
220
+ "tests/models/layoutlmv2/test_processing_layoutlmv2.py::LayoutLMv2ProcessorTest::test_model_input_names": {
221
+ "count": 1,
222
+ "errors": {
223
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 1
224
+ },
225
+ "jobs": [
226
+ "tests_exotic_models"
227
+ ],
228
+ "variants": [
229
+ "tests/models/layoutlmv2/test_processing_layoutlmv2.py::LayoutLMv2ProcessorTest::test_model_input_names"
230
+ ]
231
+ }
232
+ },
233
+ "by_model": {
234
+ "qwen3_omni_moe": {
235
+ "count": 9,
236
+ "errors": {
237
+ "/usr/local/lib/python3.10/site-packages/transformers/integrations/moe.py:179: RuntimeError: expected data_ptr to be aligned to 16 bytes": 9
238
+ },
239
+ "tests": [
240
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_batched_and_grouped_inference_1_fp32",
241
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_08_fp32_pad_left_sdpa_kernels",
242
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_09_fp32_pad_left",
243
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_10_fp32_pad_left_no_attn_mask_sdpa_kernels",
244
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_11_fp32_pad_left_no_attn_mask",
245
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_12_fp32_pad_right_sdpa_kernels",
246
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_13_fp32_pad_right",
247
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_14_fp32_pad_right_no_attn_mask_sdpa_kernels",
248
+ "tests/models/qwen3_omni_moe/test_modeling_qwen3_omni_moe.py::Qwen3OmniMoeThinkerForConditionalGenerationModelTest::test_eager_matches_sdpa_inference_15_fp32_pad_right_no_attn_mask"
249
+ ]
250
+ },
251
+ "markuplm": {
252
+ "count": 1,
253
+ "errors": {
254
+ "/usr/local/lib/python3.10/unittest/case.py:675: AssertionError: TokenizersBackend(name_or_path='/tmp/tmptemvc1sw', vocab_size=22, model_max_length=1000000000000000019884624838656, padding_side='right', truncation_side='right', special_tokens={'bos_token': '<s>', 'eos_token': '</s>', 'unk_token': '<unk>', 'sep_token': '</s>', 'pad_token': '<pad>', 'cls_token': '<s>', 'mask_token': '<mask>'}, added_tokens_decoder={": 1
255
+ },
256
+ "tests": [
257
+ "tests/models/markuplm/test_processing_markuplm.py::MarkupLMProcessorTest::test_save_load_pretrained_default"
258
+ ]
259
+ },
260
+ "paligemma": {
261
+ "count": 1,
262
+ "errors": {
263
+ "/usr/local/lib/python3.10/unittest/case.py:687: AssertionError: False is not true": 1
264
+ },
265
+ "tests": [
266
+ "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationModelTest::test_assisted_decoding_matches_greedy_search_1_same"
267
+ ]
268
+ },
269
+ "udop": {
270
+ "count": 4,
271
+ "errors": {
272
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 4
273
+ },
274
+ "tests": [
275
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_image_processor_defaults_preserved_by_image_kwargs",
276
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_image_processor_kwargs",
277
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_kwargs_overrides_default_tokenizer_kwargs",
278
+ "tests/models/udop/test_processing_udop.py::UdopProcessorTest::test_model_input_names"
279
+ ]
280
+ },
281
+ "layoutlmv2": {
282
+ "count": 1,
283
+ "errors": {
284
+ "/usr/local/lib/python3.10/site-packages/transformers/tokenization_utils_tokenizers.py:853: TypeError: TextInputSequence must be str": 1
285
+ },
286
+ "tests": [
287
+ "tests/models/layoutlmv2/test_processing_layoutlmv2.py::LayoutLMv2ProcessorTest::test_model_input_names"
288
+ ]
289
+ }
290
+ }
291
+ }