Upload 2025-12-22/runs/25849-20423647274/ci_results_run_models_gpu/model_results.json with huggingface_hub
Browse files
2025-12-22/runs/25849-20423647274/ci_results_run_models_gpu/model_results.json
ADDED
|
@@ -0,0 +1,403 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"models_llava": {
|
| 3 |
+
"failed": {
|
| 4 |
+
"PyTorch": {
|
| 5 |
+
"unclassified": 0,
|
| 6 |
+
"single": 10,
|
| 7 |
+
"multi": 9
|
| 8 |
+
},
|
| 9 |
+
"Tokenizers": {
|
| 10 |
+
"unclassified": 0,
|
| 11 |
+
"single": 0,
|
| 12 |
+
"multi": 0
|
| 13 |
+
},
|
| 14 |
+
"Pipelines": {
|
| 15 |
+
"unclassified": 0,
|
| 16 |
+
"single": 0,
|
| 17 |
+
"multi": 0
|
| 18 |
+
},
|
| 19 |
+
"Trainer": {
|
| 20 |
+
"unclassified": 0,
|
| 21 |
+
"single": 0,
|
| 22 |
+
"multi": 0
|
| 23 |
+
},
|
| 24 |
+
"ONNX": {
|
| 25 |
+
"unclassified": 0,
|
| 26 |
+
"single": 0,
|
| 27 |
+
"multi": 0
|
| 28 |
+
},
|
| 29 |
+
"Auto": {
|
| 30 |
+
"unclassified": 0,
|
| 31 |
+
"single": 0,
|
| 32 |
+
"multi": 0
|
| 33 |
+
},
|
| 34 |
+
"Quantization": {
|
| 35 |
+
"unclassified": 0,
|
| 36 |
+
"single": 0,
|
| 37 |
+
"multi": 0
|
| 38 |
+
},
|
| 39 |
+
"Unclassified": {
|
| 40 |
+
"unclassified": 0,
|
| 41 |
+
"single": 0,
|
| 42 |
+
"multi": 0
|
| 43 |
+
}
|
| 44 |
+
},
|
| 45 |
+
"errors": 0,
|
| 46 |
+
"success": 356,
|
| 47 |
+
"skipped": 253,
|
| 48 |
+
"time_spent": [
|
| 49 |
+
296.86,
|
| 50 |
+
292.09
|
| 51 |
+
],
|
| 52 |
+
"error": false,
|
| 53 |
+
"failures": {
|
| 54 |
+
"single": [
|
| 55 |
+
{
|
| 56 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 57 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation",
|
| 61 |
+
"trace": "(line 562) AssertionError: Lists differ: [\"\\n\\nUSER: What's the difference of two imag[339 chars]ama'] != [\"\\n \\nUSER: What's the difference of two ima[362 chars]the']"
|
| 62 |
+
},
|
| 63 |
+
{
|
| 64 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_generation_no_images",
|
| 65 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 10.69 MiB is free. Process 17245 has 22.29 GiB memory in use. Of the allocated memory 21.88 GiB is allocated by PyTorch, and 24.45 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral",
|
| 69 |
+
"trace": "(line 543) RuntimeError: Input type (torch.FloatTensor) and weight type (CPUBFloat16Type) should be the same or input should be a MKLDNN tensor and weight is a dense tensor"
|
| 70 |
+
},
|
| 71 |
+
{
|
| 72 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_4bit",
|
| 73 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 40.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 28.69 MiB is free. Process 17245 has 22.27 GiB memory in use. Of the allocated memory 21.88 GiB is allocated by PyTorch, and 9.70 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 74 |
+
},
|
| 75 |
+
{
|
| 76 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_batched",
|
| 77 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 140.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 126.69 MiB is free. Process 17245 has 22.17 GiB memory in use. Of the allocated memory 21.74 GiB is allocated by PyTorch, and 51.91 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 78 |
+
},
|
| 79 |
+
{
|
| 80 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched",
|
| 81 |
+
"trace": "(line 404) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[269 chars] on'] != ['USER: \\nWhat are the things I should be cautio[271 chars] on']"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched_regression",
|
| 85 |
+
"trace": "(line 507) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[280 chars]ed.'] != ['USER: \\nWhat are the things I should be cautio[283 chars]ed.']"
|
| 86 |
+
},
|
| 87 |
+
{
|
| 88 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_single",
|
| 89 |
+
"trace": "(line 353) AssertionError: 'USER: \\nWhat are the things I should be cautiou[748 chars]ies.' != 'USER: \\nWhat are the things I should be cautio[749 chars]ies.'"
|
| 90 |
+
},
|
| 91 |
+
{
|
| 92 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_tokenizer_integration",
|
| 93 |
+
"trace": "(line 579) AssertionError: Lists differ: ['<|im_start|>', '▁system', '\\n', '▁Answer', '▁the', '▁ques[176 chars]'\\n'] != ['<|im_start|>', 'system', '\\n', 'Answer', '▁the', '▁questi[175 chars]'\\n']"
|
| 94 |
+
}
|
| 95 |
+
],
|
| 96 |
+
"multi": [
|
| 97 |
+
{
|
| 98 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 99 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 100 |
+
},
|
| 101 |
+
{
|
| 102 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation",
|
| 103 |
+
"trace": "(line 562) AssertionError: Lists differ: [\"\\n\\nUSER: What's the difference of two imag[339 chars]ama'] != [\"\\n \\nUSER: What's the difference of two ima[362 chars]the']"
|
| 104 |
+
},
|
| 105 |
+
{
|
| 106 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral",
|
| 107 |
+
"trace": "(line 543) RuntimeError: Input type (torch.FloatTensor) and weight type (CPUBFloat16Type) should be the same or input should be a MKLDNN tensor and weight is a dense tensor"
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_4bit",
|
| 111 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 40.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 16.69 MiB is free. Process 22505 has 22.28 GiB memory in use. Of the allocated memory 21.77 GiB is allocated by PyTorch, and 4.77 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 112 |
+
},
|
| 113 |
+
{
|
| 114 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_batched",
|
| 115 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 40.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 34.69 MiB is free. Process 22505 has 22.26 GiB memory in use. Of the allocated memory 21.73 GiB is allocated by PyTorch, and 30.40 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 116 |
+
},
|
| 117 |
+
{
|
| 118 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched",
|
| 119 |
+
"trace": "(line 404) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[269 chars] on'] != ['USER: \\nWhat are the things I should be cautio[271 chars] on']"
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched_regression",
|
| 123 |
+
"trace": "(line 507) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[280 chars]ed.'] != ['USER: \\nWhat are the things I should be cautio[283 chars]ed.']"
|
| 124 |
+
},
|
| 125 |
+
{
|
| 126 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_single",
|
| 127 |
+
"trace": "(line 353) AssertionError: 'USER: \\nWhat are the things I should be cautiou[748 chars]ies.' != 'USER: \\nWhat are the things I should be cautio[749 chars]ies.'"
|
| 128 |
+
},
|
| 129 |
+
{
|
| 130 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_tokenizer_integration",
|
| 131 |
+
"trace": "(line 579) AssertionError: Lists differ: ['<|im_start|>', '▁system', '\\n', '▁Answer', '▁the', '▁ques[176 chars]'\\n'] != ['<|im_start|>', 'system', '\\n', 'Answer', '▁the', '▁questi[175 chars]'\\n']"
|
| 132 |
+
}
|
| 133 |
+
]
|
| 134 |
+
},
|
| 135 |
+
"job_link": {
|
| 136 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144221",
|
| 137 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144228"
|
| 138 |
+
},
|
| 139 |
+
"captured_info": {
|
| 140 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144221#step:16:1",
|
| 141 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144228#step:16:1"
|
| 142 |
+
}
|
| 143 |
+
},
|
| 144 |
+
"models_llava_next": {
|
| 145 |
+
"failed": {
|
| 146 |
+
"PyTorch": {
|
| 147 |
+
"unclassified": 0,
|
| 148 |
+
"single": 2,
|
| 149 |
+
"multi": 2
|
| 150 |
+
},
|
| 151 |
+
"Tokenizers": {
|
| 152 |
+
"unclassified": 0,
|
| 153 |
+
"single": 0,
|
| 154 |
+
"multi": 0
|
| 155 |
+
},
|
| 156 |
+
"Pipelines": {
|
| 157 |
+
"unclassified": 0,
|
| 158 |
+
"single": 0,
|
| 159 |
+
"multi": 0
|
| 160 |
+
},
|
| 161 |
+
"Trainer": {
|
| 162 |
+
"unclassified": 0,
|
| 163 |
+
"single": 0,
|
| 164 |
+
"multi": 0
|
| 165 |
+
},
|
| 166 |
+
"ONNX": {
|
| 167 |
+
"unclassified": 0,
|
| 168 |
+
"single": 0,
|
| 169 |
+
"multi": 0
|
| 170 |
+
},
|
| 171 |
+
"Auto": {
|
| 172 |
+
"unclassified": 0,
|
| 173 |
+
"single": 0,
|
| 174 |
+
"multi": 0
|
| 175 |
+
},
|
| 176 |
+
"Quantization": {
|
| 177 |
+
"unclassified": 0,
|
| 178 |
+
"single": 0,
|
| 179 |
+
"multi": 0
|
| 180 |
+
},
|
| 181 |
+
"Unclassified": {
|
| 182 |
+
"unclassified": 0,
|
| 183 |
+
"single": 0,
|
| 184 |
+
"multi": 0
|
| 185 |
+
}
|
| 186 |
+
},
|
| 187 |
+
"errors": 0,
|
| 188 |
+
"success": 359,
|
| 189 |
+
"skipped": 139,
|
| 190 |
+
"time_spent": [
|
| 191 |
+
461.77,
|
| 192 |
+
477.34
|
| 193 |
+
],
|
| 194 |
+
"error": false,
|
| 195 |
+
"failures": {
|
| 196 |
+
"multi": [
|
| 197 |
+
{
|
| 198 |
+
"line": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 199 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 200 |
+
},
|
| 201 |
+
{
|
| 202 |
+
"line": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationIntegrationTest::test_small_model_integration_test_unk_token",
|
| 203 |
+
"trace": "(line 424) AssertionError: '[INS[24 chars]his image? [/INST] The image appears to be a [185 chars]hart' != '[INS[24 chars]his image? [/INST] The image appears to be a[186 chars]hart'"
|
| 204 |
+
}
|
| 205 |
+
],
|
| 206 |
+
"single": [
|
| 207 |
+
{
|
| 208 |
+
"line": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 209 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 210 |
+
},
|
| 211 |
+
{
|
| 212 |
+
"line": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationIntegrationTest::test_small_model_integration_test_unk_token",
|
| 213 |
+
"trace": "(line 424) AssertionError: '[INS[24 chars]his image? [/INST] The image appears to be a [185 chars]hart' != '[INS[24 chars]his image? [/INST] The image appears to be a[186 chars]hart'"
|
| 214 |
+
}
|
| 215 |
+
]
|
| 216 |
+
},
|
| 217 |
+
"job_link": {
|
| 218 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144225",
|
| 219 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144216"
|
| 220 |
+
},
|
| 221 |
+
"captured_info": {
|
| 222 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144225#step:16:1",
|
| 223 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144216#step:16:1"
|
| 224 |
+
}
|
| 225 |
+
},
|
| 226 |
+
"models_llava_next_video": {
|
| 227 |
+
"failed": {
|
| 228 |
+
"PyTorch": {
|
| 229 |
+
"unclassified": 0,
|
| 230 |
+
"single": 3,
|
| 231 |
+
"multi": 3
|
| 232 |
+
},
|
| 233 |
+
"Tokenizers": {
|
| 234 |
+
"unclassified": 0,
|
| 235 |
+
"single": 0,
|
| 236 |
+
"multi": 0
|
| 237 |
+
},
|
| 238 |
+
"Pipelines": {
|
| 239 |
+
"unclassified": 0,
|
| 240 |
+
"single": 0,
|
| 241 |
+
"multi": 0
|
| 242 |
+
},
|
| 243 |
+
"Trainer": {
|
| 244 |
+
"unclassified": 0,
|
| 245 |
+
"single": 0,
|
| 246 |
+
"multi": 0
|
| 247 |
+
},
|
| 248 |
+
"ONNX": {
|
| 249 |
+
"unclassified": 0,
|
| 250 |
+
"single": 0,
|
| 251 |
+
"multi": 0
|
| 252 |
+
},
|
| 253 |
+
"Auto": {
|
| 254 |
+
"unclassified": 0,
|
| 255 |
+
"single": 0,
|
| 256 |
+
"multi": 0
|
| 257 |
+
},
|
| 258 |
+
"Quantization": {
|
| 259 |
+
"unclassified": 0,
|
| 260 |
+
"single": 0,
|
| 261 |
+
"multi": 0
|
| 262 |
+
},
|
| 263 |
+
"Unclassified": {
|
| 264 |
+
"unclassified": 0,
|
| 265 |
+
"single": 0,
|
| 266 |
+
"multi": 0
|
| 267 |
+
}
|
| 268 |
+
},
|
| 269 |
+
"errors": 0,
|
| 270 |
+
"success": 365,
|
| 271 |
+
"skipped": 107,
|
| 272 |
+
"time_spent": [
|
| 273 |
+
234.03,
|
| 274 |
+
226.99
|
| 275 |
+
],
|
| 276 |
+
"error": false,
|
| 277 |
+
"failures": {
|
| 278 |
+
"single": [
|
| 279 |
+
{
|
| 280 |
+
"line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 281 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 282 |
+
},
|
| 283 |
+
{
|
| 284 |
+
"line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test",
|
| 285 |
+
"trace": "(line 377) AssertionError: 'USER[154 chars]hile wearing a pair of glasses that are too la[25 chars] are' != 'USER[154 chars]hile another child is attempting to read the s[45 chars]eems'"
|
| 286 |
+
},
|
| 287 |
+
{
|
| 288 |
+
"line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch_matches_single",
|
| 289 |
+
"trace": "(line 469) AssertionError: 'USER: \\nWhy is this video funny? ASSISTA[254 chars]them' != \"USER: \\nWhy is this video funny? ASSISTA[231 chars]eyes\""
|
| 290 |
+
}
|
| 291 |
+
],
|
| 292 |
+
"multi": [
|
| 293 |
+
{
|
| 294 |
+
"line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 295 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 296 |
+
},
|
| 297 |
+
{
|
| 298 |
+
"line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test",
|
| 299 |
+
"trace": "(line 377) AssertionError: 'USER[154 chars]hile wearing a pair of glasses that are too la[25 chars] are' != 'USER[154 chars]hile another child is attempting to read the s[45 chars]eems'"
|
| 300 |
+
},
|
| 301 |
+
{
|
| 302 |
+
"line": "tests/models/llava_next_video/test_modeling_llava_next_video.py::LlavaNextVideoForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch_matches_single",
|
| 303 |
+
"trace": "(line 469) AssertionError: 'USER: \\nWhy is this video funny? ASSISTA[254 chars]them' != \"USER: \\nWhy is this video funny? ASSISTA[231 chars]eyes\""
|
| 304 |
+
}
|
| 305 |
+
]
|
| 306 |
+
},
|
| 307 |
+
"job_link": {
|
| 308 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144232",
|
| 309 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144213"
|
| 310 |
+
},
|
| 311 |
+
"captured_info": {
|
| 312 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144232#step:16:1",
|
| 313 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144213#step:16:1"
|
| 314 |
+
}
|
| 315 |
+
},
|
| 316 |
+
"models_llava_onevision": {
|
| 317 |
+
"failed": {
|
| 318 |
+
"PyTorch": {
|
| 319 |
+
"unclassified": 0,
|
| 320 |
+
"single": 1,
|
| 321 |
+
"multi": 1
|
| 322 |
+
},
|
| 323 |
+
"Tokenizers": {
|
| 324 |
+
"unclassified": 0,
|
| 325 |
+
"single": 0,
|
| 326 |
+
"multi": 0
|
| 327 |
+
},
|
| 328 |
+
"Pipelines": {
|
| 329 |
+
"unclassified": 0,
|
| 330 |
+
"single": 0,
|
| 331 |
+
"multi": 0
|
| 332 |
+
},
|
| 333 |
+
"Trainer": {
|
| 334 |
+
"unclassified": 0,
|
| 335 |
+
"single": 0,
|
| 336 |
+
"multi": 0
|
| 337 |
+
},
|
| 338 |
+
"ONNX": {
|
| 339 |
+
"unclassified": 0,
|
| 340 |
+
"single": 0,
|
| 341 |
+
"multi": 0
|
| 342 |
+
},
|
| 343 |
+
"Auto": {
|
| 344 |
+
"unclassified": 0,
|
| 345 |
+
"single": 0,
|
| 346 |
+
"multi": 0
|
| 347 |
+
},
|
| 348 |
+
"Quantization": {
|
| 349 |
+
"unclassified": 0,
|
| 350 |
+
"single": 0,
|
| 351 |
+
"multi": 0
|
| 352 |
+
},
|
| 353 |
+
"Unclassified": {
|
| 354 |
+
"unclassified": 0,
|
| 355 |
+
"single": 2,
|
| 356 |
+
"multi": 2
|
| 357 |
+
}
|
| 358 |
+
},
|
| 359 |
+
"errors": 0,
|
| 360 |
+
"success": 411,
|
| 361 |
+
"skipped": 115,
|
| 362 |
+
"time_spent": [
|
| 363 |
+
179.8,
|
| 364 |
+
179.54
|
| 365 |
+
],
|
| 366 |
+
"error": false,
|
| 367 |
+
"failures": {
|
| 368 |
+
"multi": [
|
| 369 |
+
{
|
| 370 |
+
"line": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 371 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 372 |
+
},
|
| 373 |
+
{
|
| 374 |
+
"line": "tests/models/llava_onevision/test_processing_llava_onevision.py::LlavaOnevisionProcessorTest::test_apply_chat_template_video_0",
|
| 375 |
+
"trace": "(line 707) TypeError: Incorrect format used for video. Should be an url linking to an video or a local path."
|
| 376 |
+
},
|
| 377 |
+
{
|
| 378 |
+
"line": "tests/models/llava_onevision/test_processing_llava_onevision.py::LlavaOnevisionProcessorTest::test_apply_chat_template_video_1",
|
| 379 |
+
"trace": "(line 707) TypeError: Incorrect format used for video. Should be an url linking to an video or a local path."
|
| 380 |
+
}
|
| 381 |
+
],
|
| 382 |
+
"single": [
|
| 383 |
+
{
|
| 384 |
+
"line": "tests/models/llava_onevision/test_modeling_llava_onevision.py::LlavaOnevisionForConditionalGenerationModelTest::test_all_tensors_are_parameter_or_buffer",
|
| 385 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 386 |
+
},
|
| 387 |
+
{
|
| 388 |
+
"line": "tests/models/llava_onevision/test_processing_llava_onevision.py::LlavaOnevisionProcessorTest::test_apply_chat_template_video_0",
|
| 389 |
+
"trace": "(line 707) TypeError: Incorrect format used for video. Should be an url linking to an video or a local path."
|
| 390 |
+
},
|
| 391 |
+
{
|
| 392 |
+
"line": "tests/models/llava_onevision/test_processing_llava_onevision.py::LlavaOnevisionProcessorTest::test_apply_chat_template_video_1",
|
| 393 |
+
"trace": "(line 707) TypeError: Incorrect format used for video. Should be an url linking to an video or a local path."
|
| 394 |
+
}
|
| 395 |
+
]
|
| 396 |
+
},
|
| 397 |
+
"job_link": {
|
| 398 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144255",
|
| 399 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20423647274/job/58680144234"
|
| 400 |
+
},
|
| 401 |
+
"captured_info": {}
|
| 402 |
+
}
|
| 403 |
+
}
|