hf-transformers-bot commited on
Commit
f288200
·
verified ·
1 Parent(s): 3b6ecd3

Upload 2026-01-14/runs/26786-21002369414/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2026-01-14/runs/26786-21002369414/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,116 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_deepseek_v2": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 6,
7
+ "multi": 6
8
+ },
9
+ "Tokenizers": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Pipelines": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Trainer": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "ONNX": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Auto": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "Quantization": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Unclassified": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ }
44
+ },
45
+ "errors": 0,
46
+ "success": 267,
47
+ "skipped": 177,
48
+ "time_spent": [
49
+ 233.6,
50
+ 224.94
51
+ ],
52
+ "error": false,
53
+ "failures": {
54
+ "single": [
55
+ {
56
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_cpu_offload",
57
+ "trace": "(line 2847) AssertionError: Tensor-likes are not close!"
58
+ },
59
+ {
60
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_bin",
61
+ "trace": "(line 2765) AssertionError: Tensor-likes are not close!"
62
+ },
63
+ {
64
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_safetensors",
65
+ "trace": "(line 2805) AssertionError: Tensor-likes are not close!"
66
+ },
67
+ {
68
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_batch_fa2",
69
+ "trace": "(line 5113) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 704.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 182.69 MiB is free. Process 47930 has 22.12 GiB memory in use. Of the allocated memory 20.58 GiB is allocated by PyTorch, with 22.00 MiB allocated in private pools (e.g., CUDA Graphs), and 982.75 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
70
+ },
71
+ {
72
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_deepseek_v2_lite",
73
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 400.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 182.69 MiB is free. Process 47930 has 22.12 GiB memory in use. Of the allocated memory 20.58 GiB is allocated by PyTorch, with 22.00 MiB allocated in private pools (e.g., CUDA Graphs), and 982.75 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
74
+ },
75
+ {
76
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_logits_eager",
77
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 400.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 182.69 MiB is free. Process 47930 has 22.12 GiB memory in use. Of the allocated memory 21.18 GiB is allocated by PyTorch, with 22.00 MiB allocated in private pools (e.g., CUDA Graphs), and 369.74 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
78
+ }
79
+ ],
80
+ "multi": [
81
+ {
82
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_cpu_offload",
83
+ "trace": "(line 2847) AssertionError: Tensor-likes are not close!"
84
+ },
85
+ {
86
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_bin",
87
+ "trace": "(line 2765) AssertionError: Tensor-likes are not close!"
88
+ },
89
+ {
90
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2ModelTest::test_disk_offload_safetensors",
91
+ "trace": "(line 2805) AssertionError: Tensor-likes are not close!"
92
+ },
93
+ {
94
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_batch_fa2",
95
+ "trace": "(line 5113) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 704.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 224.69 MiB is free. Process 694893 has 22.08 GiB memory in use. Of the allocated memory 20.24 GiB is allocated by PyTorch, with 22.00 MiB allocated in private pools (e.g., CUDA Graphs), and 1.18 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
96
+ },
97
+ {
98
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_deepseek_v2_lite",
99
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 400.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 224.69 MiB is free. Process 694893 has 22.08 GiB memory in use. Of the allocated memory 20.24 GiB is allocated by PyTorch, with 22.00 MiB allocated in private pools (e.g., CUDA Graphs), and 1.18 GiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
100
+ },
101
+ {
102
+ "line": "tests/models/deepseek_v2/test_modeling_deepseek_v2.py::DeepseekV2IntegrationTest::test_logits_eager",
103
+ "trace": "(line 692) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 400.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 224.69 MiB is free. Process 694893 has 22.08 GiB memory in use. Of the allocated memory 20.72 GiB is allocated by PyTorch, with 22.00 MiB allocated in private pools (e.g., CUDA Graphs), and 714.86 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
104
+ }
105
+ ]
106
+ },
107
+ "job_link": {
108
+ "single": "https://github.com/huggingface/transformers/actions/runs/21002369414/job/60375777589",
109
+ "multi": "https://github.com/huggingface/transformers/actions/runs/21002369414/job/60375777586"
110
+ },
111
+ "captured_info": {
112
+ "single": "https://github.com/huggingface/transformers/actions/runs/21002369414/job/60375777589#step:16:1",
113
+ "multi": "https://github.com/huggingface/transformers/actions/runs/21002369414/job/60375777586#step:16:1"
114
+ }
115
+ }
116
+ }