Upload 2025-12-22/runs/25858-20428481246/ci_results_run_models_gpu/model_results.json with huggingface_hub
Browse files
2025-12-22/runs/25858-20428481246/ci_results_run_models_gpu/model_results.json
ADDED
|
@@ -0,0 +1,1972 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"models_afmoe": {
|
| 3 |
+
"failed": {
|
| 4 |
+
"PyTorch": {
|
| 5 |
+
"unclassified": 0,
|
| 6 |
+
"single": 2,
|
| 7 |
+
"multi": 2
|
| 8 |
+
},
|
| 9 |
+
"Tokenizers": {
|
| 10 |
+
"unclassified": 0,
|
| 11 |
+
"single": 0,
|
| 12 |
+
"multi": 0
|
| 13 |
+
},
|
| 14 |
+
"Pipelines": {
|
| 15 |
+
"unclassified": 0,
|
| 16 |
+
"single": 0,
|
| 17 |
+
"multi": 0
|
| 18 |
+
},
|
| 19 |
+
"Trainer": {
|
| 20 |
+
"unclassified": 0,
|
| 21 |
+
"single": 0,
|
| 22 |
+
"multi": 0
|
| 23 |
+
},
|
| 24 |
+
"ONNX": {
|
| 25 |
+
"unclassified": 0,
|
| 26 |
+
"single": 0,
|
| 27 |
+
"multi": 0
|
| 28 |
+
},
|
| 29 |
+
"Auto": {
|
| 30 |
+
"unclassified": 0,
|
| 31 |
+
"single": 0,
|
| 32 |
+
"multi": 0
|
| 33 |
+
},
|
| 34 |
+
"Quantization": {
|
| 35 |
+
"unclassified": 0,
|
| 36 |
+
"single": 0,
|
| 37 |
+
"multi": 0
|
| 38 |
+
},
|
| 39 |
+
"Unclassified": {
|
| 40 |
+
"unclassified": 0,
|
| 41 |
+
"single": 0,
|
| 42 |
+
"multi": 0
|
| 43 |
+
}
|
| 44 |
+
},
|
| 45 |
+
"errors": 0,
|
| 46 |
+
"success": 227,
|
| 47 |
+
"skipped": 223,
|
| 48 |
+
"time_spent": [
|
| 49 |
+
116.07,
|
| 50 |
+
117.97
|
| 51 |
+
],
|
| 52 |
+
"error": false,
|
| 53 |
+
"failures": {
|
| 54 |
+
"single": [
|
| 55 |
+
{
|
| 56 |
+
"line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_flex_attention_with_grads",
|
| 57 |
+
"trace": "(line 619) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_tem_fused_0 Required: 147968 Hardware limit:101376 Reducing block sizes or `num_stages` may help."
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_sdpa_can_dispatch_on_flash",
|
| 61 |
+
"trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
|
| 62 |
+
}
|
| 63 |
+
],
|
| 64 |
+
"multi": [
|
| 65 |
+
{
|
| 66 |
+
"line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_flex_attention_with_grads",
|
| 67 |
+
"trace": "(line 619) torch._inductor.exc.InductorError: RuntimeError: No valid triton configs. OutOfMemoryError: out of resource: triton_tem_fused_0 Required: 147968 Hardware limit:101376 Reducing block sizes or `num_stages` may help."
|
| 68 |
+
},
|
| 69 |
+
{
|
| 70 |
+
"line": "tests/models/afmoe/test_modeling_afmoe.py::AfmoeModelTest::test_sdpa_can_dispatch_on_flash",
|
| 71 |
+
"trace": "(line 92) RuntimeError: No available kernel. Aborting execution."
|
| 72 |
+
}
|
| 73 |
+
]
|
| 74 |
+
},
|
| 75 |
+
"job_link": {
|
| 76 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087524",
|
| 77 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087502"
|
| 78 |
+
},
|
| 79 |
+
"captured_info": {}
|
| 80 |
+
},
|
| 81 |
+
"models_aimv2": {
|
| 82 |
+
"failed": {
|
| 83 |
+
"PyTorch": {
|
| 84 |
+
"unclassified": 0,
|
| 85 |
+
"single": 2,
|
| 86 |
+
"multi": 2
|
| 87 |
+
},
|
| 88 |
+
"Tokenizers": {
|
| 89 |
+
"unclassified": 0,
|
| 90 |
+
"single": 0,
|
| 91 |
+
"multi": 0
|
| 92 |
+
},
|
| 93 |
+
"Pipelines": {
|
| 94 |
+
"unclassified": 0,
|
| 95 |
+
"single": 0,
|
| 96 |
+
"multi": 0
|
| 97 |
+
},
|
| 98 |
+
"Trainer": {
|
| 99 |
+
"unclassified": 0,
|
| 100 |
+
"single": 0,
|
| 101 |
+
"multi": 0
|
| 102 |
+
},
|
| 103 |
+
"ONNX": {
|
| 104 |
+
"unclassified": 0,
|
| 105 |
+
"single": 0,
|
| 106 |
+
"multi": 0
|
| 107 |
+
},
|
| 108 |
+
"Auto": {
|
| 109 |
+
"unclassified": 0,
|
| 110 |
+
"single": 0,
|
| 111 |
+
"multi": 0
|
| 112 |
+
},
|
| 113 |
+
"Quantization": {
|
| 114 |
+
"unclassified": 0,
|
| 115 |
+
"single": 0,
|
| 116 |
+
"multi": 0
|
| 117 |
+
},
|
| 118 |
+
"Unclassified": {
|
| 119 |
+
"unclassified": 0,
|
| 120 |
+
"single": 0,
|
| 121 |
+
"multi": 0
|
| 122 |
+
}
|
| 123 |
+
},
|
| 124 |
+
"errors": 0,
|
| 125 |
+
"success": 451,
|
| 126 |
+
"skipped": 341,
|
| 127 |
+
"time_spent": [
|
| 128 |
+
106.28,
|
| 129 |
+
133.03
|
| 130 |
+
],
|
| 131 |
+
"error": false,
|
| 132 |
+
"failures": {
|
| 133 |
+
"single": [
|
| 134 |
+
{
|
| 135 |
+
"line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2VisionModelTest::test_batching_equivalence",
|
| 136 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in Aimv2VisionModel for key=last_hidden_state."
|
| 137 |
+
},
|
| 138 |
+
{
|
| 139 |
+
"line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2ModelTest::test_batching_equivalence",
|
| 140 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in Aimv2Model for key=logits_per_image."
|
| 141 |
+
}
|
| 142 |
+
],
|
| 143 |
+
"multi": [
|
| 144 |
+
{
|
| 145 |
+
"line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2VisionModelTest::test_batching_equivalence",
|
| 146 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in Aimv2VisionModel for key=last_hidden_state."
|
| 147 |
+
},
|
| 148 |
+
{
|
| 149 |
+
"line": "tests/models/aimv2/test_modeling_aimv2.py::Aimv2ModelTest::test_batching_equivalence",
|
| 150 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in Aimv2Model for key=logits_per_image."
|
| 151 |
+
}
|
| 152 |
+
]
|
| 153 |
+
},
|
| 154 |
+
"job_link": {
|
| 155 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087639",
|
| 156 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087468"
|
| 157 |
+
},
|
| 158 |
+
"captured_info": {
|
| 159 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087639#step:16:1",
|
| 160 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087468#step:16:1"
|
| 161 |
+
}
|
| 162 |
+
},
|
| 163 |
+
"models_albert": {
|
| 164 |
+
"failed": {
|
| 165 |
+
"PyTorch": {
|
| 166 |
+
"unclassified": 0,
|
| 167 |
+
"single": 0,
|
| 168 |
+
"multi": 0
|
| 169 |
+
},
|
| 170 |
+
"Tokenizers": {
|
| 171 |
+
"unclassified": 0,
|
| 172 |
+
"single": 0,
|
| 173 |
+
"multi": 0
|
| 174 |
+
},
|
| 175 |
+
"Pipelines": {
|
| 176 |
+
"unclassified": 0,
|
| 177 |
+
"single": 0,
|
| 178 |
+
"multi": 0
|
| 179 |
+
},
|
| 180 |
+
"Trainer": {
|
| 181 |
+
"unclassified": 0,
|
| 182 |
+
"single": 0,
|
| 183 |
+
"multi": 0
|
| 184 |
+
},
|
| 185 |
+
"ONNX": {
|
| 186 |
+
"unclassified": 0,
|
| 187 |
+
"single": 0,
|
| 188 |
+
"multi": 0
|
| 189 |
+
},
|
| 190 |
+
"Auto": {
|
| 191 |
+
"unclassified": 0,
|
| 192 |
+
"single": 0,
|
| 193 |
+
"multi": 0
|
| 194 |
+
},
|
| 195 |
+
"Quantization": {
|
| 196 |
+
"unclassified": 0,
|
| 197 |
+
"single": 0,
|
| 198 |
+
"multi": 0
|
| 199 |
+
},
|
| 200 |
+
"Unclassified": {
|
| 201 |
+
"unclassified": 0,
|
| 202 |
+
"single": 0,
|
| 203 |
+
"multi": 0
|
| 204 |
+
}
|
| 205 |
+
},
|
| 206 |
+
"errors": 0,
|
| 207 |
+
"success": 282,
|
| 208 |
+
"skipped": 180,
|
| 209 |
+
"time_spent": [
|
| 210 |
+
160.03,
|
| 211 |
+
148.09
|
| 212 |
+
],
|
| 213 |
+
"error": false,
|
| 214 |
+
"failures": {},
|
| 215 |
+
"job_link": {
|
| 216 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087491",
|
| 217 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087551"
|
| 218 |
+
},
|
| 219 |
+
"captured_info": {}
|
| 220 |
+
},
|
| 221 |
+
"models_align": {
|
| 222 |
+
"failed": {
|
| 223 |
+
"PyTorch": {
|
| 224 |
+
"unclassified": 0,
|
| 225 |
+
"single": 0,
|
| 226 |
+
"multi": 2
|
| 227 |
+
},
|
| 228 |
+
"Tokenizers": {
|
| 229 |
+
"unclassified": 0,
|
| 230 |
+
"single": 0,
|
| 231 |
+
"multi": 0
|
| 232 |
+
},
|
| 233 |
+
"Pipelines": {
|
| 234 |
+
"unclassified": 0,
|
| 235 |
+
"single": 0,
|
| 236 |
+
"multi": 0
|
| 237 |
+
},
|
| 238 |
+
"Trainer": {
|
| 239 |
+
"unclassified": 0,
|
| 240 |
+
"single": 0,
|
| 241 |
+
"multi": 0
|
| 242 |
+
},
|
| 243 |
+
"ONNX": {
|
| 244 |
+
"unclassified": 0,
|
| 245 |
+
"single": 0,
|
| 246 |
+
"multi": 0
|
| 247 |
+
},
|
| 248 |
+
"Auto": {
|
| 249 |
+
"unclassified": 0,
|
| 250 |
+
"single": 0,
|
| 251 |
+
"multi": 0
|
| 252 |
+
},
|
| 253 |
+
"Quantization": {
|
| 254 |
+
"unclassified": 0,
|
| 255 |
+
"single": 0,
|
| 256 |
+
"multi": 0
|
| 257 |
+
},
|
| 258 |
+
"Unclassified": {
|
| 259 |
+
"unclassified": 0,
|
| 260 |
+
"single": 0,
|
| 261 |
+
"multi": 0
|
| 262 |
+
}
|
| 263 |
+
},
|
| 264 |
+
"errors": 0,
|
| 265 |
+
"success": 317,
|
| 266 |
+
"skipped": 585,
|
| 267 |
+
"time_spent": [
|
| 268 |
+
85.5,
|
| 269 |
+
94.66
|
| 270 |
+
],
|
| 271 |
+
"error": false,
|
| 272 |
+
"failures": {
|
| 273 |
+
"multi": [
|
| 274 |
+
{
|
| 275 |
+
"line": "tests/models/align/test_modeling_align.py::AlignTextModelTest::test_model_parallelism",
|
| 276 |
+
"trace": "(line 590) RuntimeError: Expected all tensors to be on the same device, but got mat2 is on cuda:1, different from other tensors on cuda:0 (when checking argument in method wrapper_CUDA_bmm)"
|
| 277 |
+
},
|
| 278 |
+
{
|
| 279 |
+
"line": "tests/models/align/test_modeling_align.py::AlignTextModelTest::test_multi_gpu_data_parallel_forward",
|
| 280 |
+
"trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
|
| 281 |
+
}
|
| 282 |
+
]
|
| 283 |
+
},
|
| 284 |
+
"job_link": {
|
| 285 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087515",
|
| 286 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087545"
|
| 287 |
+
},
|
| 288 |
+
"captured_info": {}
|
| 289 |
+
},
|
| 290 |
+
"models_altclip": {
|
| 291 |
+
"failed": {
|
| 292 |
+
"PyTorch": {
|
| 293 |
+
"unclassified": 0,
|
| 294 |
+
"single": 2,
|
| 295 |
+
"multi": 4
|
| 296 |
+
},
|
| 297 |
+
"Tokenizers": {
|
| 298 |
+
"unclassified": 0,
|
| 299 |
+
"single": 0,
|
| 300 |
+
"multi": 0
|
| 301 |
+
},
|
| 302 |
+
"Pipelines": {
|
| 303 |
+
"unclassified": 0,
|
| 304 |
+
"single": 0,
|
| 305 |
+
"multi": 0
|
| 306 |
+
},
|
| 307 |
+
"Trainer": {
|
| 308 |
+
"unclassified": 0,
|
| 309 |
+
"single": 0,
|
| 310 |
+
"multi": 0
|
| 311 |
+
},
|
| 312 |
+
"ONNX": {
|
| 313 |
+
"unclassified": 0,
|
| 314 |
+
"single": 0,
|
| 315 |
+
"multi": 0
|
| 316 |
+
},
|
| 317 |
+
"Auto": {
|
| 318 |
+
"unclassified": 0,
|
| 319 |
+
"single": 0,
|
| 320 |
+
"multi": 0
|
| 321 |
+
},
|
| 322 |
+
"Quantization": {
|
| 323 |
+
"unclassified": 0,
|
| 324 |
+
"single": 0,
|
| 325 |
+
"multi": 0
|
| 326 |
+
},
|
| 327 |
+
"Unclassified": {
|
| 328 |
+
"unclassified": 0,
|
| 329 |
+
"single": 0,
|
| 330 |
+
"multi": 0
|
| 331 |
+
}
|
| 332 |
+
},
|
| 333 |
+
"errors": 0,
|
| 334 |
+
"success": 316,
|
| 335 |
+
"skipped": 582,
|
| 336 |
+
"time_spent": [
|
| 337 |
+
240.24,
|
| 338 |
+
220.73
|
| 339 |
+
],
|
| 340 |
+
"error": false,
|
| 341 |
+
"failures": {
|
| 342 |
+
"multi": [
|
| 343 |
+
{
|
| 344 |
+
"line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPVisionModelTest::test_batching_equivalence",
|
| 345 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in AltCLIPVisionModel for key=last_hidden_state."
|
| 346 |
+
},
|
| 347 |
+
{
|
| 348 |
+
"line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPTextModelTest::test_multi_gpu_data_parallel_forward",
|
| 349 |
+
"trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
|
| 350 |
+
},
|
| 351 |
+
{
|
| 352 |
+
"line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_batching_equivalence",
|
| 353 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in AltCLIPModel for key=vision_model_output."
|
| 354 |
+
},
|
| 355 |
+
{
|
| 356 |
+
"line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_multi_gpu_data_parallel_forward",
|
| 357 |
+
"trace": "(line 769) StopIteration: Caught StopIteration in replica 1 on device 1."
|
| 358 |
+
}
|
| 359 |
+
],
|
| 360 |
+
"single": [
|
| 361 |
+
{
|
| 362 |
+
"line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPVisionModelTest::test_batching_equivalence",
|
| 363 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in AltCLIPVisionModel for key=last_hidden_state."
|
| 364 |
+
},
|
| 365 |
+
{
|
| 366 |
+
"line": "tests/models/altclip/test_modeling_altclip.py::AltCLIPModelTest::test_batching_equivalence",
|
| 367 |
+
"trace": "(line 1372) AssertionError: Batched and Single row outputs are not equal in AltCLIPModel for key=vision_model_output."
|
| 368 |
+
}
|
| 369 |
+
]
|
| 370 |
+
},
|
| 371 |
+
"job_link": {
|
| 372 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087477",
|
| 373 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087480"
|
| 374 |
+
},
|
| 375 |
+
"captured_info": {
|
| 376 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087477#step:16:1",
|
| 377 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087480#step:16:1"
|
| 378 |
+
}
|
| 379 |
+
},
|
| 380 |
+
"models_apertus": {
|
| 381 |
+
"failed": {
|
| 382 |
+
"PyTorch": {
|
| 383 |
+
"unclassified": 0,
|
| 384 |
+
"single": 0,
|
| 385 |
+
"multi": 0
|
| 386 |
+
},
|
| 387 |
+
"Tokenizers": {
|
| 388 |
+
"unclassified": 0,
|
| 389 |
+
"single": 0,
|
| 390 |
+
"multi": 0
|
| 391 |
+
},
|
| 392 |
+
"Pipelines": {
|
| 393 |
+
"unclassified": 0,
|
| 394 |
+
"single": 0,
|
| 395 |
+
"multi": 0
|
| 396 |
+
},
|
| 397 |
+
"Trainer": {
|
| 398 |
+
"unclassified": 0,
|
| 399 |
+
"single": 0,
|
| 400 |
+
"multi": 0
|
| 401 |
+
},
|
| 402 |
+
"ONNX": {
|
| 403 |
+
"unclassified": 0,
|
| 404 |
+
"single": 0,
|
| 405 |
+
"multi": 0
|
| 406 |
+
},
|
| 407 |
+
"Auto": {
|
| 408 |
+
"unclassified": 0,
|
| 409 |
+
"single": 0,
|
| 410 |
+
"multi": 0
|
| 411 |
+
},
|
| 412 |
+
"Quantization": {
|
| 413 |
+
"unclassified": 0,
|
| 414 |
+
"single": 0,
|
| 415 |
+
"multi": 0
|
| 416 |
+
},
|
| 417 |
+
"Unclassified": {
|
| 418 |
+
"unclassified": 0,
|
| 419 |
+
"single": 0,
|
| 420 |
+
"multi": 0
|
| 421 |
+
}
|
| 422 |
+
},
|
| 423 |
+
"errors": 0,
|
| 424 |
+
"success": 257,
|
| 425 |
+
"skipped": 197,
|
| 426 |
+
"time_spent": [
|
| 427 |
+
99.96,
|
| 428 |
+
98.93
|
| 429 |
+
],
|
| 430 |
+
"error": false,
|
| 431 |
+
"failures": {},
|
| 432 |
+
"job_link": {
|
| 433 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087488",
|
| 434 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087511"
|
| 435 |
+
},
|
| 436 |
+
"captured_info": {
|
| 437 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087488#step:16:1",
|
| 438 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087511#step:16:1"
|
| 439 |
+
}
|
| 440 |
+
},
|
| 441 |
+
"models_arcee": {
|
| 442 |
+
"failed": {
|
| 443 |
+
"PyTorch": {
|
| 444 |
+
"unclassified": 0,
|
| 445 |
+
"single": 0,
|
| 446 |
+
"multi": 0
|
| 447 |
+
},
|
| 448 |
+
"Tokenizers": {
|
| 449 |
+
"unclassified": 0,
|
| 450 |
+
"single": 0,
|
| 451 |
+
"multi": 0
|
| 452 |
+
},
|
| 453 |
+
"Pipelines": {
|
| 454 |
+
"unclassified": 0,
|
| 455 |
+
"single": 0,
|
| 456 |
+
"multi": 0
|
| 457 |
+
},
|
| 458 |
+
"Trainer": {
|
| 459 |
+
"unclassified": 0,
|
| 460 |
+
"single": 0,
|
| 461 |
+
"multi": 0
|
| 462 |
+
},
|
| 463 |
+
"ONNX": {
|
| 464 |
+
"unclassified": 0,
|
| 465 |
+
"single": 0,
|
| 466 |
+
"multi": 0
|
| 467 |
+
},
|
| 468 |
+
"Auto": {
|
| 469 |
+
"unclassified": 0,
|
| 470 |
+
"single": 0,
|
| 471 |
+
"multi": 0
|
| 472 |
+
},
|
| 473 |
+
"Quantization": {
|
| 474 |
+
"unclassified": 0,
|
| 475 |
+
"single": 0,
|
| 476 |
+
"multi": 0
|
| 477 |
+
},
|
| 478 |
+
"Unclassified": {
|
| 479 |
+
"unclassified": 0,
|
| 480 |
+
"single": 0,
|
| 481 |
+
"multi": 0
|
| 482 |
+
}
|
| 483 |
+
},
|
| 484 |
+
"errors": 0,
|
| 485 |
+
"success": 269,
|
| 486 |
+
"skipped": 193,
|
| 487 |
+
"time_spent": [
|
| 488 |
+
149.12,
|
| 489 |
+
151.39
|
| 490 |
+
],
|
| 491 |
+
"error": false,
|
| 492 |
+
"failures": {},
|
| 493 |
+
"job_link": {
|
| 494 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087514",
|
| 495 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087525"
|
| 496 |
+
},
|
| 497 |
+
"captured_info": {
|
| 498 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087514#step:16:1",
|
| 499 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087525#step:16:1"
|
| 500 |
+
}
|
| 501 |
+
},
|
| 502 |
+
"models_aria": {
|
| 503 |
+
"failed": {
|
| 504 |
+
"PyTorch": {
|
| 505 |
+
"unclassified": 0,
|
| 506 |
+
"single": 0,
|
| 507 |
+
"multi": 0
|
| 508 |
+
},
|
| 509 |
+
"Tokenizers": {
|
| 510 |
+
"unclassified": 0,
|
| 511 |
+
"single": 0,
|
| 512 |
+
"multi": 0
|
| 513 |
+
},
|
| 514 |
+
"Pipelines": {
|
| 515 |
+
"unclassified": 0,
|
| 516 |
+
"single": 0,
|
| 517 |
+
"multi": 0
|
| 518 |
+
},
|
| 519 |
+
"Trainer": {
|
| 520 |
+
"unclassified": 0,
|
| 521 |
+
"single": 0,
|
| 522 |
+
"multi": 0
|
| 523 |
+
},
|
| 524 |
+
"ONNX": {
|
| 525 |
+
"unclassified": 0,
|
| 526 |
+
"single": 0,
|
| 527 |
+
"multi": 0
|
| 528 |
+
},
|
| 529 |
+
"Auto": {
|
| 530 |
+
"unclassified": 0,
|
| 531 |
+
"single": 0,
|
| 532 |
+
"multi": 0
|
| 533 |
+
},
|
| 534 |
+
"Quantization": {
|
| 535 |
+
"unclassified": 0,
|
| 536 |
+
"single": 0,
|
| 537 |
+
"multi": 0
|
| 538 |
+
},
|
| 539 |
+
"Unclassified": {
|
| 540 |
+
"unclassified": 0,
|
| 541 |
+
"single": 0,
|
| 542 |
+
"multi": 0
|
| 543 |
+
}
|
| 544 |
+
},
|
| 545 |
+
"errors": 0,
|
| 546 |
+
"success": 316,
|
| 547 |
+
"skipped": 178,
|
| 548 |
+
"time_spent": [
|
| 549 |
+
220.81,
|
| 550 |
+
232.71
|
| 551 |
+
],
|
| 552 |
+
"error": false,
|
| 553 |
+
"failures": {},
|
| 554 |
+
"job_link": {
|
| 555 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087542",
|
| 556 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087519"
|
| 557 |
+
},
|
| 558 |
+
"captured_info": {}
|
| 559 |
+
},
|
| 560 |
+
"models_audioflamingo3": {
|
| 561 |
+
"failed": {
|
| 562 |
+
"PyTorch": {
|
| 563 |
+
"unclassified": 0,
|
| 564 |
+
"single": 0,
|
| 565 |
+
"multi": 0
|
| 566 |
+
},
|
| 567 |
+
"Tokenizers": {
|
| 568 |
+
"unclassified": 0,
|
| 569 |
+
"single": 0,
|
| 570 |
+
"multi": 0
|
| 571 |
+
},
|
| 572 |
+
"Pipelines": {
|
| 573 |
+
"unclassified": 0,
|
| 574 |
+
"single": 0,
|
| 575 |
+
"multi": 0
|
| 576 |
+
},
|
| 577 |
+
"Trainer": {
|
| 578 |
+
"unclassified": 0,
|
| 579 |
+
"single": 0,
|
| 580 |
+
"multi": 0
|
| 581 |
+
},
|
| 582 |
+
"ONNX": {
|
| 583 |
+
"unclassified": 0,
|
| 584 |
+
"single": 0,
|
| 585 |
+
"multi": 0
|
| 586 |
+
},
|
| 587 |
+
"Auto": {
|
| 588 |
+
"unclassified": 0,
|
| 589 |
+
"single": 0,
|
| 590 |
+
"multi": 0
|
| 591 |
+
},
|
| 592 |
+
"Quantization": {
|
| 593 |
+
"unclassified": 0,
|
| 594 |
+
"single": 0,
|
| 595 |
+
"multi": 0
|
| 596 |
+
},
|
| 597 |
+
"Unclassified": {
|
| 598 |
+
"unclassified": 0,
|
| 599 |
+
"single": 0,
|
| 600 |
+
"multi": 0
|
| 601 |
+
}
|
| 602 |
+
},
|
| 603 |
+
"errors": 0,
|
| 604 |
+
"success": 269,
|
| 605 |
+
"skipped": 161,
|
| 606 |
+
"time_spent": [
|
| 607 |
+
161.29,
|
| 608 |
+
145.32
|
| 609 |
+
],
|
| 610 |
+
"error": false,
|
| 611 |
+
"failures": {},
|
| 612 |
+
"job_link": {
|
| 613 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087490",
|
| 614 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087534"
|
| 615 |
+
},
|
| 616 |
+
"captured_info": {}
|
| 617 |
+
},
|
| 618 |
+
"models_auto": {
|
| 619 |
+
"failed": {
|
| 620 |
+
"PyTorch": {
|
| 621 |
+
"unclassified": 0,
|
| 622 |
+
"single": 0,
|
| 623 |
+
"multi": 0
|
| 624 |
+
},
|
| 625 |
+
"Tokenizers": {
|
| 626 |
+
"unclassified": 0,
|
| 627 |
+
"single": 3,
|
| 628 |
+
"multi": 3
|
| 629 |
+
},
|
| 630 |
+
"Pipelines": {
|
| 631 |
+
"unclassified": 0,
|
| 632 |
+
"single": 0,
|
| 633 |
+
"multi": 0
|
| 634 |
+
},
|
| 635 |
+
"Trainer": {
|
| 636 |
+
"unclassified": 0,
|
| 637 |
+
"single": 0,
|
| 638 |
+
"multi": 0
|
| 639 |
+
},
|
| 640 |
+
"ONNX": {
|
| 641 |
+
"unclassified": 0,
|
| 642 |
+
"single": 0,
|
| 643 |
+
"multi": 0
|
| 644 |
+
},
|
| 645 |
+
"Auto": {
|
| 646 |
+
"unclassified": 0,
|
| 647 |
+
"single": 0,
|
| 648 |
+
"multi": 0
|
| 649 |
+
},
|
| 650 |
+
"Quantization": {
|
| 651 |
+
"unclassified": 0,
|
| 652 |
+
"single": 0,
|
| 653 |
+
"multi": 0
|
| 654 |
+
},
|
| 655 |
+
"Unclassified": {
|
| 656 |
+
"unclassified": 0,
|
| 657 |
+
"single": 0,
|
| 658 |
+
"multi": 0
|
| 659 |
+
}
|
| 660 |
+
},
|
| 661 |
+
"errors": 0,
|
| 662 |
+
"success": 228,
|
| 663 |
+
"skipped": 14,
|
| 664 |
+
"time_spent": [
|
| 665 |
+
154.6,
|
| 666 |
+
182.05
|
| 667 |
+
],
|
| 668 |
+
"error": false,
|
| 669 |
+
"failures": {
|
| 670 |
+
"multi": [
|
| 671 |
+
{
|
| 672 |
+
"line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_custom_tokenizer_from_hub",
|
| 673 |
+
"trace": "(line 257) ValueError: Couldn't instantiate the backend tokenizer from one of:"
|
| 674 |
+
},
|
| 675 |
+
{
|
| 676 |
+
"line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_from_pretrained_dynamic_tokenizer",
|
| 677 |
+
"trace": "(line 163) AssertionError: ValueError not raised"
|
| 678 |
+
},
|
| 679 |
+
{
|
| 680 |
+
"line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_tokenizer_identifier_non_existent",
|
| 681 |
+
"trace": "(line 163) AssertionError: \"julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier\" does not match \"401 Client Error. (Request ID: Root=1-69491a01-16e22a3e78a66a3363e0ed8a;74765c98-8731-496d-b0c0-0ea1ef76966a)"
|
| 682 |
+
}
|
| 683 |
+
],
|
| 684 |
+
"single": [
|
| 685 |
+
{
|
| 686 |
+
"line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_custom_tokenizer_from_hub",
|
| 687 |
+
"trace": "(line 257) ValueError: Couldn't instantiate the backend tokenizer from one of:"
|
| 688 |
+
},
|
| 689 |
+
{
|
| 690 |
+
"line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_from_pretrained_dynamic_tokenizer",
|
| 691 |
+
"trace": "(line 163) AssertionError: ValueError not raised"
|
| 692 |
+
},
|
| 693 |
+
{
|
| 694 |
+
"line": "tests/models/auto/test_tokenization_auto.py::AutoTokenizerTest::test_tokenizer_identifier_non_existent",
|
| 695 |
+
"trace": "(line 163) AssertionError: \"julien-c/herlolip-not-exists is not a local folder and is not a valid model identifier\" does not match \"401 Client Error. (Request ID: Root=1-694919d7-5a23e2d40fedc5350454127a;f2ca7b39-6d08-4a0b-82c6-0dd56ac0ef4f)"
|
| 696 |
+
}
|
| 697 |
+
]
|
| 698 |
+
},
|
| 699 |
+
"job_link": {
|
| 700 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087483",
|
| 701 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087495"
|
| 702 |
+
},
|
| 703 |
+
"captured_info": {}
|
| 704 |
+
},
|
| 705 |
+
"models_bamba": {
|
| 706 |
+
"failed": {
|
| 707 |
+
"PyTorch": {
|
| 708 |
+
"unclassified": 0,
|
| 709 |
+
"single": 3,
|
| 710 |
+
"multi": 3
|
| 711 |
+
},
|
| 712 |
+
"Tokenizers": {
|
| 713 |
+
"unclassified": 0,
|
| 714 |
+
"single": 0,
|
| 715 |
+
"multi": 0
|
| 716 |
+
},
|
| 717 |
+
"Pipelines": {
|
| 718 |
+
"unclassified": 0,
|
| 719 |
+
"single": 0,
|
| 720 |
+
"multi": 0
|
| 721 |
+
},
|
| 722 |
+
"Trainer": {
|
| 723 |
+
"unclassified": 0,
|
| 724 |
+
"single": 0,
|
| 725 |
+
"multi": 0
|
| 726 |
+
},
|
| 727 |
+
"ONNX": {
|
| 728 |
+
"unclassified": 0,
|
| 729 |
+
"single": 0,
|
| 730 |
+
"multi": 0
|
| 731 |
+
},
|
| 732 |
+
"Auto": {
|
| 733 |
+
"unclassified": 0,
|
| 734 |
+
"single": 0,
|
| 735 |
+
"multi": 0
|
| 736 |
+
},
|
| 737 |
+
"Quantization": {
|
| 738 |
+
"unclassified": 0,
|
| 739 |
+
"single": 0,
|
| 740 |
+
"multi": 0
|
| 741 |
+
},
|
| 742 |
+
"Unclassified": {
|
| 743 |
+
"unclassified": 0,
|
| 744 |
+
"single": 0,
|
| 745 |
+
"multi": 0
|
| 746 |
+
}
|
| 747 |
+
},
|
| 748 |
+
"errors": 0,
|
| 749 |
+
"success": 211,
|
| 750 |
+
"skipped": 223,
|
| 751 |
+
"time_spent": [
|
| 752 |
+
75.8,
|
| 753 |
+
74.8
|
| 754 |
+
],
|
| 755 |
+
"error": false,
|
| 756 |
+
"failures": {
|
| 757 |
+
"single": [
|
| 758 |
+
{
|
| 759 |
+
"line": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_sdpa_can_compile_dynamic",
|
| 760 |
+
"trace": "(line 1677) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16)), **{'attn_mask': FakeTensor(..., device='cuda:0', size=(s84, 1, s34, s30), dtype=torch.float16), 'dropout_p': 0.0, 'scale': 0.353553390593274, 'is_causal': False}): got RuntimeError('expand: attempting to expand a dimension of length s30 -> s87!')"
|
| 761 |
+
},
|
| 762 |
+
{
|
| 763 |
+
"line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_batched_generate_with_padding",
|
| 764 |
+
"trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.40 GiB is free. Process 26017 has 18.90 GiB memory in use. Of the allocated memory 18.54 GiB is allocated by PyTorch, and 4.28 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 765 |
+
},
|
| 766 |
+
{
|
| 767 |
+
"line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_generate",
|
| 768 |
+
"trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.25 GiB is free. Process 26017 has 19.04 GiB memory in use. Of the allocated memory 18.67 GiB is allocated by PyTorch, and 11.38 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 769 |
+
}
|
| 770 |
+
],
|
| 771 |
+
"multi": [
|
| 772 |
+
{
|
| 773 |
+
"line": "tests/models/bamba/test_modeling_bamba.py::BambaModelTest::test_sdpa_can_compile_dynamic",
|
| 774 |
+
"trace": "(line 1677) torch._dynamo.exc.TorchRuntimeError: Dynamo failed to run FX node with fake tensors: call_function <built-in function scaled_dot_product_attention>(*(FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16), FakeTensor(..., device='cuda:0', size=(s47, 4, s87, 8), dtype=torch.float16)), **{'attn_mask': FakeTensor(..., device='cuda:0', size=(s84, 1, s34, s30), dtype=torch.float16), 'dropout_p': 0.0, 'scale': 0.353553390593274, 'is_causal': False}): got RuntimeError('expand: attempting to expand a dimension of length s30 -> s87!')"
|
| 775 |
+
},
|
| 776 |
+
{
|
| 777 |
+
"line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_batched_generate_with_padding",
|
| 778 |
+
"trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 8.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.27 GiB is free. Process 25692 has 19.03 GiB memory in use. Of the allocated memory 18.54 GiB is allocated by PyTorch, and 12.15 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 779 |
+
},
|
| 780 |
+
{
|
| 781 |
+
"line": "tests/models/bamba/test_modeling_bamba.py::BambaModelIntegrationTest::test_simple_generate",
|
| 782 |
+
"trace": "(line 904) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 4.00 GiB. GPU 0 has a total capacity of 22.30 GiB of which 3.12 GiB is free. Process 25692 has 19.17 GiB memory in use. Of the allocated memory 18.68 GiB is allocated by PyTorch, and 19.25 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 783 |
+
}
|
| 784 |
+
]
|
| 785 |
+
},
|
| 786 |
+
"job_link": {
|
| 787 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087498",
|
| 788 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087500"
|
| 789 |
+
},
|
| 790 |
+
"captured_info": {}
|
| 791 |
+
},
|
| 792 |
+
"models_bark": {
|
| 793 |
+
"failed": {
|
| 794 |
+
"PyTorch": {
|
| 795 |
+
"unclassified": 0,
|
| 796 |
+
"single": 1,
|
| 797 |
+
"multi": 1
|
| 798 |
+
},
|
| 799 |
+
"Tokenizers": {
|
| 800 |
+
"unclassified": 0,
|
| 801 |
+
"single": 0,
|
| 802 |
+
"multi": 0
|
| 803 |
+
},
|
| 804 |
+
"Pipelines": {
|
| 805 |
+
"unclassified": 0,
|
| 806 |
+
"single": 0,
|
| 807 |
+
"multi": 0
|
| 808 |
+
},
|
| 809 |
+
"Trainer": {
|
| 810 |
+
"unclassified": 0,
|
| 811 |
+
"single": 0,
|
| 812 |
+
"multi": 0
|
| 813 |
+
},
|
| 814 |
+
"ONNX": {
|
| 815 |
+
"unclassified": 0,
|
| 816 |
+
"single": 0,
|
| 817 |
+
"multi": 0
|
| 818 |
+
},
|
| 819 |
+
"Auto": {
|
| 820 |
+
"unclassified": 0,
|
| 821 |
+
"single": 0,
|
| 822 |
+
"multi": 0
|
| 823 |
+
},
|
| 824 |
+
"Quantization": {
|
| 825 |
+
"unclassified": 0,
|
| 826 |
+
"single": 0,
|
| 827 |
+
"multi": 0
|
| 828 |
+
},
|
| 829 |
+
"Unclassified": {
|
| 830 |
+
"unclassified": 0,
|
| 831 |
+
"single": 0,
|
| 832 |
+
"multi": 0
|
| 833 |
+
}
|
| 834 |
+
},
|
| 835 |
+
"errors": 0,
|
| 836 |
+
"success": 410,
|
| 837 |
+
"skipped": 480,
|
| 838 |
+
"time_spent": [
|
| 839 |
+
793.41,
|
| 840 |
+
748.43
|
| 841 |
+
],
|
| 842 |
+
"error": false,
|
| 843 |
+
"failures": {
|
| 844 |
+
"single": [
|
| 845 |
+
{
|
| 846 |
+
"line": "tests/models/bark/test_modeling_bark.py::BarkModelIntegrationTests::test_generate_end_to_end_with_offload",
|
| 847 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 848 |
+
}
|
| 849 |
+
],
|
| 850 |
+
"multi": [
|
| 851 |
+
{
|
| 852 |
+
"line": "tests/models/bark/test_modeling_bark.py::BarkModelIntegrationTests::test_generate_end_to_end_with_offload",
|
| 853 |
+
"trace": "(line 2542) RuntimeError: Expected all tensors to be on the same device, but got index is on cuda:0, different from other tensors on cpu (when checking argument in method wrapper_CUDA__index_select)"
|
| 854 |
+
}
|
| 855 |
+
]
|
| 856 |
+
},
|
| 857 |
+
"job_link": {
|
| 858 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087556",
|
| 859 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087478"
|
| 860 |
+
},
|
| 861 |
+
"captured_info": {}
|
| 862 |
+
},
|
| 863 |
+
"models_bart": {
|
| 864 |
+
"failed": {
|
| 865 |
+
"PyTorch": {
|
| 866 |
+
"unclassified": 0,
|
| 867 |
+
"single": 1,
|
| 868 |
+
"multi": 1
|
| 869 |
+
},
|
| 870 |
+
"Tokenizers": {
|
| 871 |
+
"unclassified": 0,
|
| 872 |
+
"single": 0,
|
| 873 |
+
"multi": 0
|
| 874 |
+
},
|
| 875 |
+
"Pipelines": {
|
| 876 |
+
"unclassified": 0,
|
| 877 |
+
"single": 0,
|
| 878 |
+
"multi": 0
|
| 879 |
+
},
|
| 880 |
+
"Trainer": {
|
| 881 |
+
"unclassified": 0,
|
| 882 |
+
"single": 0,
|
| 883 |
+
"multi": 0
|
| 884 |
+
},
|
| 885 |
+
"ONNX": {
|
| 886 |
+
"unclassified": 0,
|
| 887 |
+
"single": 0,
|
| 888 |
+
"multi": 0
|
| 889 |
+
},
|
| 890 |
+
"Auto": {
|
| 891 |
+
"unclassified": 0,
|
| 892 |
+
"single": 0,
|
| 893 |
+
"multi": 0
|
| 894 |
+
},
|
| 895 |
+
"Quantization": {
|
| 896 |
+
"unclassified": 0,
|
| 897 |
+
"single": 0,
|
| 898 |
+
"multi": 0
|
| 899 |
+
},
|
| 900 |
+
"Unclassified": {
|
| 901 |
+
"unclassified": 0,
|
| 902 |
+
"single": 0,
|
| 903 |
+
"multi": 0
|
| 904 |
+
}
|
| 905 |
+
},
|
| 906 |
+
"errors": 0,
|
| 907 |
+
"success": 484,
|
| 908 |
+
"skipped": 314,
|
| 909 |
+
"time_spent": [
|
| 910 |
+
212.49,
|
| 911 |
+
229.82
|
| 912 |
+
],
|
| 913 |
+
"error": false,
|
| 914 |
+
"failures": {
|
| 915 |
+
"single": [
|
| 916 |
+
{
|
| 917 |
+
"line": "tests/models/bart/test_modeling_bart.py::BartModelIntegrationTests::test_inference_no_head",
|
| 918 |
+
"trace": "(line 907) AssertionError: The values for attribute 'dtype' do not match: torch.float16 != torch.float32."
|
| 919 |
+
}
|
| 920 |
+
],
|
| 921 |
+
"multi": [
|
| 922 |
+
{
|
| 923 |
+
"line": "tests/models/bart/test_modeling_bart.py::BartModelIntegrationTests::test_inference_no_head",
|
| 924 |
+
"trace": "(line 907) AssertionError: The values for attribute 'dtype' do not match: torch.float16 != torch.float32."
|
| 925 |
+
}
|
| 926 |
+
]
|
| 927 |
+
},
|
| 928 |
+
"job_link": {
|
| 929 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087548",
|
| 930 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087503"
|
| 931 |
+
},
|
| 932 |
+
"captured_info": {
|
| 933 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087548#step:16:1",
|
| 934 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087503#step:16:1"
|
| 935 |
+
}
|
| 936 |
+
},
|
| 937 |
+
"models_bert": {
|
| 938 |
+
"failed": {
|
| 939 |
+
"PyTorch": {
|
| 940 |
+
"unclassified": 0,
|
| 941 |
+
"single": 0,
|
| 942 |
+
"multi": 0
|
| 943 |
+
},
|
| 944 |
+
"Tokenizers": {
|
| 945 |
+
"unclassified": 0,
|
| 946 |
+
"single": 0,
|
| 947 |
+
"multi": 0
|
| 948 |
+
},
|
| 949 |
+
"Pipelines": {
|
| 950 |
+
"unclassified": 0,
|
| 951 |
+
"single": 0,
|
| 952 |
+
"multi": 0
|
| 953 |
+
},
|
| 954 |
+
"Trainer": {
|
| 955 |
+
"unclassified": 0,
|
| 956 |
+
"single": 0,
|
| 957 |
+
"multi": 0
|
| 958 |
+
},
|
| 959 |
+
"ONNX": {
|
| 960 |
+
"unclassified": 0,
|
| 961 |
+
"single": 0,
|
| 962 |
+
"multi": 0
|
| 963 |
+
},
|
| 964 |
+
"Auto": {
|
| 965 |
+
"unclassified": 0,
|
| 966 |
+
"single": 0,
|
| 967 |
+
"multi": 0
|
| 968 |
+
},
|
| 969 |
+
"Quantization": {
|
| 970 |
+
"unclassified": 0,
|
| 971 |
+
"single": 0,
|
| 972 |
+
"multi": 0
|
| 973 |
+
},
|
| 974 |
+
"Unclassified": {
|
| 975 |
+
"unclassified": 0,
|
| 976 |
+
"single": 0,
|
| 977 |
+
"multi": 0
|
| 978 |
+
}
|
| 979 |
+
},
|
| 980 |
+
"errors": 0,
|
| 981 |
+
"success": 355,
|
| 982 |
+
"skipped": 211,
|
| 983 |
+
"time_spent": [
|
| 984 |
+
207.12,
|
| 985 |
+
219.81
|
| 986 |
+
],
|
| 987 |
+
"error": false,
|
| 988 |
+
"failures": {},
|
| 989 |
+
"job_link": {
|
| 990 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087571",
|
| 991 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087419"
|
| 992 |
+
},
|
| 993 |
+
"captured_info": {}
|
| 994 |
+
},
|
| 995 |
+
"models_bert_generation": {
|
| 996 |
+
"failed": {
|
| 997 |
+
"PyTorch": {
|
| 998 |
+
"unclassified": 0,
|
| 999 |
+
"single": 0,
|
| 1000 |
+
"multi": 0
|
| 1001 |
+
},
|
| 1002 |
+
"Tokenizers": {
|
| 1003 |
+
"unclassified": 0,
|
| 1004 |
+
"single": 0,
|
| 1005 |
+
"multi": 0
|
| 1006 |
+
},
|
| 1007 |
+
"Pipelines": {
|
| 1008 |
+
"unclassified": 0,
|
| 1009 |
+
"single": 0,
|
| 1010 |
+
"multi": 0
|
| 1011 |
+
},
|
| 1012 |
+
"Trainer": {
|
| 1013 |
+
"unclassified": 0,
|
| 1014 |
+
"single": 0,
|
| 1015 |
+
"multi": 0
|
| 1016 |
+
},
|
| 1017 |
+
"ONNX": {
|
| 1018 |
+
"unclassified": 0,
|
| 1019 |
+
"single": 0,
|
| 1020 |
+
"multi": 0
|
| 1021 |
+
},
|
| 1022 |
+
"Auto": {
|
| 1023 |
+
"unclassified": 0,
|
| 1024 |
+
"single": 0,
|
| 1025 |
+
"multi": 0
|
| 1026 |
+
},
|
| 1027 |
+
"Quantization": {
|
| 1028 |
+
"unclassified": 0,
|
| 1029 |
+
"single": 0,
|
| 1030 |
+
"multi": 0
|
| 1031 |
+
},
|
| 1032 |
+
"Unclassified": {
|
| 1033 |
+
"unclassified": 0,
|
| 1034 |
+
"single": 0,
|
| 1035 |
+
"multi": 0
|
| 1036 |
+
}
|
| 1037 |
+
},
|
| 1038 |
+
"errors": 0,
|
| 1039 |
+
"success": 347,
|
| 1040 |
+
"skipped": 219,
|
| 1041 |
+
"time_spent": [
|
| 1042 |
+
104.08,
|
| 1043 |
+
78.95
|
| 1044 |
+
],
|
| 1045 |
+
"error": false,
|
| 1046 |
+
"failures": {},
|
| 1047 |
+
"job_link": {
|
| 1048 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087433",
|
| 1049 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087496"
|
| 1050 |
+
},
|
| 1051 |
+
"captured_info": {}
|
| 1052 |
+
},
|
| 1053 |
+
"models_big_bird": {
|
| 1054 |
+
"failed": {
|
| 1055 |
+
"PyTorch": {
|
| 1056 |
+
"unclassified": 0,
|
| 1057 |
+
"single": 1,
|
| 1058 |
+
"multi": 1
|
| 1059 |
+
},
|
| 1060 |
+
"Tokenizers": {
|
| 1061 |
+
"unclassified": 0,
|
| 1062 |
+
"single": 0,
|
| 1063 |
+
"multi": 0
|
| 1064 |
+
},
|
| 1065 |
+
"Pipelines": {
|
| 1066 |
+
"unclassified": 0,
|
| 1067 |
+
"single": 0,
|
| 1068 |
+
"multi": 0
|
| 1069 |
+
},
|
| 1070 |
+
"Trainer": {
|
| 1071 |
+
"unclassified": 0,
|
| 1072 |
+
"single": 0,
|
| 1073 |
+
"multi": 0
|
| 1074 |
+
},
|
| 1075 |
+
"ONNX": {
|
| 1076 |
+
"unclassified": 0,
|
| 1077 |
+
"single": 0,
|
| 1078 |
+
"multi": 0
|
| 1079 |
+
},
|
| 1080 |
+
"Auto": {
|
| 1081 |
+
"unclassified": 0,
|
| 1082 |
+
"single": 0,
|
| 1083 |
+
"multi": 0
|
| 1084 |
+
},
|
| 1085 |
+
"Quantization": {
|
| 1086 |
+
"unclassified": 0,
|
| 1087 |
+
"single": 0,
|
| 1088 |
+
"multi": 0
|
| 1089 |
+
},
|
| 1090 |
+
"Unclassified": {
|
| 1091 |
+
"unclassified": 0,
|
| 1092 |
+
"single": 0,
|
| 1093 |
+
"multi": 0
|
| 1094 |
+
}
|
| 1095 |
+
},
|
| 1096 |
+
"errors": 0,
|
| 1097 |
+
"success": 242,
|
| 1098 |
+
"skipped": 246,
|
| 1099 |
+
"time_spent": [
|
| 1100 |
+
108.01,
|
| 1101 |
+
123.79
|
| 1102 |
+
],
|
| 1103 |
+
"error": false,
|
| 1104 |
+
"failures": {
|
| 1105 |
+
"multi": [
|
| 1106 |
+
{
|
| 1107 |
+
"line": "tests/models/big_bird/test_modeling_big_bird.py::BigBirdModelIntegrationTest::test_fill_mask",
|
| 1108 |
+
"trace": "(line 911) AssertionError: '' != 'happiness'"
|
| 1109 |
+
}
|
| 1110 |
+
],
|
| 1111 |
+
"single": [
|
| 1112 |
+
{
|
| 1113 |
+
"line": "tests/models/big_bird/test_modeling_big_bird.py::BigBirdModelIntegrationTest::test_fill_mask",
|
| 1114 |
+
"trace": "(line 911) AssertionError: '' != 'happiness'"
|
| 1115 |
+
}
|
| 1116 |
+
]
|
| 1117 |
+
},
|
| 1118 |
+
"job_link": {
|
| 1119 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087506",
|
| 1120 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087416"
|
| 1121 |
+
},
|
| 1122 |
+
"captured_info": {
|
| 1123 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087506#step:16:1",
|
| 1124 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087416#step:16:1"
|
| 1125 |
+
}
|
| 1126 |
+
},
|
| 1127 |
+
"models_gemma": {
|
| 1128 |
+
"failed": {
|
| 1129 |
+
"PyTorch": {
|
| 1130 |
+
"unclassified": 0,
|
| 1131 |
+
"single": 3,
|
| 1132 |
+
"multi": 3
|
| 1133 |
+
},
|
| 1134 |
+
"Tokenizers": {
|
| 1135 |
+
"unclassified": 0,
|
| 1136 |
+
"single": 0,
|
| 1137 |
+
"multi": 0
|
| 1138 |
+
},
|
| 1139 |
+
"Pipelines": {
|
| 1140 |
+
"unclassified": 0,
|
| 1141 |
+
"single": 0,
|
| 1142 |
+
"multi": 0
|
| 1143 |
+
},
|
| 1144 |
+
"Trainer": {
|
| 1145 |
+
"unclassified": 0,
|
| 1146 |
+
"single": 0,
|
| 1147 |
+
"multi": 0
|
| 1148 |
+
},
|
| 1149 |
+
"ONNX": {
|
| 1150 |
+
"unclassified": 0,
|
| 1151 |
+
"single": 0,
|
| 1152 |
+
"multi": 0
|
| 1153 |
+
},
|
| 1154 |
+
"Auto": {
|
| 1155 |
+
"unclassified": 0,
|
| 1156 |
+
"single": 0,
|
| 1157 |
+
"multi": 0
|
| 1158 |
+
},
|
| 1159 |
+
"Quantization": {
|
| 1160 |
+
"unclassified": 0,
|
| 1161 |
+
"single": 0,
|
| 1162 |
+
"multi": 0
|
| 1163 |
+
},
|
| 1164 |
+
"Unclassified": {
|
| 1165 |
+
"unclassified": 0,
|
| 1166 |
+
"single": 0,
|
| 1167 |
+
"multi": 0
|
| 1168 |
+
}
|
| 1169 |
+
},
|
| 1170 |
+
"errors": 0,
|
| 1171 |
+
"success": 383,
|
| 1172 |
+
"skipped": 197,
|
| 1173 |
+
"time_spent": [
|
| 1174 |
+
957.86,
|
| 1175 |
+
946.82
|
| 1176 |
+
],
|
| 1177 |
+
"error": false,
|
| 1178 |
+
"failures": {
|
| 1179 |
+
"single": [
|
| 1180 |
+
{
|
| 1181 |
+
"line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_export_static_cache",
|
| 1182 |
+
"trace": "(line 436) AssertionError: Lists differ: ['Hel[87 chars] in the 1990s. I have been looking on the internet and I have'] != ['Hel[87 chars] in the 1990s. I have looked on the internet and I have found']"
|
| 1183 |
+
},
|
| 1184 |
+
{
|
| 1185 |
+
"line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_4bit",
|
| 1186 |
+
"trace": "(line 329) AssertionError: Lists differ: ['Hel[42 chars]I am using a pic16f877a. I', 'Hi today I am go[68 chars]est'] != ['Hel[42 chars]I am trying to make a program that will take a[110 chars]!:)']"
|
| 1187 |
+
},
|
| 1188 |
+
{
|
| 1189 |
+
"line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_fp16_static_cache",
|
| 1190 |
+
"trace": "(line 299) AssertionError: Lists differ: ['Hel[27 chars]a 1999 4.0L 4x4. I', 'Hi today I am going to s[51 chars] 3D'] != ['Hel[27 chars]a 1995 3000gt SL. I have a', 'Hi today I am go[59 chars] 3D']"
|
| 1191 |
+
}
|
| 1192 |
+
],
|
| 1193 |
+
"multi": [
|
| 1194 |
+
{
|
| 1195 |
+
"line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_export_static_cache",
|
| 1196 |
+
"trace": "(line 436) AssertionError: Lists differ: ['Hel[87 chars] in the 1990s. I have been looking on the internet and I have'] != ['Hel[87 chars] in the 1990s. I have looked on the internet and I have found']"
|
| 1197 |
+
},
|
| 1198 |
+
{
|
| 1199 |
+
"line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_4bit",
|
| 1200 |
+
"trace": "(line 329) AssertionError: Lists differ: ['Hel[42 chars]I am using a pic16f877a. I', 'Hi today I am go[68 chars]est'] != ['Hel[42 chars]I am trying to make a program that will take a[110 chars]!:)']"
|
| 1201 |
+
},
|
| 1202 |
+
{
|
| 1203 |
+
"line": "tests/models/gemma/test_modeling_gemma.py::GemmaIntegrationTest::test_model_7b_fp16_static_cache",
|
| 1204 |
+
"trace": "(line 299) AssertionError: Lists differ: ['Hel[27 chars]a 1999 4.0L 4x4. I', 'Hi today I am going to s[51 chars] 3D'] != ['Hel[27 chars]a 1995 3000gt SL. I have a', 'Hi today I am go[59 chars] 3D']"
|
| 1205 |
+
}
|
| 1206 |
+
]
|
| 1207 |
+
},
|
| 1208 |
+
"job_link": {
|
| 1209 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087423",
|
| 1210 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087559"
|
| 1211 |
+
},
|
| 1212 |
+
"captured_info": {
|
| 1213 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087423#step:16:1",
|
| 1214 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087559#step:16:1"
|
| 1215 |
+
}
|
| 1216 |
+
},
|
| 1217 |
+
"models_gemma2": {
|
| 1218 |
+
"failed": {
|
| 1219 |
+
"PyTorch": {
|
| 1220 |
+
"unclassified": 0,
|
| 1221 |
+
"single": 4,
|
| 1222 |
+
"multi": 4
|
| 1223 |
+
},
|
| 1224 |
+
"Tokenizers": {
|
| 1225 |
+
"unclassified": 0,
|
| 1226 |
+
"single": 0,
|
| 1227 |
+
"multi": 0
|
| 1228 |
+
},
|
| 1229 |
+
"Pipelines": {
|
| 1230 |
+
"unclassified": 0,
|
| 1231 |
+
"single": 0,
|
| 1232 |
+
"multi": 0
|
| 1233 |
+
},
|
| 1234 |
+
"Trainer": {
|
| 1235 |
+
"unclassified": 0,
|
| 1236 |
+
"single": 0,
|
| 1237 |
+
"multi": 0
|
| 1238 |
+
},
|
| 1239 |
+
"ONNX": {
|
| 1240 |
+
"unclassified": 0,
|
| 1241 |
+
"single": 0,
|
| 1242 |
+
"multi": 0
|
| 1243 |
+
},
|
| 1244 |
+
"Auto": {
|
| 1245 |
+
"unclassified": 0,
|
| 1246 |
+
"single": 0,
|
| 1247 |
+
"multi": 0
|
| 1248 |
+
},
|
| 1249 |
+
"Quantization": {
|
| 1250 |
+
"unclassified": 0,
|
| 1251 |
+
"single": 0,
|
| 1252 |
+
"multi": 0
|
| 1253 |
+
},
|
| 1254 |
+
"Unclassified": {
|
| 1255 |
+
"unclassified": 0,
|
| 1256 |
+
"single": 1,
|
| 1257 |
+
"multi": 1
|
| 1258 |
+
}
|
| 1259 |
+
},
|
| 1260 |
+
"errors": 0,
|
| 1261 |
+
"success": 272,
|
| 1262 |
+
"skipped": 206,
|
| 1263 |
+
"time_spent": [
|
| 1264 |
+
585.94,
|
| 1265 |
+
456.96
|
| 1266 |
+
],
|
| 1267 |
+
"error": false,
|
| 1268 |
+
"failures": {
|
| 1269 |
+
"single": [
|
| 1270 |
+
{
|
| 1271 |
+
"line": "[100%]",
|
| 1272 |
+
"trace": "(line 2774) Failed: (subprocess) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1273 |
+
},
|
| 1274 |
+
{
|
| 1275 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_2b_pipeline_bf16_flex_attention",
|
| 1276 |
+
"trace": "(line 62) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1277 |
+
},
|
| 1278 |
+
{
|
| 1279 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_9b_bf16_flex_attention",
|
| 1280 |
+
"trace": "(line 224) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1281 |
+
},
|
| 1282 |
+
{
|
| 1283 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_9b_fp16",
|
| 1284 |
+
"trace": "(line 224) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1285 |
+
},
|
| 1286 |
+
{
|
| 1287 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_9b_pipeline_bf16",
|
| 1288 |
+
"trace": "Cannot retrieve error message."
|
| 1289 |
+
}
|
| 1290 |
+
],
|
| 1291 |
+
"multi": [
|
| 1292 |
+
{
|
| 1293 |
+
"line": "[100%]",
|
| 1294 |
+
"trace": "(line 2774) Failed: (subprocess) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1295 |
+
},
|
| 1296 |
+
{
|
| 1297 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_2b_pipeline_bf16_flex_attention",
|
| 1298 |
+
"trace": "(line 244) RuntimeError: CUDA driver error: an illegal memory access was encountered"
|
| 1299 |
+
},
|
| 1300 |
+
{
|
| 1301 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_9b_bf16_flex_attention",
|
| 1302 |
+
"trace": "(line 224) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1303 |
+
},
|
| 1304 |
+
{
|
| 1305 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_9b_fp16",
|
| 1306 |
+
"trace": "(line 224) torch.AcceleratorError: CUDA error: an illegal memory access was encountered"
|
| 1307 |
+
},
|
| 1308 |
+
{
|
| 1309 |
+
"line": "tests/models/gemma2/test_modeling_gemma2.py::Gemma2IntegrationTest::test_model_9b_pipeline_bf16",
|
| 1310 |
+
"trace": "Cannot retrieve error message."
|
| 1311 |
+
}
|
| 1312 |
+
]
|
| 1313 |
+
},
|
| 1314 |
+
"job_link": {
|
| 1315 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087474",
|
| 1316 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087547"
|
| 1317 |
+
},
|
| 1318 |
+
"captured_info": {}
|
| 1319 |
+
},
|
| 1320 |
+
"models_gemma3": {
|
| 1321 |
+
"failed": {
|
| 1322 |
+
"PyTorch": {
|
| 1323 |
+
"unclassified": 0,
|
| 1324 |
+
"single": 1,
|
| 1325 |
+
"multi": 1
|
| 1326 |
+
},
|
| 1327 |
+
"Tokenizers": {
|
| 1328 |
+
"unclassified": 0,
|
| 1329 |
+
"single": 0,
|
| 1330 |
+
"multi": 0
|
| 1331 |
+
},
|
| 1332 |
+
"Pipelines": {
|
| 1333 |
+
"unclassified": 0,
|
| 1334 |
+
"single": 0,
|
| 1335 |
+
"multi": 0
|
| 1336 |
+
},
|
| 1337 |
+
"Trainer": {
|
| 1338 |
+
"unclassified": 0,
|
| 1339 |
+
"single": 0,
|
| 1340 |
+
"multi": 0
|
| 1341 |
+
},
|
| 1342 |
+
"ONNX": {
|
| 1343 |
+
"unclassified": 0,
|
| 1344 |
+
"single": 0,
|
| 1345 |
+
"multi": 0
|
| 1346 |
+
},
|
| 1347 |
+
"Auto": {
|
| 1348 |
+
"unclassified": 0,
|
| 1349 |
+
"single": 0,
|
| 1350 |
+
"multi": 0
|
| 1351 |
+
},
|
| 1352 |
+
"Quantization": {
|
| 1353 |
+
"unclassified": 0,
|
| 1354 |
+
"single": 0,
|
| 1355 |
+
"multi": 0
|
| 1356 |
+
},
|
| 1357 |
+
"Unclassified": {
|
| 1358 |
+
"unclassified": 0,
|
| 1359 |
+
"single": 0,
|
| 1360 |
+
"multi": 0
|
| 1361 |
+
}
|
| 1362 |
+
},
|
| 1363 |
+
"errors": 0,
|
| 1364 |
+
"success": 592,
|
| 1365 |
+
"skipped": 360,
|
| 1366 |
+
"time_spent": [
|
| 1367 |
+
524.12,
|
| 1368 |
+
530.42
|
| 1369 |
+
],
|
| 1370 |
+
"error": false,
|
| 1371 |
+
"failures": {
|
| 1372 |
+
"multi": [
|
| 1373 |
+
{
|
| 1374 |
+
"line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_dynamic_sliding_window_is_default",
|
| 1375 |
+
"trace": "(line 676) AssertionError: 'DynamicSlidingWindowLayer' unexpectedly found in 'DynamicCache(layers=[DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer])'"
|
| 1376 |
+
}
|
| 1377 |
+
],
|
| 1378 |
+
"single": [
|
| 1379 |
+
{
|
| 1380 |
+
"line": "tests/models/gemma3/test_modeling_gemma3.py::Gemma3IntegrationTest::test_dynamic_sliding_window_is_default",
|
| 1381 |
+
"trace": "(line 676) AssertionError: 'DynamicSlidingWindowLayer' unexpectedly found in 'DynamicCache(layers=[DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer, DynamicLayer, DynamicSlidingWindowLayer, DynamicSlidingWindowLayer])'"
|
| 1382 |
+
}
|
| 1383 |
+
]
|
| 1384 |
+
},
|
| 1385 |
+
"job_link": {
|
| 1386 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087558",
|
| 1387 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087428"
|
| 1388 |
+
},
|
| 1389 |
+
"captured_info": {
|
| 1390 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087558#step:16:1",
|
| 1391 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087428#step:16:1"
|
| 1392 |
+
}
|
| 1393 |
+
},
|
| 1394 |
+
"models_llama": {
|
| 1395 |
+
"failed": {
|
| 1396 |
+
"PyTorch": {
|
| 1397 |
+
"unclassified": 0,
|
| 1398 |
+
"single": 0,
|
| 1399 |
+
"multi": 0
|
| 1400 |
+
},
|
| 1401 |
+
"Tokenizers": {
|
| 1402 |
+
"unclassified": 0,
|
| 1403 |
+
"single": 5,
|
| 1404 |
+
"multi": 5
|
| 1405 |
+
},
|
| 1406 |
+
"Pipelines": {
|
| 1407 |
+
"unclassified": 0,
|
| 1408 |
+
"single": 0,
|
| 1409 |
+
"multi": 0
|
| 1410 |
+
},
|
| 1411 |
+
"Trainer": {
|
| 1412 |
+
"unclassified": 0,
|
| 1413 |
+
"single": 0,
|
| 1414 |
+
"multi": 0
|
| 1415 |
+
},
|
| 1416 |
+
"ONNX": {
|
| 1417 |
+
"unclassified": 0,
|
| 1418 |
+
"single": 0,
|
| 1419 |
+
"multi": 0
|
| 1420 |
+
},
|
| 1421 |
+
"Auto": {
|
| 1422 |
+
"unclassified": 0,
|
| 1423 |
+
"single": 0,
|
| 1424 |
+
"multi": 0
|
| 1425 |
+
},
|
| 1426 |
+
"Quantization": {
|
| 1427 |
+
"unclassified": 0,
|
| 1428 |
+
"single": 0,
|
| 1429 |
+
"multi": 0
|
| 1430 |
+
},
|
| 1431 |
+
"Unclassified": {
|
| 1432 |
+
"unclassified": 0,
|
| 1433 |
+
"single": 0,
|
| 1434 |
+
"multi": 0
|
| 1435 |
+
}
|
| 1436 |
+
},
|
| 1437 |
+
"errors": 0,
|
| 1438 |
+
"success": 375,
|
| 1439 |
+
"skipped": 195,
|
| 1440 |
+
"time_spent": [
|
| 1441 |
+
309.66,
|
| 1442 |
+
320.09
|
| 1443 |
+
],
|
| 1444 |
+
"error": false,
|
| 1445 |
+
"failures": {
|
| 1446 |
+
"multi": [
|
| 1447 |
+
{
|
| 1448 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_added_tokens_serialization",
|
| 1449 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1450 |
+
},
|
| 1451 |
+
{
|
| 1452 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask",
|
| 1453 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1454 |
+
},
|
| 1455 |
+
{
|
| 1456 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
|
| 1457 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1458 |
+
},
|
| 1459 |
+
{
|
| 1460 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_padding_side_in_kwargs",
|
| 1461 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1462 |
+
},
|
| 1463 |
+
{
|
| 1464 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_truncation_side_in_kwargs",
|
| 1465 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1466 |
+
}
|
| 1467 |
+
],
|
| 1468 |
+
"single": [
|
| 1469 |
+
{
|
| 1470 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_added_tokens_serialization",
|
| 1471 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1472 |
+
},
|
| 1473 |
+
{
|
| 1474 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask",
|
| 1475 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1476 |
+
},
|
| 1477 |
+
{
|
| 1478 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_chat_template_return_assistant_tokens_mask_truncated",
|
| 1479 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1480 |
+
},
|
| 1481 |
+
{
|
| 1482 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_padding_side_in_kwargs",
|
| 1483 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1484 |
+
},
|
| 1485 |
+
{
|
| 1486 |
+
"line": "tests/models/llama/test_tokenization_llama.py::LlamaTokenizationTest::test_truncation_side_in_kwargs",
|
| 1487 |
+
"trace": "(line 483) OSError: You are trying to access a gated repo."
|
| 1488 |
+
}
|
| 1489 |
+
]
|
| 1490 |
+
},
|
| 1491 |
+
"job_link": {
|
| 1492 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087562",
|
| 1493 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087484"
|
| 1494 |
+
},
|
| 1495 |
+
"captured_info": {}
|
| 1496 |
+
},
|
| 1497 |
+
"models_llava": {
|
| 1498 |
+
"failed": {
|
| 1499 |
+
"PyTorch": {
|
| 1500 |
+
"unclassified": 0,
|
| 1501 |
+
"single": 11,
|
| 1502 |
+
"multi": 10
|
| 1503 |
+
},
|
| 1504 |
+
"Tokenizers": {
|
| 1505 |
+
"unclassified": 0,
|
| 1506 |
+
"single": 0,
|
| 1507 |
+
"multi": 0
|
| 1508 |
+
},
|
| 1509 |
+
"Pipelines": {
|
| 1510 |
+
"unclassified": 0,
|
| 1511 |
+
"single": 0,
|
| 1512 |
+
"multi": 0
|
| 1513 |
+
},
|
| 1514 |
+
"Trainer": {
|
| 1515 |
+
"unclassified": 0,
|
| 1516 |
+
"single": 0,
|
| 1517 |
+
"multi": 0
|
| 1518 |
+
},
|
| 1519 |
+
"ONNX": {
|
| 1520 |
+
"unclassified": 0,
|
| 1521 |
+
"single": 0,
|
| 1522 |
+
"multi": 0
|
| 1523 |
+
},
|
| 1524 |
+
"Auto": {
|
| 1525 |
+
"unclassified": 0,
|
| 1526 |
+
"single": 0,
|
| 1527 |
+
"multi": 0
|
| 1528 |
+
},
|
| 1529 |
+
"Quantization": {
|
| 1530 |
+
"unclassified": 0,
|
| 1531 |
+
"single": 0,
|
| 1532 |
+
"multi": 0
|
| 1533 |
+
},
|
| 1534 |
+
"Unclassified": {
|
| 1535 |
+
"unclassified": 0,
|
| 1536 |
+
"single": 0,
|
| 1537 |
+
"multi": 0
|
| 1538 |
+
}
|
| 1539 |
+
},
|
| 1540 |
+
"errors": 0,
|
| 1541 |
+
"success": 356,
|
| 1542 |
+
"skipped": 251,
|
| 1543 |
+
"time_spent": [
|
| 1544 |
+
292.01,
|
| 1545 |
+
298.65
|
| 1546 |
+
],
|
| 1547 |
+
"error": false,
|
| 1548 |
+
"failures": {
|
| 1549 |
+
"single": [
|
| 1550 |
+
{
|
| 1551 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation",
|
| 1552 |
+
"trace": "(line 562) AssertionError: Lists differ: [\"\\n\\nUSER: What's the difference of two imag[339 chars]ama'] != [\"\\n \\nUSER: What's the difference of two ima[362 chars]the']"
|
| 1553 |
+
},
|
| 1554 |
+
{
|
| 1555 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_generation_no_images",
|
| 1556 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 32.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 20.69 MiB is free. Process 76530 has 22.28 GiB memory in use. Of the allocated memory 21.87 GiB is allocated by PyTorch, and 25.34 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 1557 |
+
},
|
| 1558 |
+
{
|
| 1559 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral",
|
| 1560 |
+
"trace": "(line 543) RuntimeError: Input type (torch.FloatTensor) and weight type (CPUBFloat16Type) should be the same or input should be a MKLDNN tensor and weight is a dense tensor"
|
| 1561 |
+
},
|
| 1562 |
+
{
|
| 1563 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_4bit",
|
| 1564 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 50.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 44.69 MiB is free. Process 76530 has 22.25 GiB memory in use. Of the allocated memory 21.83 GiB is allocated by PyTorch, and 44.99 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 1565 |
+
},
|
| 1566 |
+
{
|
| 1567 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_batched",
|
| 1568 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 140.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 32.69 MiB is free. Process 76530 has 22.26 GiB memory in use. Of the allocated memory 21.86 GiB is allocated by PyTorch, and 18.76 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 1569 |
+
},
|
| 1570 |
+
{
|
| 1571 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test",
|
| 1572 |
+
"trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::BFloat16 != float"
|
| 1573 |
+
},
|
| 1574 |
+
{
|
| 1575 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch",
|
| 1576 |
+
"trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::BFloat16 != float"
|
| 1577 |
+
},
|
| 1578 |
+
{
|
| 1579 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched",
|
| 1580 |
+
"trace": "(line 404) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[269 chars] on'] != ['USER: \\nWhat are the things I should be cautio[271 chars] on']"
|
| 1581 |
+
},
|
| 1582 |
+
{
|
| 1583 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched_regression",
|
| 1584 |
+
"trace": "(line 507) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[280 chars]ed.'] != ['USER: \\nWhat are the things I should be cautio[283 chars]ed.']"
|
| 1585 |
+
},
|
| 1586 |
+
{
|
| 1587 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_single",
|
| 1588 |
+
"trace": "(line 353) AssertionError: 'USER: \\nWhat are the things I should be cautiou[748 chars]ies.' != 'USER: \\nWhat are the things I should be cautio[749 chars]ies.'"
|
| 1589 |
+
},
|
| 1590 |
+
{
|
| 1591 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_tokenizer_integration",
|
| 1592 |
+
"trace": "(line 579) AssertionError: Lists differ: ['<|im_start|>', '▁system', '\\n', '▁Answer', '▁the', '▁ques[176 chars]'\\n'] != ['<|im_start|>', 'system', '\\n', 'Answer', '▁the', '▁questi[175 chars]'\\n']"
|
| 1593 |
+
}
|
| 1594 |
+
],
|
| 1595 |
+
"multi": [
|
| 1596 |
+
{
|
| 1597 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_batched_generation",
|
| 1598 |
+
"trace": "(line 562) AssertionError: Lists differ: [\"\\n\\nUSER: What's the difference of two imag[339 chars]ama'] != [\"\\n \\nUSER: What's the difference of two ima[362 chars]the']"
|
| 1599 |
+
},
|
| 1600 |
+
{
|
| 1601 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral",
|
| 1602 |
+
"trace": "(line 543) RuntimeError: Input type (torch.FloatTensor) and weight type (CPUBFloat16Type) should be the same or input should be a MKLDNN tensor and weight is a dense tensor"
|
| 1603 |
+
},
|
| 1604 |
+
{
|
| 1605 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_4bit",
|
| 1606 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 40.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 12.69 MiB is free. Process 127830 has 22.28 GiB memory in use. Of the allocated memory 21.77 GiB is allocated by PyTorch, and 8.52 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 1607 |
+
},
|
| 1608 |
+
{
|
| 1609 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_pixtral_batched",
|
| 1610 |
+
"trace": "(line 693) torch.OutOfMemoryError: CUDA out of memory. Tried to allocate 140.00 MiB. GPU 0 has a total capacity of 22.30 GiB of which 126.69 MiB is free. Process 127830 has 22.17 GiB memory in use. Of the allocated memory 21.65 GiB is allocated by PyTorch, and 27.35 MiB is reserved by PyTorch but unallocated. If reserved but unallocated memory is large try setting PYTORCH_CUDA_ALLOC_CONF=expandable_segments:True to avoid fragmentation. See documentation for Memory Management (https://pytorch.org/docs/stable/notes/cuda.html#environment-variables)"
|
| 1611 |
+
},
|
| 1612 |
+
{
|
| 1613 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test",
|
| 1614 |
+
"trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::BFloat16 != float"
|
| 1615 |
+
},
|
| 1616 |
+
{
|
| 1617 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_batch",
|
| 1618 |
+
"trace": "(line 134) RuntimeError: expected mat1 and mat2 to have the same dtype, but got: c10::BFloat16 != float"
|
| 1619 |
+
},
|
| 1620 |
+
{
|
| 1621 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched",
|
| 1622 |
+
"trace": "(line 404) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[269 chars] on'] != ['USER: \\nWhat are the things I should be cautio[271 chars] on']"
|
| 1623 |
+
},
|
| 1624 |
+
{
|
| 1625 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_batched_regression",
|
| 1626 |
+
"trace": "(line 507) AssertionError: Lists differ: ['USER: \\nWhat are the things I should be cautiou[280 chars]ed.'] != ['USER: \\nWhat are the things I should be cautio[283 chars]ed.']"
|
| 1627 |
+
},
|
| 1628 |
+
{
|
| 1629 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_small_model_integration_test_llama_single",
|
| 1630 |
+
"trace": "(line 353) AssertionError: 'USER: \\nWhat are the things I should be cautiou[748 chars]ies.' != 'USER: \\nWhat are the things I should be cautio[749 chars]ies.'"
|
| 1631 |
+
},
|
| 1632 |
+
{
|
| 1633 |
+
"line": "tests/models/llava/test_modeling_llava.py::LlavaForConditionalGenerationIntegrationTest::test_tokenizer_integration",
|
| 1634 |
+
"trace": "(line 579) AssertionError: Lists differ: ['<|im_start|>', '▁system', '\\n', '▁Answer', '▁the', '▁ques[176 chars]'\\n'] != ['<|im_start|>', 'system', '\\n', 'Answer', '▁the', '▁questi[175 chars]'\\n']"
|
| 1635 |
+
}
|
| 1636 |
+
]
|
| 1637 |
+
},
|
| 1638 |
+
"job_link": {
|
| 1639 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087467",
|
| 1640 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087529"
|
| 1641 |
+
},
|
| 1642 |
+
"captured_info": {
|
| 1643 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087467#step:16:1",
|
| 1644 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087529#step:16:1"
|
| 1645 |
+
}
|
| 1646 |
+
},
|
| 1647 |
+
"models_llava_next": {
|
| 1648 |
+
"failed": {
|
| 1649 |
+
"PyTorch": {
|
| 1650 |
+
"unclassified": 0,
|
| 1651 |
+
"single": 1,
|
| 1652 |
+
"multi": 1
|
| 1653 |
+
},
|
| 1654 |
+
"Tokenizers": {
|
| 1655 |
+
"unclassified": 0,
|
| 1656 |
+
"single": 0,
|
| 1657 |
+
"multi": 0
|
| 1658 |
+
},
|
| 1659 |
+
"Pipelines": {
|
| 1660 |
+
"unclassified": 0,
|
| 1661 |
+
"single": 0,
|
| 1662 |
+
"multi": 0
|
| 1663 |
+
},
|
| 1664 |
+
"Trainer": {
|
| 1665 |
+
"unclassified": 0,
|
| 1666 |
+
"single": 0,
|
| 1667 |
+
"multi": 0
|
| 1668 |
+
},
|
| 1669 |
+
"ONNX": {
|
| 1670 |
+
"unclassified": 0,
|
| 1671 |
+
"single": 0,
|
| 1672 |
+
"multi": 0
|
| 1673 |
+
},
|
| 1674 |
+
"Auto": {
|
| 1675 |
+
"unclassified": 0,
|
| 1676 |
+
"single": 0,
|
| 1677 |
+
"multi": 0
|
| 1678 |
+
},
|
| 1679 |
+
"Quantization": {
|
| 1680 |
+
"unclassified": 0,
|
| 1681 |
+
"single": 0,
|
| 1682 |
+
"multi": 0
|
| 1683 |
+
},
|
| 1684 |
+
"Unclassified": {
|
| 1685 |
+
"unclassified": 0,
|
| 1686 |
+
"single": 0,
|
| 1687 |
+
"multi": 0
|
| 1688 |
+
}
|
| 1689 |
+
},
|
| 1690 |
+
"errors": 0,
|
| 1691 |
+
"success": 363,
|
| 1692 |
+
"skipped": 137,
|
| 1693 |
+
"time_spent": [
|
| 1694 |
+
960.3,
|
| 1695 |
+
974.72
|
| 1696 |
+
],
|
| 1697 |
+
"error": false,
|
| 1698 |
+
"failures": {
|
| 1699 |
+
"multi": [
|
| 1700 |
+
{
|
| 1701 |
+
"line": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationIntegrationTest::test_small_model_integration_test_unk_token",
|
| 1702 |
+
"trace": "(line 424) AssertionError: '[INS[24 chars]his image? [/INST] The image appears to be a [185 chars]hart' != '[INS[24 chars]his image? [/INST] The image appears to be a[186 chars]hart'"
|
| 1703 |
+
}
|
| 1704 |
+
],
|
| 1705 |
+
"single": [
|
| 1706 |
+
{
|
| 1707 |
+
"line": "tests/models/llava_next/test_modeling_llava_next.py::LlavaNextForConditionalGenerationIntegrationTest::test_small_model_integration_test_unk_token",
|
| 1708 |
+
"trace": "(line 424) AssertionError: '[INS[24 chars]his image? [/INST] The image appears to be a [185 chars]hart' != '[INS[24 chars]his image? [/INST] The image appears to be a[186 chars]hart'"
|
| 1709 |
+
}
|
| 1710 |
+
]
|
| 1711 |
+
},
|
| 1712 |
+
"job_link": {
|
| 1713 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087582",
|
| 1714 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087530"
|
| 1715 |
+
},
|
| 1716 |
+
"captured_info": {
|
| 1717 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087582#step:16:1",
|
| 1718 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087530#step:16:1"
|
| 1719 |
+
}
|
| 1720 |
+
},
|
| 1721 |
+
"models_mistral": {
|
| 1722 |
+
"failed": {
|
| 1723 |
+
"PyTorch": {
|
| 1724 |
+
"unclassified": 0,
|
| 1725 |
+
"single": 0,
|
| 1726 |
+
"multi": 0
|
| 1727 |
+
},
|
| 1728 |
+
"Tokenizers": {
|
| 1729 |
+
"unclassified": 0,
|
| 1730 |
+
"single": 0,
|
| 1731 |
+
"multi": 0
|
| 1732 |
+
},
|
| 1733 |
+
"Pipelines": {
|
| 1734 |
+
"unclassified": 0,
|
| 1735 |
+
"single": 0,
|
| 1736 |
+
"multi": 0
|
| 1737 |
+
},
|
| 1738 |
+
"Trainer": {
|
| 1739 |
+
"unclassified": 0,
|
| 1740 |
+
"single": 0,
|
| 1741 |
+
"multi": 0
|
| 1742 |
+
},
|
| 1743 |
+
"ONNX": {
|
| 1744 |
+
"unclassified": 0,
|
| 1745 |
+
"single": 0,
|
| 1746 |
+
"multi": 0
|
| 1747 |
+
},
|
| 1748 |
+
"Auto": {
|
| 1749 |
+
"unclassified": 0,
|
| 1750 |
+
"single": 0,
|
| 1751 |
+
"multi": 0
|
| 1752 |
+
},
|
| 1753 |
+
"Quantization": {
|
| 1754 |
+
"unclassified": 0,
|
| 1755 |
+
"single": 0,
|
| 1756 |
+
"multi": 0
|
| 1757 |
+
},
|
| 1758 |
+
"Unclassified": {
|
| 1759 |
+
"unclassified": 0,
|
| 1760 |
+
"single": 0,
|
| 1761 |
+
"multi": 0
|
| 1762 |
+
}
|
| 1763 |
+
},
|
| 1764 |
+
"errors": 0,
|
| 1765 |
+
"success": 295,
|
| 1766 |
+
"skipped": 187,
|
| 1767 |
+
"time_spent": [
|
| 1768 |
+
233.01,
|
| 1769 |
+
234.31
|
| 1770 |
+
],
|
| 1771 |
+
"error": false,
|
| 1772 |
+
"failures": {},
|
| 1773 |
+
"job_link": {
|
| 1774 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087520",
|
| 1775 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087561"
|
| 1776 |
+
},
|
| 1777 |
+
"captured_info": {
|
| 1778 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087520#step:16:1",
|
| 1779 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087561#step:16:1"
|
| 1780 |
+
}
|
| 1781 |
+
},
|
| 1782 |
+
"models_paligemma": {
|
| 1783 |
+
"failed": {
|
| 1784 |
+
"PyTorch": {
|
| 1785 |
+
"unclassified": 0,
|
| 1786 |
+
"single": 1,
|
| 1787 |
+
"multi": 1
|
| 1788 |
+
},
|
| 1789 |
+
"Tokenizers": {
|
| 1790 |
+
"unclassified": 0,
|
| 1791 |
+
"single": 0,
|
| 1792 |
+
"multi": 0
|
| 1793 |
+
},
|
| 1794 |
+
"Pipelines": {
|
| 1795 |
+
"unclassified": 0,
|
| 1796 |
+
"single": 0,
|
| 1797 |
+
"multi": 0
|
| 1798 |
+
},
|
| 1799 |
+
"Trainer": {
|
| 1800 |
+
"unclassified": 0,
|
| 1801 |
+
"single": 0,
|
| 1802 |
+
"multi": 0
|
| 1803 |
+
},
|
| 1804 |
+
"ONNX": {
|
| 1805 |
+
"unclassified": 0,
|
| 1806 |
+
"single": 0,
|
| 1807 |
+
"multi": 0
|
| 1808 |
+
},
|
| 1809 |
+
"Auto": {
|
| 1810 |
+
"unclassified": 0,
|
| 1811 |
+
"single": 0,
|
| 1812 |
+
"multi": 0
|
| 1813 |
+
},
|
| 1814 |
+
"Quantization": {
|
| 1815 |
+
"unclassified": 0,
|
| 1816 |
+
"single": 0,
|
| 1817 |
+
"multi": 0
|
| 1818 |
+
},
|
| 1819 |
+
"Unclassified": {
|
| 1820 |
+
"unclassified": 0,
|
| 1821 |
+
"single": 0,
|
| 1822 |
+
"multi": 0
|
| 1823 |
+
}
|
| 1824 |
+
},
|
| 1825 |
+
"errors": 0,
|
| 1826 |
+
"success": 282,
|
| 1827 |
+
"skipped": 164,
|
| 1828 |
+
"time_spent": [
|
| 1829 |
+
328.08,
|
| 1830 |
+
358.38
|
| 1831 |
+
],
|
| 1832 |
+
"error": false,
|
| 1833 |
+
"failures": {
|
| 1834 |
+
"multi": [
|
| 1835 |
+
{
|
| 1836 |
+
"line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_integration_detection_bug",
|
| 1837 |
+
"trace": "(line 676) AssertionError: 'detect shoe\\n<loc0045><loc0309><loc0708><loc0646> shoe' != 'detect shoe\\n<loc0051><loc0309><loc0708><loc0646> shoe'"
|
| 1838 |
+
}
|
| 1839 |
+
],
|
| 1840 |
+
"single": [
|
| 1841 |
+
{
|
| 1842 |
+
"line": "tests/models/paligemma/test_modeling_paligemma.py::PaliGemmaForConditionalGenerationIntegrationTest::test_integration_detection_bug",
|
| 1843 |
+
"trace": "(line 676) AssertionError: 'detect shoe\\n<loc0045><loc0309><loc0708><loc0646> shoe' != 'detect shoe\\n<loc0051><loc0309><loc0708><loc0646> shoe'"
|
| 1844 |
+
}
|
| 1845 |
+
]
|
| 1846 |
+
},
|
| 1847 |
+
"job_link": {
|
| 1848 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087607",
|
| 1849 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087446"
|
| 1850 |
+
},
|
| 1851 |
+
"captured_info": {
|
| 1852 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087607#step:16:1",
|
| 1853 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087446#step:16:1"
|
| 1854 |
+
}
|
| 1855 |
+
},
|
| 1856 |
+
"models_qwen2_vl": {
|
| 1857 |
+
"failed": {
|
| 1858 |
+
"PyTorch": {
|
| 1859 |
+
"unclassified": 0,
|
| 1860 |
+
"single": 0,
|
| 1861 |
+
"multi": 0
|
| 1862 |
+
},
|
| 1863 |
+
"Tokenizers": {
|
| 1864 |
+
"unclassified": 0,
|
| 1865 |
+
"single": 0,
|
| 1866 |
+
"multi": 0
|
| 1867 |
+
},
|
| 1868 |
+
"Pipelines": {
|
| 1869 |
+
"unclassified": 0,
|
| 1870 |
+
"single": 0,
|
| 1871 |
+
"multi": 0
|
| 1872 |
+
},
|
| 1873 |
+
"Trainer": {
|
| 1874 |
+
"unclassified": 0,
|
| 1875 |
+
"single": 0,
|
| 1876 |
+
"multi": 0
|
| 1877 |
+
},
|
| 1878 |
+
"ONNX": {
|
| 1879 |
+
"unclassified": 0,
|
| 1880 |
+
"single": 0,
|
| 1881 |
+
"multi": 0
|
| 1882 |
+
},
|
| 1883 |
+
"Auto": {
|
| 1884 |
+
"unclassified": 0,
|
| 1885 |
+
"single": 0,
|
| 1886 |
+
"multi": 0
|
| 1887 |
+
},
|
| 1888 |
+
"Quantization": {
|
| 1889 |
+
"unclassified": 0,
|
| 1890 |
+
"single": 0,
|
| 1891 |
+
"multi": 0
|
| 1892 |
+
},
|
| 1893 |
+
"Unclassified": {
|
| 1894 |
+
"unclassified": 0,
|
| 1895 |
+
"single": 0,
|
| 1896 |
+
"multi": 0
|
| 1897 |
+
}
|
| 1898 |
+
},
|
| 1899 |
+
"errors": 0,
|
| 1900 |
+
"success": 415,
|
| 1901 |
+
"skipped": 237,
|
| 1902 |
+
"time_spent": [
|
| 1903 |
+
311.12,
|
| 1904 |
+
320.23
|
| 1905 |
+
],
|
| 1906 |
+
"error": false,
|
| 1907 |
+
"failures": {},
|
| 1908 |
+
"job_link": {
|
| 1909 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087475",
|
| 1910 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087552"
|
| 1911 |
+
},
|
| 1912 |
+
"captured_info": {}
|
| 1913 |
+
},
|
| 1914 |
+
"models_t5gemma": {
|
| 1915 |
+
"failed": {
|
| 1916 |
+
"PyTorch": {
|
| 1917 |
+
"unclassified": 0,
|
| 1918 |
+
"single": 0,
|
| 1919 |
+
"multi": 0
|
| 1920 |
+
},
|
| 1921 |
+
"Tokenizers": {
|
| 1922 |
+
"unclassified": 0,
|
| 1923 |
+
"single": 0,
|
| 1924 |
+
"multi": 0
|
| 1925 |
+
},
|
| 1926 |
+
"Pipelines": {
|
| 1927 |
+
"unclassified": 0,
|
| 1928 |
+
"single": 0,
|
| 1929 |
+
"multi": 0
|
| 1930 |
+
},
|
| 1931 |
+
"Trainer": {
|
| 1932 |
+
"unclassified": 0,
|
| 1933 |
+
"single": 0,
|
| 1934 |
+
"multi": 0
|
| 1935 |
+
},
|
| 1936 |
+
"ONNX": {
|
| 1937 |
+
"unclassified": 0,
|
| 1938 |
+
"single": 0,
|
| 1939 |
+
"multi": 0
|
| 1940 |
+
},
|
| 1941 |
+
"Auto": {
|
| 1942 |
+
"unclassified": 0,
|
| 1943 |
+
"single": 0,
|
| 1944 |
+
"multi": 0
|
| 1945 |
+
},
|
| 1946 |
+
"Quantization": {
|
| 1947 |
+
"unclassified": 0,
|
| 1948 |
+
"single": 0,
|
| 1949 |
+
"multi": 0
|
| 1950 |
+
},
|
| 1951 |
+
"Unclassified": {
|
| 1952 |
+
"unclassified": 0,
|
| 1953 |
+
"single": 0,
|
| 1954 |
+
"multi": 0
|
| 1955 |
+
}
|
| 1956 |
+
},
|
| 1957 |
+
"errors": 0,
|
| 1958 |
+
"success": 419,
|
| 1959 |
+
"skipped": 285,
|
| 1960 |
+
"time_spent": [
|
| 1961 |
+
251.94,
|
| 1962 |
+
250.99
|
| 1963 |
+
],
|
| 1964 |
+
"error": false,
|
| 1965 |
+
"failures": {},
|
| 1966 |
+
"job_link": {
|
| 1967 |
+
"single": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087481",
|
| 1968 |
+
"multi": "https://github.com/huggingface/transformers/actions/runs/20428481246/job/58694087538"
|
| 1969 |
+
},
|
| 1970 |
+
"captured_info": {}
|
| 1971 |
+
}
|
| 1972 |
+
}
|