diff --git "a/variant-c/variant_c_final.log" "b/variant-c/variant_c_final.log" new file mode 100644--- /dev/null +++ "b/variant-c/variant_c_final.log" @@ -0,0 +1,1613 @@ +W0222 16:33:50.453541 41281 warnings.py:110] /workspace/quip-sharp/lib/codebook/__init__.py:6: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. + @torch.library.impl_abstract("quip_lib::decode_matvec_e8p") + +W0222 16:33:50.454899 41281 warnings.py:110] /workspace/quip-sharp/lib/codebook/__init__.py:25: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. + @torch.library.impl_abstract("quip_lib::decompress_packed_e8p") + +I0222 16:33:50.570622 41281 utils.py:148] Note: detected 96 virtual cores but NumExpr set to maximum of 64, check "NUMEXPR_MAX_THREADS" environment variable. +I0222 16:33:50.570672 41281 utils.py:151] Note: NumExpr detected 96 cores but "NUMEXPR_MAX_THREADS" not set, so enforcing safe limit of 16. +I0222 16:33:50.570697 41281 utils.py:164] NumExpr defaulting to 16 threads. +W0222 16:33:51.252380 41281 warnings.py:110] /workspace/quip-sharp/lib/utils/matmul_had.py:92: FutureWarning: `torch.library.impl_abstract` was renamed to `torch.library.register_fake`. Please use that instead; we will remove `torch.library.impl_abstract` in a future version of PyTorch. + @torch.library.impl_abstract("quip_lib::hadamard") + +I0222 16:34:24.121408 41281 quantize_decompress_robust.py:121] Loaded codebook E8P12 +I0222 16:34:24.121847 41281 quantize_decompress_robust.py:126] Progress: 6/50 done, 44 remaining +I0222 16:34:24.121875 41281 quantize_decompress_robust.py:131] Loading base model... + Loading weights: 0%| | 0/453 [00:00