diff --git "a/debug.log" "b/debug.log" --- "a/debug.log" +++ "b/debug.log" @@ -1,10 +1,7 @@ - Loading dataset from disk: 0%| | 0/205 [00:00:39] [PID:52829] Skipping import of cpp extensions due to incompatible torch version 2.9.1+cu128 for torchao version 0.13.0 -[2026-02-10 06:24:34,917] [WARNING] [accelerate.utils.dataclasses.__post_init__:1962] [PID:52829] sharding_strategy is deprecated in favor of reshard_after_forward. This will be removed in a future version of Accelerate. -[2026-02-10 07:02:45,775] [WARNING] [py.warnings._showwarnmsg:110] [PID:52829] /root/miniconda3/envs/py3.11/lib/python3.11/site-packages/torch/distributed/fsdp/fully_sharded_data_parallel.py:675: FutureWarning: FSDP.state_dict_type() and FSDP.set_state_dict_type() are being deprecated. Please use APIs, get_state_dict() and set_state_dict(), which can support different parallelisms, FSDP1, FSDP2, DDP. API doc: https://pytorch.org/docs/stable/distributed.checkpoint.html#torch.distributed.checkpoint.state_dict.get_state_dict .Tutorial: https://pytorch.org/tutorials/recipes/distributed_checkpoint_recipe.html . - warnings.warn( - -[2026-02-10 07:40:55,009] [WARNING] [py.warnings._showwarnmsg:110] [PID:52829] /root/miniconda3/envs/py3.11/lib/python3.11/site-packages/torch/distributed/fsdp/fully_sharded_data_parallel.py:675: FutureWarning: FSDP.state_dict_type() and FSDP.set_state_dict_type() are being deprecated. Please use APIs, get_state_dict() and set_state_dict(), which can support different parallelisms, FSDP1, FSDP2, DDP. API doc: https://pytorch.org/docs/stable/distributed.checkpoint.html#torch.distributed.checkpoint.state_dict.get_state_dict .Tutorial: https://pytorch.org/tutorials/recipes/distributed_checkpoint_recipe.html . + Loading dataset from disk: 0%| | 0/205 [00:00:39] [PID:6544] Skipping import of cpp extensions due to incompatible torch version 2.9.1+cu128 for torchao version 0.13.0 +[2026-02-10 13:03:33,964] [WARNING] [accelerate.utils.dataclasses.__post_init__:1962] [PID:6544] sharding_strategy is deprecated in favor of reshard_after_forward. This will be removed in a future version of Accelerate. +[2026-02-10 13:41:29,309] [WARNING] [py.warnings._showwarnmsg:110] [PID:6544] /root/miniconda3/envs/py3.11/lib/python3.11/site-packages/torch/distributed/fsdp/fully_sharded_data_parallel.py:675: FutureWarning: FSDP.state_dict_type() and FSDP.set_state_dict_type() are being deprecated. Please use APIs, get_state_dict() and set_state_dict(), which can support different parallelisms, FSDP1, FSDP2, DDP. API doc: https://pytorch.org/docs/stable/distributed.checkpoint.html#torch.distributed.checkpoint.state_dict.get_state_dict .Tutorial: https://pytorch.org/tutorials/recipes/distributed_checkpoint_recipe.html . warnings.warn(