|
|
#!/bin/bash |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
cd /root/CVPR/MemGen |
|
|
|
|
|
|
|
|
export DEBUG_MODE="true" |
|
|
export LOG_PATH="./test_output/debug_log_math_vision.txt" |
|
|
export CUDA_VISIBLE_DEVICES="0,1,2,3,4,5,6,7" |
|
|
export MAIN_PROCESS_PORT="29508" |
|
|
export NCCL_DEBUG="WARN" |
|
|
export NCCL_IB_DISABLE="1" |
|
|
export NCCL_P2P_DISABLE="0" |
|
|
export NCCL_ASYNC_DISABLE="1" |
|
|
export TORCH_DISTRIBUTED_DEBUG="OFF" |
|
|
|
|
|
|
|
|
uv run python -m accelerate.commands.launch \ |
|
|
--num_processes=8 \ |
|
|
--main_process_port=29508 \ |
|
|
--config_file=configs/zero2.yaml \ |
|
|
main.py \ |
|
|
--cfg-path configs/latent_memory/math_vision.yaml \ |
|
|
--options \ |
|
|
model.reasoner_model_name "Qwen/Qwen2.5-VL-7B-Instruct" \ |
|
|
model.weaver.weaver_model_name "Qwen/Qwen2.5-1.5B-Instruct" \ |
|
|
model.trigger.trigger_model_name "null" \ |
|
|
model.weaver.prompt_latents_len "6" \ |
|
|
model.weaver.inference_latents_len "6" \ |
|
|
model.max_prompt_aug_num "1" \ |
|
|
model.max_inference_aug_num "2" \ |
|
|
model.load_model_path "null" \ |
|
|
run.mode "train" \ |
|
|
run.train_weaver "True" \ |
|
|
run.train_trigger "False" \ |
|
|
run.train_weaver_method "grpo" \ |
|
|
run.generation.do_sample "True" \ |
|
|
run.generation.temperature "1.0" \ |
|
|
run.generation.max_response_length "1024" \ |
|
|
run.weaver.grpo.batch_size "4" \ |
|
|
run.output_dir "/root/CVPR/MemGen/test_output/math_vision" \ |
|
|
datasets.math_vision.mode "grpo" |
|
|
|
|
|
|