Delete VTimeLLM/.vscode/launch.json with huggingface_hub
Browse files- VTimeLLM/.vscode/launch.json +0 -102
VTimeLLM/.vscode/launch.json
DELETED
|
@@ -1,102 +0,0 @@
|
|
| 1 |
-
{
|
| 2 |
-
"version": "0.2.0",
|
| 3 |
-
"configurations": [
|
| 4 |
-
{
|
| 5 |
-
"name": "Debug MDPO Training",
|
| 6 |
-
"type": "python",
|
| 7 |
-
"request": "launch",
|
| 8 |
-
"program": "${workspaceFolder}/scripts/train/debug_mdpo_train.sh",
|
| 9 |
-
"console": "integratedTerminal",
|
| 10 |
-
"cwd": "${workspaceFolder}",
|
| 11 |
-
"env": {
|
| 12 |
-
"PYTHONPATH": "${workspaceFolder}:${env:PYTHONPATH}",
|
| 13 |
-
"CUDA_VISIBLE_DEVICES": "1",
|
| 14 |
-
"TRANSFORMERS_OFFLINE": "1",
|
| 15 |
-
"WANDB_PROJECT": "vtimellm"
|
| 16 |
-
},
|
| 17 |
-
"args": [],
|
| 18 |
-
"justMyCode": false,
|
| 19 |
-
"stopOnEntry": false
|
| 20 |
-
},
|
| 21 |
-
{
|
| 22 |
-
"name": "Debug Python Training Direct",
|
| 23 |
-
"type": "python",
|
| 24 |
-
"request": "launch",
|
| 25 |
-
"module": "vtimellm.train.train_dpo_mem",
|
| 26 |
-
"console": "integratedTerminal",
|
| 27 |
-
"cwd": "${workspaceFolder}",
|
| 28 |
-
"env": {
|
| 29 |
-
"PYTHONPATH": "${workspaceFolder}:${env:PYTHONPATH}",
|
| 30 |
-
"CUDA_VISIBLE_DEVICES": "1",
|
| 31 |
-
"TRANSFORMERS_OFFLINE": "1",
|
| 32 |
-
"WANDB_PROJECT": "vtimellm",
|
| 33 |
-
"RANK": "1",
|
| 34 |
-
"MASTER_PORT": "29571"
|
| 35 |
-
},
|
| 36 |
-
"args": [
|
| 37 |
-
"--deepspeed", "./scripts/zero2.json",
|
| 38 |
-
"--lora_enable", "True",
|
| 39 |
-
"--lora_r", "8",
|
| 40 |
-
"--lora_alpha", "128",
|
| 41 |
-
"--training_stage", "3",
|
| 42 |
-
"--finetuning", "True",
|
| 43 |
-
"--model_name_or_path", "./checkpoints/vicuna-7b-v1.5",
|
| 44 |
-
"--version", "v1",
|
| 45 |
-
"--data_path", "./data/activitynet/mdpo-train.json",
|
| 46 |
-
"--data_folder", "./data/activitynet/videos/train",
|
| 47 |
-
"--feat_folder", "./data/activitynet/clipvitl14-vtimellm.pth",
|
| 48 |
-
"--pretrain_mm_mlp_adapter", "./checkpoints/vtimellm-vicuna-v1-5-7b-stage1/mm_projector.bin",
|
| 49 |
-
"--stage2_path", "./checkpoints/vtimellm-vicuna-v1-5-7b-stage2",
|
| 50 |
-
"--stage3_path", "./checkpoints/vtimellm-vicuna-v1-5-7b-stage3",
|
| 51 |
-
"--stage4_path", "checkpoints/vtimellm-vicuna-v1-5-7b-activitynet-stage4",
|
| 52 |
-
"--output_dir", "./outputs/vtimellm-vicuna-v1-5-7b-activitynet-stage5",
|
| 53 |
-
"--bf16", "True",
|
| 54 |
-
"--max_steps", "100",
|
| 55 |
-
"--per_device_train_batch_size", "2",
|
| 56 |
-
"--gradient_accumulation_steps", "4",
|
| 57 |
-
"--evaluation_strategy", "no",
|
| 58 |
-
"--save_strategy", "no",
|
| 59 |
-
"--save_steps", "50000",
|
| 60 |
-
"--save_total_limit", "10",
|
| 61 |
-
"--learning_rate", "1e-6",
|
| 62 |
-
"--freeze_mm_mlp_adapter", "True",
|
| 63 |
-
"--weight_decay", "0.",
|
| 64 |
-
"--warmup_ratio", "0.1",
|
| 65 |
-
"--lr_scheduler_type", "cosine",
|
| 66 |
-
"--logging_steps", "1",
|
| 67 |
-
"--tf32", "True",
|
| 68 |
-
"--model_max_length", "2048",
|
| 69 |
-
"--gradient_checkpointing", "True",
|
| 70 |
-
"--dataloader_num_workers", "4",
|
| 71 |
-
"--lazy_preprocess", "True",
|
| 72 |
-
"--report_to", "none",
|
| 73 |
-
"--run_name", "vtimellm-vicuna-v1-5-7b-activitynet-stage5",
|
| 74 |
-
"--gamma", "0.0",
|
| 75 |
-
"--beta", "0.5",
|
| 76 |
-
"--dpo_alpha", "1.0",
|
| 77 |
-
"--train4dpo"
|
| 78 |
-
],
|
| 79 |
-
"justMyCode": false,
|
| 80 |
-
"stopOnEntry": false
|
| 81 |
-
},
|
| 82 |
-
{
|
| 83 |
-
"name": "Debug with DeepSpeed",
|
| 84 |
-
"type": "python",
|
| 85 |
-
"request": "launch",
|
| 86 |
-
"program": "${workspaceFolder}/scripts/train/debug_deepspeed.py",
|
| 87 |
-
"console": "integratedTerminal",
|
| 88 |
-
"cwd": "${workspaceFolder}",
|
| 89 |
-
"env": {
|
| 90 |
-
"PYTHONPATH": "${workspaceFolder}:${env:PYTHONPATH}",
|
| 91 |
-
"CUDA_VISIBLE_DEVICES": "1",
|
| 92 |
-
"TRANSFORMERS_OFFLINE": "1",
|
| 93 |
-
"WANDB_PROJECT": "vtimellm",
|
| 94 |
-
"RANK": "1",
|
| 95 |
-
"MASTER_PORT": "29571"
|
| 96 |
-
},
|
| 97 |
-
"args": [],
|
| 98 |
-
"justMyCode": false,
|
| 99 |
-
"stopOnEntry": false
|
| 100 |
-
}
|
| 101 |
-
]
|
| 102 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|