File size: 13,850 Bytes
fca4fc0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413

import os
import sys
import argparse
import json
import torch
from dataclasses import dataclass, field
from typing import Optional, Dict, Any

# Add project root to path
current_dir = os.getcwd()
if 'VidChain' in current_dir:
    root_dir = os.path.join(current_dir, "..", "..")
else:
    root_dir = current_dir
sys.path.append(root_dir)

try:
    import transformers
    from transformers import HfArgumentParser
    from vtimellm.train.train_dpo import (
        ModelArguments, 
        TrainingArguments, 
        DataArguments,
        train
    )
    from vtimellm.train.dataset import make_supervised_dpo_data_module
    from vtimellm.model import VTimeLLMLlamaForCausalLM
    from vtimellm.model.builder import load_lora
    from vtimellm.mm_utils import print_trainable_parameters
    from trl.trl.trainer import DPOTrainer
    print("✓ Training modules imported successfully")
except ImportError as e:
    print(f"✗ Error importing training modules: {e}")
    print("Make sure you're in the correct environment and all dependencies are installed")
    sys.exit(1)


@dataclass
class DPOConfig:
    """Configuration class for DPO training demo"""
    
    # Model paths
    model_name_or_path: str = "checkpoints/vicuna-7b-v1.5"
    stage2_path: str = "checkpoints/vtimellm-vicuna-v1-5-7b-stage2"
    stage3_path: str = "checkpoints/vtimellm-vicuna-v1-5-7b-stage3"
    stage4_path: str = "checkpoints/vtimellm-vicuna-v1-5-7b-activitynet-stage4"
    pretrain_mm_mlp_adapter: str = "checkpoints/vtimellm-vicuna-v1-5-7b-stage1/mm_projector.bin"
    
    # Data paths
    data_path: str = "data/activitynet/mdpo-train.json"
    data_folder: str = "data/activitynet/videos/train"
    feat_folder: str = "data/activitynet/clipvitl14-vtimellm.pth"
    
    # Training parameters
    output_dir: str = "outputs/vtimellm-dpo-demo"
    num_train_epochs: int = 1
    per_device_train_batch_size: int = 2
    gradient_accumulation_steps: int = 4
    learning_rate: float = 1e-6
    model_max_length: int = 2048
    
    # DPO parameters
    beta: float = 0.5  # DPO beta parameter (controls preference strength)
    dpo_alpha: float = 1.0  # DPO alpha parameter
    gamma: float = 0.0  # Additional regularization parameter
    
    # LoRA parameters
    lora_enable: bool = True
    lora_r: int = 64
    lora_alpha: int = 128
    lora_dropout: float = 0.05
    
    # Other settings
    bf16: bool = True
    gradient_checkpointing: bool = True
    dataloader_num_workers: int = 4
    logging_steps: int = 10
    save_steps: int = 500
    save_total_limit: int = 3
    warmup_ratio: float = 0.1
    weight_decay: float = 0.0
    freeze_mm_mlp_adapter: bool = True


def check_training_requirements(config: DPOConfig) -> bool:
    """Check if all requirements for training are met"""
    print("🔍 Checking training requirements...")
    
    # Check model files
    model_files = [
        config.model_name_or_path,
        config.stage2_path,
        config.stage3_path,
        config.stage4_path,
        config.pretrain_mm_mlp_adapter
    ]
    
    missing_files = []
    for file_path in model_files:
        if not os.path.exists(file_path):
            missing_files.append(file_path)
    
    if missing_files:
        print("❌ Missing model files:")
        for file_path in missing_files:
            print(f"  - {file_path}")
        print("\nPlease download the required model checkpoints.")
        return False
    
    # Check data files
    if not os.path.exists(config.data_path):
        print(f"❌ Training data not found: {config.data_path}")
        print("Please prepare your DPO training data.")
        return False
    
    # Check GPU
    if not torch.cuda.is_available():
        print("⚠ No CUDA GPU available. Training will be very slow on CPU.")
    else:
        gpu_memory = torch.cuda.get_device_properties(0).total_memory / 1024**3
        print(f"✓ GPU available: {torch.cuda.get_device_name(0)} ({gpu_memory:.1f} GB)")
        
        if gpu_memory < 16:
            print("⚠ Warning: GPU memory is less than 16GB. Consider reducing batch size.")
    
    # Check output directory
    os.makedirs(config.output_dir, exist_ok=True)
    print(f"✓ Output directory: {config.output_dir}")
    
    print("✅ Training requirements check completed")
    return True


def create_training_arguments(config: DPOConfig) -> tuple:
    """Create training arguments from config"""
    
    # Model arguments
    model_args = ModelArguments(
        model_name_or_path=config.model_name_or_path,
        stage2_path=config.stage2_path,
        stage3_path=config.stage3_path,
        stage4_path=config.stage4_path,
        pretrain_mm_mlp_adapter=config.pretrain_mm_mlp_adapter,
        version="v1"
    )
    
    # Data arguments
    data_args = DataArguments(
        data_path=config.data_path,
        data_folder=config.data_folder,
        feat_folder=config.feat_folder
    )
    
    # Training arguments
    training_args = TrainingArguments(
        output_dir=config.output_dir,
        num_train_epochs=config.num_train_epochs,
        per_device_train_batch_size=config.per_device_train_batch_size,
        gradient_accumulation_steps=config.gradient_accumulation_steps,
        learning_rate=config.learning_rate,
        model_max_length=config.model_max_length,
        bf16=config.bf16,
        gradient_checkpointing=config.gradient_checkpointing,
        dataloader_num_workers=config.dataloader_num_workers,
        logging_steps=config.logging_steps,
        save_steps=config.save_steps,
        save_total_limit=config.save_total_limit,
        warmup_ratio=config.warmup_ratio,
        weight_decay=config.weight_decay,
        freeze_mm_mlp_adapter=config.freeze_mm_mlp_adapter,
        
        # DPO specific
        beta=config.beta,
        dpo_alpha=config.dpo_alpha,
        gamma=config.gamma,
        train4dpo=True,
        
        # LoRA specific
        lora_enable=config.lora_enable,
        lora_r=config.lora_r,
        lora_alpha=config.lora_alpha,
        lora_dropout=config.lora_dropout,
        
        # Single GPU training settings
        no_cuda=False,  # Keep CUDA enabled
        local_rank=-1,  # Force single GPU by setting local_rank to -1
        dataloader_pin_memory=False,  # Disable pin memory for single GPU
        
        # Other settings
        training_stage=3,
        finetuning=True,
        evaluation_strategy="no",
        save_strategy="steps",
        lr_scheduler_type="cosine",
        tf32=True,
        report_to="none",  # Disable wandb for demo
        remove_unused_columns=False
    )
    
    return model_args, data_args, training_args


def load_and_prepare_model(model_args, training_args):
    """Load and prepare the model for DPO training"""
    print("🤖 Loading and preparing model...")
    
    # Load base model
    model = VTimeLLMLlamaForCausalLM.from_pretrained(
        model_args.model_name_or_path,
        cache_dir=training_args.cache_dir,
        torch_dtype=torch.bfloat16 if training_args.bf16 else torch.float16
    )
    model.config.use_cache = False
    
    # Load tokenizer
    tokenizer = transformers.AutoTokenizer.from_pretrained(
        model_args.model_name_or_path,
        cache_dir=training_args.cache_dir,
        model_max_length=training_args.model_max_length,
        padding_side="right",
        use_fast=False,
    )
    tokenizer.pad_token = tokenizer.unk_token
    
    # Initialize vision modules
    model.get_model().initialize_vision_modules(model_args)
    model.cuda()
    
    # Load stage 2 weights
    print("📥 Loading stage 2 weights...")
    model = load_lora(model, model_args.stage2_path)
    model = model.merge_and_unload()
    
    # Load stage 3 weights
    print("📥 Loading stage 3 weights...")
    model = load_lora(model, model_args.stage3_path)
    model = model.merge_and_unload()
    
    # Load stage 4 weights (for DPO training)
    print("📥 Loading stage 4 weights...")
    model = load_lora(model, model_args.stage4_path)
    model = model.merge_and_unload()
    
    # Add LoRA adapters for DPO training
    if training_args.lora_enable:
        from peft import LoraConfig, get_peft_model
        
        # Find target modules for LoRA
        def find_all_linear_explicit_names(model):
            cls = torch.nn.Linear
            lora_module_names = set()
            multimodal_keywords = ['mm_projector', 'vision_tower', 'vision_resampler']
            for name, module in model.named_modules():
                if any(mm_keyword in name for mm_keyword in multimodal_keywords):
                    continue
                if isinstance(module, cls):
                    lora_module_names.add(name)
            if 'lm_head' in lora_module_names:
                lora_module_names.remove('lm_head')
            return list(lora_module_names)
        
        explicit_linear = find_all_linear_explicit_names(model)
        lora_config = LoraConfig(
            r=training_args.lora_r,
            lora_alpha=training_args.lora_alpha,
            target_modules=explicit_linear,
            lora_dropout=training_args.lora_dropout,
            bias="none",
            task_type="CAUSAL_LM",
        )
        
        print("🔧 Adding LoRA adapters...")
        model = get_peft_model(model, lora_config)
        print_trainable_parameters(model)
    
    # Move to GPU
    model = model.cuda()
    
    return model, tokenizer


def save_training_results(model, trainer, training_args, config):
    """Save training results and model"""
    print("💾 Saving training results...")
    
    # Save model
    if training_args.lora_enable:
        from vtimellm.train.train_dpo import get_peft_state_maybe_zero_3, get_peft_state_non_lora_maybe_zero_3
        
        state_dict = get_peft_state_maybe_zero_3(
            model.named_parameters(), training_args.lora_bias
        )
        non_lora_state_dict = get_peft_state_non_lora_maybe_zero_3(
            model.named_parameters()
        )
        
        if training_args.local_rank in [0, -1]:
            model.config.save_pretrained(training_args.output_dir)
            model.save_pretrained(training_args.output_dir, state_dict=state_dict)
            torch.save(non_lora_state_dict, os.path.join(training_args.output_dir, 'non_lora_trainables.bin'))
    
    # Save training config
    config_dict = {
        "model_paths": {
            "model_name_or_path": config.model_name_or_path,
            "stage2_path": config.stage2_path,
            "stage3_path": config.stage3_path,
            "stage4_path": config.stage4_path,
        },
        "training_params": {
            "num_train_epochs": config.num_train_epochs,
            "per_device_train_batch_size": config.per_device_train_batch_size,
            "learning_rate": config.learning_rate,
            "beta": config.beta,
            "dpo_alpha": config.dpo_alpha,
        },
        "lora_params": {
            "lora_r": config.lora_r,
            "lora_alpha": config.lora_alpha,
            "lora_dropout": config.lora_dropout,
        }
    }
    
    with open(os.path.join(training_args.output_dir, "training_config.json"), "w") as f:
        json.dump(config_dict, f, indent=2)
    
    print(f"✅ Results saved to: {training_args.output_dir}")


def demo_dpo_training(config: DPOConfig):
    """Main function for DPO training demo"""
    print("🎬 VTimeLLM DPO Training Demo")
    print("=" * 50)
    
    # Force single GPU usage
    print("🔧 Setting up single GPU training...")
    if torch.cuda.is_available():
        # Set CUDA device to 0 (first GPU)
        torch.cuda.set_device(0)
        print(f"✓ Using GPU: {torch.cuda.get_device_name(0)}")
        
        # Set environment variables for single GPU
        os.environ['CUDA_VISIBLE_DEVICES'] = '0'
        os.environ['OMPI_COMM_WORLD_SIZE'] = '1'
        print("✓ Environment variables set for single GPU")
    else:
        print("⚠ No CUDA GPU available")
    
    # Check requirements
    if not check_training_requirements(config):
        print("❌ Training requirements not met. Please fix the issues above.")
        return False
    
    try:
        # Create training arguments
        model_args, data_args, training_args = create_training_arguments(config)
        
        # Load and prepare model
        model, tokenizer = load_and_prepare_model(model_args, training_args)
        
        # Create trainer
        trainer = create_dpo_trainer(model, tokenizer, data_args, training_args)
        
        # Train model
        train_model(trainer, training_args)
        
        # Save results
        save_training_results(model, trainer, training_args, config)
        
        print("\n🎉 DPO Training Demo Completed Successfully!")
        print(f"📁 Model saved to: {config.output_dir}")
        print("\nYou can now use the trained model for inference!")
        
        return True
        
    except Exception as e:
        print(f"❌ Training failed with error: {e}")
        import traceback
        traceback.print_exc()
        return False


def create_sample_config():
    """Create a sample configuration for students"""
    return DPOConfig(
        # Use smaller batch size for demo
        per_device_train_batch_size=1,
        gradient_accumulation_steps=2,
        num_train_epochs=1,
        logging_steps=5,
        save_steps=100,
        save_total_limit=2
    )


def main():
    """Main function with command line interface"""
    # Use default config
    config = DPOConfig()
    print("⚙️ Using default configuration")
    check_only = False
    if check_only:
        # Only check requirements
        check_training_requirements(config)
    else:
        # Run full training demo
        demo_dpo_training(config)


if __name__ == "__main__":
    main()