| # Training | |
| This repository supports finetuning SAM3 models on custom datasets in multi-node setup or local execution. The training script is located at `sam3/train.py` and uses Hydra configuration management to handle complex training setups. | |
| ## Installation | |
| ```bash | |
| cd sam3 | |
| pip install -e ".[train]" | |
| ``` | |
| ### Training Script Usage | |
| The main training script is located at `sam3/train.py`. It uses Hydra configuration management to handle complex training setups. | |
| #### Basic Usage | |
| ```bash | |
| # Example: Train on Roboflow dataset | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml | |
| # Example: Train on ODinW13 dataset | |
| python sam3/train/train.py -c configs/odinw13/odinw_text_only_train.yaml | |
| ``` | |
| Follow [`Roboflow 100-VL`](https://github.com/roboflow/rf100-vl/) to download the roboflow 100-vl datasets. Follow [`GLIP`](https://github.com/microsoft/GLIP) to download the ODinW datasets. The data folder should be organized as follows, and put your roboflow_vl_100_root and odinw_data_root in the job configs. | |
| ``` | |
| roboflow_vl_100_root: | |
| 13-lkc01 | |
| train | |
| valid | |
| test | |
| 2024-frc | |
| actions | |
| ... | |
| odinw_data_root: | |
| AerialMaritimeDrone | |
| large | |
| train | |
| valid | |
| test | |
| Aquarium | |
| ... | |
| ``` | |
| #### Command Line Arguments | |
| The training script supports several command line arguments: | |
| ```bash | |
| python sam3/train/train.py \ | |
| -c CONFIG_NAME \ | |
| [--use-cluster 0|1] \ | |
| [--partition PARTITION_NAME] \ | |
| [--account ACCOUNT_NAME] \ | |
| [--qos QOS_NAME] \ | |
| [--num-gpus NUM_GPUS] \ | |
| [--num-nodes NUM_NODES] | |
| ``` | |
| **Arguments:** | |
| - `-c, --config`: **Required.** Path to the configuration file (e.g., `sam3/train/configs/roboflow_v100_full_ft_100_images.yaml`) | |
| - `--use-cluster`: Whether to launch on a cluster (0: local, 1: cluster). Default: uses config setting | |
| - `--partition`: SLURM partition name for cluster execution | |
| - `--account`: SLURM account name for cluster execution | |
| - `--qos`: SLURM QOS (Quality of Service) setting | |
| - `--num-gpus`: Number of GPUs per node. Default: uses config setting | |
| - `--num-nodes`: Number of nodes for distributed training. Default: uses config setting | |
| #### Local Training Examples | |
| ```bash | |
| # Single GPU training | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml --use-cluster 0 --num-gpus 1 | |
| # Multi-GPU training on a single node | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml --use-cluster 0 --num-gpus 4 | |
| # Force local execution even if config specifies GPUs | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml --use-cluster 0 | |
| ``` | |
| #### Cluster Training Examples | |
| ```bash | |
| # Basic cluster training with default settings from config | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml --use-cluster 1 | |
| # Cluster training with specific SLURM settings | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml \ | |
| --use-cluster 1 \ | |
| --partition gpu_partition \ | |
| --account my_account \ | |
| --qos high_priority \ | |
| --num-gpus 8 \ | |
| --num-nodes 2 | |
| ``` | |
| ### Configuration Files | |
| Training configurations are stored in `sam3/train/configs/`. The configuration files use Hydra's YAML format and support: | |
| - **Dataset Configuration**: Data paths, transforms, and loading parameters | |
| - **Model Configuration**: Architecture settings, checkpoint paths, and model parameters | |
| - **Training Configuration**: Batch sizes, learning rates, optimization settings | |
| - **Launcher Configuration**: Distributed training and cluster settings | |
| - **Logging Configuration**: TensorBoard, experiment tracking, and output directories | |
| #### Key Configuration Sections | |
| ```yaml | |
| # Paths to datasets and checkpoints | |
| paths: | |
| bpe_path: /path/to/bpe/file | |
| dataset_root: /path/to/dataset | |
| experiment_log_dir: /path/to/logs | |
| # Launcher settings for local/cluster execution | |
| launcher: | |
| num_nodes: 1 | |
| gpus_per_node: 2 | |
| experiment_log_dir: ${paths.experiment_log_dir} | |
| # Cluster execution settings | |
| submitit: | |
| use_cluster: True | |
| timeout_hour: 72 | |
| cpus_per_task: 10 | |
| partition: null | |
| account: null | |
| ``` | |
| ### Monitoring Training | |
| The training script automatically sets up logging and saves outputs to the experiment directory: | |
| ```bash | |
| # Logs are saved to the experiment_log_dir specified in config | |
| experiment_log_dir/ | |
| ├── config.yaml # Original configuration | |
| ├── config_resolved.yaml # Resolved configuration with all variables expanded | |
| ├── checkpoints/ # Model checkpoints (if skip_checkpointing=False) | |
| ├── tensorboard/ # TensorBoard logs | |
| ├── logs/ # Text logs | |
| └── submitit_logs/ # Cluster job logs (if using cluster) | |
| ``` | |
| You can monitor training progress using TensorBoard: | |
| ```bash | |
| tensorboard --logdir /path/to/experiment_log_dir/tensorboard | |
| ``` | |
| ### Job Arrays for Dataset Sweeps | |
| The Roboflow and ODinW configuration supports job arrays for training multiple models on different datasets: | |
| This feature is specifically enabled via, | |
| ```yaml | |
| submitit: | |
| job_array: | |
| num_tasks: 100 | |
| task_index: 0 | |
| ``` | |
| The configuration includes a complete list of 100 Roboflow supercategories, and the `submitit.job_array.task_index` automatically selects which dataset to use based on the array job index. | |
| ```bash | |
| # Submit job array to train on different Roboflow datasets | |
| # The job array index selects which dataset from all_roboflow_supercategories | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_full_ft_100_images.yaml \ | |
| --use-cluster 1 | |
| ``` | |
| ### Reproduce ODinW13 10-shot results | |
| Running the following job will give the results on the ODinW13 seed 300, see `odinw_train.train_file: fewshot_train_shot10_seed300` in the config file. | |
| ```bash | |
| # Example: Train on ODinW13 dataset | |
| python sam3/train/train.py -c configs/odinw13/odinw_text_only_train.yaml | |
| ``` | |
| Change `odinw_train.train_file` to `fewshot_train_shot10_seed30` and `fewshot_train_shot10_seed3` to get the results for the other two seeds. Final results are aggregated from the three seeds. Notice that a small number of jobs may diverge during training, in which case we just use the last checkpoint's result before it diverges. | |
| ### Eval Script Usage | |
| With a similar setup as the training config, the training script `sam3/train.py` can also be used for evaluation, too, when setting `trainer.mode = val` in the job config. Run the following job will give the results on the zero-shot results on RF100-VL and ODinW13 datasets. | |
| ```bash | |
| # Example: Evaluate on Roboflow dataset | |
| python sam3/train/train.py -c configs/roboflow_v100/roboflow_v100_eval.yaml | |
| # Example: Evaluate on ODinW13 dataset | |
| python sam3/train/train.py -c configs/odinw13/odinw_text_only.yaml | |
| ``` | |