File size: 2,149 Bytes
87b03df
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
{
    "dataset": {
      "align_stage_components": [
        "download/llava-laion-cc-sbu-558k/chat.json",
        "download/llava-laion-cc-sbu-558k"
      ],
      "dataset_id": "llava-lvis4v-lrv",
      "dataset_root_dir": "/opt/ml/input/data/training/skaramcheti/datasets/prismatic-vlms",
      "finetune_stage_components": [
        "download/llava-v1.5-instruct/llava_v1_5_lvis4v_lrv_mix1231k.json",
        "download/llava-v1.5-instruct"
      ],
      "type": "llava-lvis4v-lrv"
    },
    "hf_token": ".hf_token",
    "model": {
      "align_epochs": 1,
      "align_global_batch_size": 256,
      "align_learning_rate": 0.001,
      "align_lr_scheduler_type": "linear-warmup+cosine-decay",
      "align_max_grad_norm": 1.0,
      "align_max_steps": null,
      "align_per_device_batch_size": 16,
      "align_train_strategy": "fsdp-shard-grad-op",
      "align_warmup_ratio": 0.03,
      "align_weight_decay": 0.0,
      "arch_specifier": "no-align+fused-gelu-mlp",
      "enable_gradient_checkpointing": true,
      "enable_mixed_precision_training": true,
      "finetune_epochs": 2,
      "finetune_global_batch_size": 128,
      "finetune_learning_rate": 2e-05,
      "finetune_lr_scheduler_type": "linear-warmup+cosine-decay",
      "finetune_max_grad_norm": 1.0,
      "finetune_max_steps": null,
      "finetune_per_device_batch_size": 16,
      "finetune_train_strategy": "fsdp-full-shard",
      "finetune_warmup_ratio": 0.03,
      "finetune_weight_decay": 0.1,
      "image_resize_strategy": "resize-naive",
      "llm_backbone_id": "llama2-7b-pure",
      "llm_max_length": 2048,
      "model_id": "prism-dinosiglip-224px+7b",
      "reduce_in_full_precision": false,
      "type": "prism-dinosiglip-224px+7b",
      "vision_backbone_id": "dinosiglip-vit-so-224px"
    },
    "pretrained_checkpoint": null,
    "run_id": "llava-lvis4v-lrv+prism-dinosiglip-224px+7b+stage-finetune+x7",
    "run_root_dir": "/opt/ml/input/data/training/x-prismatic-vlms/runs",
    "seed": 7,
    "stage": "finetune",
    "trackers": [
      "jsonl",
      "wandb"
    ],
    "wandb_entity": "stanford-voltron",
    "wandb_project": "onyx-vlms"
  }