zheedong commited on
Commit
8bd6512
·
1 Parent(s): a18294a

CLIP Train update

Browse files
CLIP_last_layer/checkpoints/epoch=0-step=68000.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:657fc27c3f97ae064e7850389cc5f28876a1e584b771eebceed259bf12e1cf01
3
+ size 17576378479
CLIP_last_layer/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lm_model_name": "meta-llama/Meta-Llama-3-8B-Instruct", "lr": 0.0001, "weight_decay": 0.01, "warmup_steps": 100, "vision_encoder_name": "clip", "vision_hidden_size": 1024, "freeze_vision": true, "use_lora": true, "vision_layer": -1, "num_workers": 8, "val_num_workers": 1, "shuffle_buffer": 10000, "resampled": false, "max_length": 32, "training_epochs": 1, "batch_size": 16, "devices": 2, "lora_r": 8, "lora_alpha": 16, "lora_dropout": 0.05, "num_sanity_val_steps": 2, "val_check_interval": 1000, "limit_val_batches": 16, "num_training_step": 68750}
CLIP_last_layer/events.out.tfevents.1764619008.ubuntu ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db912c8932c42e3b40f212ebd97a1366b93a363c30f367424bd9f62648ace140
3
+ size 1678295
CLIP_last_layer/hparams.yaml ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ config:
2
+ lm_model_name: meta-llama/Meta-Llama-3-8B-Instruct
3
+ lr: 0.0001
4
+ weight_decay: 0.01
5
+ warmup_steps: 100
6
+ vision_hidden_size: 1024
7
+ freeze_vision: true
8
+ use_lora: true
9
+ vision_layer: -1
10
+ num_workers: 8
11
+ val_num_workers: 1
12
+ shuffle_buffer: 10000
13
+ resampled: false
14
+ max_length: 32
15
+ training_epochs: 1
16
+ batch_size: 16
17
+ devices: 2
18
+ lora_r: 8
19
+ lora_alpha: 16
20
+ lora_dropout: 0.05
21
+ num_sanity_val_steps: 2
22
+ val_check_interval: 1000
23
+ limit_val_batches: 16
24
+ num_training_step: 68750
version_0_vggt_last_layer/config.json DELETED
@@ -1 +0,0 @@
1
- {"lm_model_name": "meta-llama/Meta-Llama-3-8B-Instruct", "lr": 0.0001, "weight_decay": 0.01, "warmup_steps": 100, "vision_hidden_size": 2048, "freeze_vision": true, "use_lora": true, "vision_layer": -1, "num_workers": 8, "val_num_workers": 1, "shuffle_buffer": 10000, "resampled": false, "max_length": 32, "training_epochs": 1, "batch_size": 16, "devices": 2, "lora_r": 8, "lora_alpha": 16, "lora_dropout": 0.05, "num_sanity_val_steps": 2, "val_check_interval": 1000, "limit_val_batches": 16, "num_training_step": 68750}
 
 
{version_0_vggt_last_layer → vggt_last_layer}/checkpoints/epoch=0-step=68000.ckpt RENAMED
File without changes
vggt_last_layer/config.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"lm_model_name": "meta-llama/Meta-Llama-3-8B-Instruct", "lr": 0.0001, "weight_decay": 0.01, "warmup_steps": 100, "vision_encoder_name": "vggt", "vision_hidden_size": 2048, "freeze_vision": true, "use_lora": true, "vision_layer": -1, "num_workers": 8, "val_num_workers": 1, "shuffle_buffer": 10000, "resampled": false, "max_length": 32, "training_epochs": 1, "batch_size": 16, "devices": 2, "lora_r": 8, "lora_alpha": 16, "lora_dropout": 0.05, "num_sanity_val_steps": 2, "val_check_interval": 1000, "limit_val_batches": 16, "num_training_step": 68750}
{version_0_vggt_last_layer → vggt_last_layer}/events.out.tfevents.1763589865.ubuntu RENAMED
File without changes
{version_0_vggt_last_layer → vggt_last_layer}/hparams.yaml RENAMED
File without changes