speedinghzl commited on
Commit
d71bf69
·
verified ·
1 Parent(s): fc494b4

Upload folder using huggingface_hub

Browse files
logs-lr1e-3/clipcls_vit_b16_s512m_bs16k/checkpoints/epoch_1.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a505c21768d54c953977e6c87de5bf2cf2228fd2c9aad83c46bea1028d1b6b99
3
+ size 2252180672
logs-lr1e-3/clipcls_vit_b16_s512m_bs16k/checkpoints/results.jsonl ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {"imagenet-zeroshot-val-top1": 0.3911, "imagenet-zeroshot-val-top5": 0.67312}
2
+ {"imagenet-zeroshot-val-top1": 0.48068, "imagenet-zeroshot-val-top5": 0.76568}
3
+ {"imagenet-zeroshot-val-top1": 0.54642, "imagenet-zeroshot-val-top5": 0.8211}
4
+ {"imagenet-zeroshot-val-top1": 0.6014, "imagenet-zeroshot-val-top5": 0.86306}
5
+ {"imagenet-zeroshot-val-top1": 0.63588, "imagenet-zeroshot-val-top5": 0.88492}
6
+ {"imagenet-zeroshot-val-top1": 0.63514, "imagenet-zeroshot-val-top5": 0.88496}
logs-lr1e-3/clipcls_vit_b16_s512m_bs16k/out.log ADDED
The diff for this file is too large to render. See raw diff
 
logs-lr1e-3/clipcls_vit_b16_s512m_bs16k/params.txt ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ accum_freq: 1
2
+ aug_cfg: {}
3
+ batch_size: 2048
4
+ beta1: 0.9
5
+ beta2: 0.98
6
+ checkpoint_path: ./logs-lr1e-3/clipcls_vit_b16_s512m_bs16k/checkpoints
7
+ coca_caption_loss_weight: 2.0
8
+ coca_contrastive_loss_weight: 1.0
9
+ copy_codebase: False
10
+ csv_caption_key: title
11
+ csv_img_key: filepath
12
+ csv_separator:
13
+ dataset_resampled: False
14
+ dataset_type: webdataset
15
+ ddp_static_graph: True
16
+ debug: False
17
+ delete_prev_step_ckpt: True
18
+ delete_previous_checkpoint: False
19
+ device: cuda:0
20
+ dist_backend: nccl
21
+ dist_url: env://
22
+ distill: False
23
+ distill_model: None
24
+ distill_pretrained: None
25
+ distributed: True
26
+ epochs: 1
27
+ epochs_cooldown: None
28
+ eps: 1e-06
29
+ force_custom_text: False
30
+ force_image_size: 224
31
+ force_patch_dropout: None
32
+ force_quick_gelu: False
33
+ gather_with_grad: True
34
+ global_batch_size: 16384
35
+ grad_checkpointing: True
36
+ grad_clip_norm: None
37
+ horovod: False
38
+ image_interpolation: None
39
+ image_mean: None
40
+ image_resize_mode: None
41
+ image_std: None
42
+ imagenet_v2: None
43
+ imagenet_val: /mnt/bn/zilongdata-hl/dataset/imagenet/val
44
+ local_loss: True
45
+ local_rank: 0
46
+ lock_image: False
47
+ lock_image_freeze_bn_stats: False
48
+ lock_image_unlocked_groups: 0
49
+ lock_text: False
50
+ lock_text_freeze_layer_norm: False
51
+ lock_text_unlocked_layers: 0
52
+ log_every_n_steps: 128
53
+ log_level: 20
54
+ log_local: False
55
+ log_path: ./logs-lr1e-3/clipcls_vit_b16_s512m_bs16k/out.log
56
+ logs: ./logs-lr1e-3
57
+ lr: 0.001
58
+ lr_cooldown_end: 0.0
59
+ lr_cooldown_power: 1.0
60
+ lr_scheduler: cosine
61
+ model: CLIPCLS-ViT-B-16
62
+ name: clipcls_vit_b16_s512m_bs16k
63
+ no_set_device_rank: False
64
+ precision: amp_bfloat16
65
+ pretrained:
66
+ pretrained_image:
67
+ pretrained_text:
68
+ rank: 0
69
+ remote_sync: None
70
+ remote_sync_frequency: 300
71
+ remote_sync_protocol: s3
72
+ report_to: wandb
73
+ resume: None
74
+ save_every_n_steps: 6104
75
+ save_frequency: 1
76
+ save_most_recent: False
77
+ seed: 0
78
+ siglip: False
79
+ skip_scheduler: False
80
+ tensorboard: False
81
+ tensorboard_path:
82
+ torchcompile: False
83
+ torchscript: False
84
+ trace: False
85
+ train_data: /mnt/bn/zilongdata-hl/dataset/datacomp-1b-webdataset/{000000..140146}.tar
86
+ train_data_upsampling_factors: None
87
+ train_num_samples: 512000000
88
+ use_bn_sync: False
89
+ use_bnb_linear: None
90
+ val_data: None
91
+ val_frequency: 1
92
+ val_num_samples: None
93
+ val_steps: 6104
94
+ wandb: True
95
+ wandb_notes:
96
+ wandb_project_name: cls-clip-batch-size
97
+ warmup: 500
98
+ wd: 0.2
99
+ workers: 1
100
+ world_size: 8
101
+ zeroshot_frequency: 2
102
+ zeroshot_steps: 6104