htlou commited on
Commit
ebbb000
·
verified ·
1 Parent(s): dfbb9d8

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. arguments.yaml +11 -9
  3. environ.txt +21 -59
  4. script.sh +23 -26
  5. slice_14649/added_tokens.json +4 -0
  6. slice_14649/chat_template.json +3 -0
  7. slice_14649/config.json +85 -0
  8. slice_14649/preprocessor_config.json +52 -0
  9. slice_14649/processor_config.json +7 -0
  10. slice_14649/pytorch_model.bin +3 -0
  11. slice_14649/special_tokens_map.json +31 -0
  12. slice_14649/tokenizer.json +0 -0
  13. slice_14649/tokenizer.model +3 -0
  14. slice_14649/tokenizer_config.json +69 -0
  15. slice_19532/added_tokens.json +4 -0
  16. slice_19532/chat_template.json +3 -0
  17. slice_19532/config.json +85 -0
  18. slice_19532/preprocessor_config.json +52 -0
  19. slice_19532/processor_config.json +7 -0
  20. slice_19532/pytorch_model.bin +3 -0
  21. slice_19532/special_tokens_map.json +31 -0
  22. slice_19532/tokenizer.json +0 -0
  23. slice_19532/tokenizer.model +3 -0
  24. slice_19532/tokenizer_config.json +69 -0
  25. slice_24415/added_tokens.json +4 -0
  26. slice_24415/chat_template.json +3 -0
  27. slice_24415/config.json +85 -0
  28. slice_24415/preprocessor_config.json +52 -0
  29. slice_24415/processor_config.json +7 -0
  30. slice_24415/pytorch_model.bin +3 -0
  31. slice_24415/special_tokens_map.json +31 -0
  32. slice_24415/tokenizer.json +0 -0
  33. slice_24415/tokenizer.model +3 -0
  34. slice_24415/tokenizer_config.json +69 -0
  35. slice_29298/added_tokens.json +4 -0
  36. slice_29298/chat_template.json +3 -0
  37. slice_29298/config.json +85 -0
  38. slice_29298/preprocessor_config.json +52 -0
  39. slice_29298/processor_config.json +7 -0
  40. slice_29298/pytorch_model.bin +3 -0
  41. slice_29298/special_tokens_map.json +31 -0
  42. slice_29298/tokenizer.json +0 -0
  43. slice_29298/tokenizer.model +3 -0
  44. slice_29298/tokenizer_config.json +69 -0
  45. slice_4883/added_tokens.json +4 -0
  46. slice_4883/chat_template.json +3 -0
  47. slice_4883/config.json +85 -0
  48. slice_4883/preprocessor_config.json +52 -0
  49. slice_4883/processor_config.json +7 -0
  50. slice_4883/pytorch_model.bin +3 -0
.gitattributes CHANGED
@@ -34,3 +34,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  wandb/run-20250331_180548-xid0f47f/run-xid0f47f.wandb filter=lfs diff=lfs merge=lfs -text
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  wandb/run-20250331_180548-xid0f47f/run-xid0f47f.wandb filter=lfs diff=lfs merge=lfs -text
37
+ wandb/run-20250329_055156-k1abll4c/run-k1abll4c.wandb filter=lfs diff=lfs merge=lfs -text
38
+ wandb/run-20250330_120229-onfejhxh/run-onfejhxh.wandb filter=lfs diff=lfs merge=lfs -text
arguments.yaml CHANGED
@@ -1,25 +1,27 @@
1
  data_cfgs:
2
  eval_data_files: {}
3
  eval_datasets: {}
 
4
  eval_optional_args: []
5
  eval_size: {}
6
  eval_split: {}
7
  eval_subset: {}
8
  eval_template: {}
 
9
  train_data_files: {}
10
- train_datasets: /aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-40
11
  train_name: text-image-to-text
12
  train_optional_args: []
13
  train_size: {}
14
  train_split: train
15
- train_template: AA_TI2T_LLAVA
16
  logger_cfgs:
17
  cache_dir: {}
18
  log_project: align-anything
19
- log_run_name: dpo
20
  log_type: wandb
21
- output_dir: ../outputs/llava_1.6_mistral_7b_val/top1-40
22
- save_total_limit: 3
23
  model_cfgs:
24
  model_max_length: 4096
25
  model_name_or_path: /aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf
@@ -29,6 +31,7 @@ train_cfgs:
29
  adam_betas:
30
  - 0.9
31
  - 0.95
 
32
  bf16: true
33
  ds_cfgs: ds_z3_config.json
34
  epochs: 3
@@ -38,16 +41,15 @@ train_cfgs:
38
  freeze_language_model: false
39
  freeze_mm_proj: false
40
  freeze_vision_tower: true
41
- gradient_accumulation_steps: 1
42
  gradient_checkpointing: true
43
- learning_rate: 1.0e-06
44
  load_checkpoint: false
45
  lr_scheduler_type: cosine
46
  lr_warmup_ratio: 0.03
 
47
  per_device_eval_batch_size: 1
48
  per_device_train_batch_size: 1
49
- regularization: 0.001
50
  save_checkpoint: false
51
- scale_coeff: 0.1
52
  seed: 42
53
  weight_decay: 0.0
 
1
  data_cfgs:
2
  eval_data_files: {}
3
  eval_datasets: {}
4
+ eval_name: {}
5
  eval_optional_args: []
6
  eval_size: {}
7
  eval_split: {}
8
  eval_subset: {}
9
  eval_template: {}
10
+ load_multi_datasets: false
11
  train_data_files: {}
12
+ train_datasets: /aifs4su/yaodong/hantao/datasets/MMInstruct-GPT4V_mistral-7b_cosi_cut/merged/top1-40
13
  train_name: text-image-to-text
14
  train_optional_args: []
15
  train_size: {}
16
  train_split: train
17
+ train_template: MM_TI2T_LLAVA
18
  logger_cfgs:
19
  cache_dir: {}
20
  log_project: align-anything
21
+ log_run_name: sft
22
  log_type: wandb
23
+ output_dir: ../outputs/LLAVA_7B_cosi/top1-40
24
+ save_total_limit: 6
25
  model_cfgs:
26
  model_max_length: 4096
27
  model_name_or_path: /aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf
 
31
  adam_betas:
32
  - 0.9
33
  - 0.95
34
+ adam_epsilon: 1.0e-08
35
  bf16: true
36
  ds_cfgs: ds_z3_config.json
37
  epochs: 3
 
41
  freeze_language_model: false
42
  freeze_mm_proj: false
43
  freeze_vision_tower: true
44
+ gradient_accumulation_steps: 16
45
  gradient_checkpointing: true
46
+ learning_rate: 2.0e-05
47
  load_checkpoint: false
48
  lr_scheduler_type: cosine
49
  lr_warmup_ratio: 0.03
50
+ max_grad_norm: 1.0
51
  per_device_eval_batch_size: 1
52
  per_device_train_batch_size: 1
 
53
  save_checkpoint: false
 
54
  seed: 42
55
  weight_decay: 0.0
environ.txt CHANGED
@@ -84,75 +84,34 @@ BASH_FUNC_switchml%%=() { typeset swfound=1;
84
  BUILD=x86_64-conda-linux-gnu
85
  CC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
86
  CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
87
- CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
88
  CMAKE_ARGS=-DCMAKE_AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ar -DCMAKE_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib -DCMAKE_LINKER=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld -DCMAKE_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip -DCMAKE_BUILD_TYPE=Release
89
  CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama:/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot/usr
90
  CMD_WLM_CLUSTER_NAME=slurm
91
- CONDA_BACKUP_ADDR2LINE=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-addr2line
92
- CONDA_BACKUP_AR=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ar
93
- CONDA_BACKUP_AS=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-as
94
- CONDA_BACKUP_BUILD=x86_64-conda-linux-gnu
95
- CONDA_BACKUP_CC=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cc
96
- CONDA_BACKUP_CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cc
97
- CONDA_BACKUP_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
98
- CONDA_BACKUP_CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_anything:/aifs4su/yaodong/miniconda3/envs/hantao_anything/x86_64-conda-linux-gnu/sysroot/usr
99
- CONDA_BACKUP_CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/x86_64-conda-linux-gnu/sysroot
100
- CONDA_BACKUP_CPP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cpp
101
- CONDA_BACKUP_CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
102
- CONDA_BACKUP_CXX=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++
103
- CONDA_BACKUP_CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++filt
104
- CONDA_BACKUP_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
105
- CONDA_BACKUP_CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++
106
- CONDA_BACKUP_DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
107
- CONDA_BACKUP_DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
108
- CONDA_BACKUP_DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
109
- CONDA_BACKUP_ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-elfedit
110
- CONDA_BACKUP_GCC=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc
111
- CONDA_BACKUP_GCC_AR=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-ar
112
- CONDA_BACKUP_GCC_NM=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-nm
113
- CONDA_BACKUP_GCC_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-ranlib
114
- CONDA_BACKUP_GPROF=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gprof
115
- CONDA_BACKUP_GXX=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-g++
116
- CONDA_BACKUP_HOST=x86_64-conda-linux-gnu
117
- CONDA_BACKUP_LD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ld
118
- CONDA_BACKUP_LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib
119
- CONDA_BACKUP_LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ld.gold
120
- CONDA_BACKUP_NM=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-nm
121
- CONDA_BACKUP_OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-objcopy
122
- CONDA_BACKUP_OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-objdump
123
- CONDA_BACKUP_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ranlib
124
- CONDA_BACKUP_READELF=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-readelf
125
- CONDA_BACKUP_SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-size
126
- CONDA_BACKUP_STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-strings
127
- CONDA_BACKUP_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-strip
128
- CONDA_BACKUP__CONDA_PYTHON_SYSCONFIGDATA_NAME=_sysconfigdata_x86_64_conda_cos7_linux_gnu
129
- CONDA_BACKUP_build_alias=x86_64-conda-linux-gnu
130
- CONDA_BACKUP_host_alias=x86_64-conda-linux-gnu
131
  CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot
132
  CONDA_DEFAULT_ENV=hantao_llama
133
  CONDA_EXE=/aifs4su/yaodong/miniconda3/bin/conda
134
  CONDA_PREFIX=/aifs4su/yaodong/miniconda3/envs/hantao_llama
135
  CONDA_PREFIX_1=/aifs4su/yaodong/miniconda3
136
- CONDA_PREFIX_2=/aifs4su/yaodong/miniconda3/envs/hantao_anything
137
  CONDA_PROMPT_MODIFIER=(hantao_llama)
138
  CONDA_PYTHON_EXE=/aifs4su/yaodong/miniconda3/bin/python
139
- CONDA_SHLVL=3
140
  CPATH=/cm/shared/apps/slurm/current/include
141
  CPATH_modshare=/cm/shared/apps/slurm/current/include:1
142
  CPP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cpp
143
- CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
144
  CROSS_RANK=0
145
  CROSS_SIZE=1
146
  CUDA_MODULE_LOADING=LAZY
147
  CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
148
  CXX=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
149
  CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++filt
150
- CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
151
  CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
152
  DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1028/bus
153
- DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
154
- DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
155
- DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
156
  ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-elfedit
157
  ENABLE_LMOD=0
158
  GCC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc
@@ -172,7 +131,7 @@ KMP_DUPLICATE_LIB_OK=True
172
  KMP_INIT_AT_FORK=FALSE
173
  LANG=C.UTF-8
174
  LD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld
175
- LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
176
  LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld.gold
177
  LD_LIBRARY_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/../../lib64:/usr/mpi/gcc/openmpi-4.1.7a1/lib:/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64
178
  LD_LIBRARY_PATH_modshare=/cm/shared/apps/slurm/current/lib64:1:/usr/mpi/gcc/openmpi-4.1.7a1/lib:1:/cm/shared/apps/slurm/current/lib64/slurm:1
@@ -192,11 +151,12 @@ LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd
192
  MANPATH=/usr/mpi/gcc/openmpi-4.1.7a1/share/man:/cm/shared/apps/slurm/current/man:/cm/local/apps/environment-modules/4.5.3/share/man:/usr/local/man:/usr/local/share/man:/usr/share/man:/cm/local/apps/environment-modules/current/share/man:/cm/local/apps/environment-modules/current/share/man
193
  MANPATH_modshare=/usr/local/share/man:1:/usr/mpi/gcc/openmpi-4.1.7a1/share/man:1:/cm/local/apps/environment-modules/current/share/man:1:/cm/local/apps/environment-modules/4.5.3/share/man:1:/usr/local/man:1:/usr/share/man:1:/cm/shared/apps/slurm/current/man:1
194
  MASTER_ADDR=127.0.0.1
195
- MASTER_PORT=51401
196
  MIG_PARTED_CHECKPOINT_FILE=/var/lib/nvidia-mig-manager/checkpoint.json
197
  MIG_PARTED_CONFIG_FILE=/etc/nvidia-mig-manager/config.yaml
198
  MIG_PARTED_HOOKS_FILE=/etc/nvidia-mig-manager/hooks.yaml
199
  MODULEPATH=/cm/local/modulefiles:/cm/shared/modulefiles
 
200
  MODULESHOME=/cm/local/apps/environment-modules/4.5.3
201
  MODULES_CMD=/cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl
202
  MODULES_SET_SHELL_STARTUP=0
@@ -209,12 +169,14 @@ NVCC_PREPEND_FLAGS_BACKUP= -ccbin=/aifs4su/yaodong/miniconda3/bin/x86_64-conda-l
209
  NVITOP_MONITOR_MODE=colorful
210
  OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objcopy
211
  OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objdump
212
- OLDPWD=/aifs4su/yaodong/hantao/align-anything/scripts/llava
213
  PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin:/aifs4su/yaodong/miniconda3/condabin:/usr/mpi/gcc/openmpi-4.1.7a1/bin:/usr/lpp/mmfs/bin:/usr/local/cuda/bin:/opt/bin:/usr/lpp/mmfs/bin:/cm/shared/apps/slurm/current/sbin:/cm/shared/apps/slurm/current/bin:/usr/local/cuda/bin:/opt/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/sbin:/usr/sbin:/cm/local/apps/environment-modules/4.5.3/bin
214
  PATH_modshare=/usr/mpi/gcc/openmpi-4.1.7a1/bin:1:/opt/bin/:1:/usr/bin:1:/usr/local/bin:1:/cm/shared/apps/slurm/current/bin:1:/cm/shared/apps/slurm/current/sbin:1:/bin:1:/snap/bin:1:/sbin:1:/usr/sbin:1:/cm/local/apps/environment-modules/4.5.3/bin:1:/usr/games:1:/usr/local/sbin:1:/usr/lpp/mmfs/bin:1:/usr/local/cuda/bin:1:/usr/local/games:1
215
  PWD=/aifs4su/yaodong/hantao/align-anything/scripts
216
  PYTHONHASHSEED=42
217
- PYTHONPATH=/aifs4su/yaodong/hantao/align-anything/scripts:/aifs4su/yaodong/hantao/align-anything/scripts:/aifs4su/yaodong/hantao/align-anything/scripts:/aifs4su/yaodong/hantao/align-anything/scripts
 
 
218
  RANK=0
219
  RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib
220
  READELF=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-readelf
@@ -222,24 +184,24 @@ SHELL=/bin/bash
222
  SHLVL=3
223
  SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-size
224
  SLURM_CONF=/cm/shared/apps/slurm/var/etc/slurm/slurm.conf
225
- SSH_CLIENT=10.33.4.229 44942 22
226
- SSH_CONNECTION=10.33.4.51 48576 10.33.4.230 22
227
- SSH_TTY=/dev/pts/2
228
  STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strings
229
  STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip
230
  TERM=screen
231
  TERM_PROGRAM=tmux
232
  TERM_PROGRAM_VERSION=3.2a
233
- TMUX=/tmp/tmux-1028/default,2764504,18
234
- TMUX_PANE=%18
235
  USER=yangyaodong
236
  WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
237
- WANDB_SERVICE=2-1862653-tcp-localhost-37721
238
  WORLD_SIZE=8
239
  XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
240
  XDG_RUNTIME_DIR=/run/user/1028
241
  XDG_SESSION_CLASS=user
242
- XDG_SESSION_ID=63754
243
  XDG_SESSION_TYPE=tty
244
  ZERO_STAGE=3
245
  _=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/deepspeed
 
84
  BUILD=x86_64-conda-linux-gnu
85
  CC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
86
  CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
87
+ CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
88
  CMAKE_ARGS=-DCMAKE_AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ar -DCMAKE_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib -DCMAKE_LINKER=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld -DCMAKE_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip -DCMAKE_BUILD_TYPE=Release
89
  CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama:/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot/usr
90
  CMD_WLM_CLUSTER_NAME=slurm
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
  CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot
92
  CONDA_DEFAULT_ENV=hantao_llama
93
  CONDA_EXE=/aifs4su/yaodong/miniconda3/bin/conda
94
  CONDA_PREFIX=/aifs4su/yaodong/miniconda3/envs/hantao_llama
95
  CONDA_PREFIX_1=/aifs4su/yaodong/miniconda3
 
96
  CONDA_PROMPT_MODIFIER=(hantao_llama)
97
  CONDA_PYTHON_EXE=/aifs4su/yaodong/miniconda3/bin/python
98
+ CONDA_SHLVL=2
99
  CPATH=/cm/shared/apps/slurm/current/include
100
  CPATH_modshare=/cm/shared/apps/slurm/current/include:1
101
  CPP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cpp
102
+ CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
103
  CROSS_RANK=0
104
  CROSS_SIZE=1
105
  CUDA_MODULE_LOADING=LAZY
106
  CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
107
  CXX=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
108
  CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++filt
109
+ CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
110
  CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
111
  DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1028/bus
112
+ DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include
113
+ DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include
114
+ DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include
115
  ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-elfedit
116
  ENABLE_LMOD=0
117
  GCC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc
 
131
  KMP_INIT_AT_FORK=FALSE
132
  LANG=C.UTF-8
133
  LD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld
134
+ LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
135
  LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld.gold
136
  LD_LIBRARY_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/../../lib64:/usr/mpi/gcc/openmpi-4.1.7a1/lib:/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64
137
  LD_LIBRARY_PATH_modshare=/cm/shared/apps/slurm/current/lib64:1:/usr/mpi/gcc/openmpi-4.1.7a1/lib:1:/cm/shared/apps/slurm/current/lib64/slurm:1
 
151
  MANPATH=/usr/mpi/gcc/openmpi-4.1.7a1/share/man:/cm/shared/apps/slurm/current/man:/cm/local/apps/environment-modules/4.5.3/share/man:/usr/local/man:/usr/local/share/man:/usr/share/man:/cm/local/apps/environment-modules/current/share/man:/cm/local/apps/environment-modules/current/share/man
152
  MANPATH_modshare=/usr/local/share/man:1:/usr/mpi/gcc/openmpi-4.1.7a1/share/man:1:/cm/local/apps/environment-modules/current/share/man:1:/cm/local/apps/environment-modules/4.5.3/share/man:1:/usr/local/man:1:/usr/share/man:1:/cm/shared/apps/slurm/current/man:1
153
  MASTER_ADDR=127.0.0.1
154
+ MASTER_PORT=46590
155
  MIG_PARTED_CHECKPOINT_FILE=/var/lib/nvidia-mig-manager/checkpoint.json
156
  MIG_PARTED_CONFIG_FILE=/etc/nvidia-mig-manager/config.yaml
157
  MIG_PARTED_HOOKS_FILE=/etc/nvidia-mig-manager/hooks.yaml
158
  MODULEPATH=/cm/local/modulefiles:/cm/shared/modulefiles
159
+ MODULEPATH_modshare=/cm/local/modulefiles:1:/cm/shared/modulefiles:3
160
  MODULESHOME=/cm/local/apps/environment-modules/4.5.3
161
  MODULES_CMD=/cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl
162
  MODULES_SET_SHELL_STARTUP=0
 
169
  NVITOP_MONITOR_MODE=colorful
170
  OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objcopy
171
  OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objdump
172
+ OLDPWD=/home/yangyaodong
173
  PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin:/aifs4su/yaodong/miniconda3/condabin:/usr/mpi/gcc/openmpi-4.1.7a1/bin:/usr/lpp/mmfs/bin:/usr/local/cuda/bin:/opt/bin:/usr/lpp/mmfs/bin:/cm/shared/apps/slurm/current/sbin:/cm/shared/apps/slurm/current/bin:/usr/local/cuda/bin:/opt/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/sbin:/usr/sbin:/cm/local/apps/environment-modules/4.5.3/bin
174
  PATH_modshare=/usr/mpi/gcc/openmpi-4.1.7a1/bin:1:/opt/bin/:1:/usr/bin:1:/usr/local/bin:1:/cm/shared/apps/slurm/current/bin:1:/cm/shared/apps/slurm/current/sbin:1:/bin:1:/snap/bin:1:/sbin:1:/usr/sbin:1:/cm/local/apps/environment-modules/4.5.3/bin:1:/usr/games:1:/usr/local/sbin:1:/usr/lpp/mmfs/bin:1:/usr/local/cuda/bin:1:/usr/local/games:1
175
  PWD=/aifs4su/yaodong/hantao/align-anything/scripts
176
  PYTHONHASHSEED=42
177
+ PYTHONPATH=/aifs4su/yaodong/hantao/align-anything/scripts
178
+ QT_QPA_FONTDIR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/qt/fonts
179
+ QT_QPA_PLATFORM_PLUGIN_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/qt/plugins
180
  RANK=0
181
  RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib
182
  READELF=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-readelf
 
184
  SHLVL=3
185
  SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-size
186
  SLURM_CONF=/cm/shared/apps/slurm/var/etc/slurm/slurm.conf
187
+ SSH_CLIENT=10.33.4.51 34180 22
188
+ SSH_CONNECTION=10.33.4.51 34180 10.33.4.231 22
189
+ SSH_TTY=/dev/pts/8
190
  STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strings
191
  STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip
192
  TERM=screen
193
  TERM_PROGRAM=tmux
194
  TERM_PROGRAM_VERSION=3.2a
195
+ TMUX=/tmp/tmux-1028/default,1588807,0
196
+ TMUX_PANE=%0
197
  USER=yangyaodong
198
  WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
199
+ WANDB_SERVICE=2-1593022-tcp-localhost-46867
200
  WORLD_SIZE=8
201
  XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
202
  XDG_RUNTIME_DIR=/run/user/1028
203
  XDG_SESSION_CLASS=user
204
+ XDG_SESSION_ID=1733
205
  XDG_SESSION_TYPE=tty
206
  ZERO_STAGE=3
207
  _=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/deepspeed
script.sh CHANGED
@@ -15,38 +15,35 @@
15
  # limitations under the License.
16
  # ==============================================================================
17
 
18
- DATASETS_NAME=("top1-10" "top1-20" "top1-30" "top1-40" "top1-50" "top1-60" "top1-70" "top1-80" "top1-90" "top1-100")
19
- # DATASETS_NAME=("top1-10")
20
- #
21
- MODEL_NAME_OR_PATH="/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf" # model path
22
 
23
- # HOSTFILE="/aifs4su/yaodong/hantao/align-anything/scripts/.hostfile"
24
 
25
- for DATASET_NAME in ${DATASETS_NAME[@]}; do
26
- TRAIN_DATASETS="/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/${DATASET_NAME}" # dataset path
27
- TRAIN_TEMPLATE="AA_TI2T_LLAVA" # dataset template
28
  TRAIN_NAME="text-image-to-text" # dataset name
29
  TRAIN_SPLIT="train" # split the dataset
30
 
31
- OUTPUT_DIR="../outputs/llava_1.6_mistral_7b_val/${DATASET_NAME}" # output dir
32
 
33
- # For wandb online logging
34
- export WANDB_API_KEY="7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33"
35
 
36
- # Source the setup script
37
- source ./setup.sh
38
 
39
- # Execute deepspeed command
40
- deepspeed \
41
- --master_port ${MASTER_PORT} \
42
- --module align_anything.trainers.text_image_to_text.dpo \
43
- --model_name_or_path ${MODEL_NAME_OR_PATH} \
44
- --train_datasets ${TRAIN_DATASETS} \
45
- --train_template ${TRAIN_TEMPLATE} \
46
- --train_split ${TRAIN_SPLIT} \
47
- --train_name ${TRAIN_NAME} \
48
- --output_dir ${OUTPUT_DIR} \
49
- --save_total_limit 3 \
50
- --train_batch_size 8 \
51
- --epochs 3
52
  done
 
15
  # limitations under the License.
16
  # ==============================================================================
17
 
18
+ DATASETS_NAME=("top1-40")
 
 
 
19
 
20
+ MODEL_NAME_OR_PATH="/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf" # model path
21
 
22
+ for DATASET_NAME in "${DATASETS_NAME[@]}"; do
23
+ TRAIN_DATASETS="/aifs4su/yaodong/hantao/datasets/MMInstruct-GPT4V_mistral-7b_cosi_cut/merged/${DATASET_NAME}" # dataset path
24
+ TRAIN_TEMPLATE="MM_TI2T_LLAVA" # dataset template
25
  TRAIN_NAME="text-image-to-text" # dataset name
26
  TRAIN_SPLIT="train" # split the dataset
27
 
28
+ OUTPUT_DIR="../outputs/LLAVA_7B_cosi/${DATASET_NAME}" # output dir
29
 
30
+ # For wandb online logging
31
+ export WANDB_API_KEY="7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33"
32
 
33
+ # Source the setup script
34
+ source ./setup.sh
35
 
36
+ # Execute deepspeed command
37
+ deepspeed \
38
+ --master_port ${MASTER_PORT} \
39
+ --module align_anything.trainers.text_image_to_text.sft \
40
+ --model_name_or_path ${MODEL_NAME_OR_PATH} \
41
+ --train_datasets ${TRAIN_DATASETS} \
42
+ --train_template ${TRAIN_TEMPLATE} \
43
+ --train_split ${TRAIN_SPLIT} \
44
+ --train_name ${TRAIN_NAME} \
45
+ --output_dir ${OUTPUT_DIR} \
46
+ --save_total_limit 6 \
47
+ --train_batch_size 16 \
48
+ --epochs 3
49
  done
slice_14649/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
slice_14649/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
3
+ }
slice_14649/config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "ignore_index": -100,
9
+ "image_grid_pinpoints": [
10
+ [
11
+ 336,
12
+ 672
13
+ ],
14
+ [
15
+ 672,
16
+ 336
17
+ ],
18
+ [
19
+ 672,
20
+ 672
21
+ ],
22
+ [
23
+ 1008,
24
+ 336
25
+ ],
26
+ [
27
+ 336,
28
+ 1008
29
+ ]
30
+ ],
31
+ "image_seq_length": 576,
32
+ "image_token_index": 32000,
33
+ "model_type": "llava_next",
34
+ "multimodal_projector_bias": true,
35
+ "pad_token_id": 32001,
36
+ "projector_hidden_act": "gelu",
37
+ "text_config": {
38
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
39
+ "architectures": [
40
+ "MistralForCausalLM"
41
+ ],
42
+ "attention_dropout": 0.0,
43
+ "head_dim": 128,
44
+ "hidden_act": "silu",
45
+ "hidden_size": 4096,
46
+ "initializer_range": 0.02,
47
+ "intermediate_size": 14336,
48
+ "max_position_embeddings": 32768,
49
+ "model_type": "mistral",
50
+ "num_attention_heads": 32,
51
+ "num_hidden_layers": 32,
52
+ "num_key_value_heads": 8,
53
+ "rms_norm_eps": 1e-05,
54
+ "rope_theta": 1000000.0,
55
+ "sliding_window": null,
56
+ "torch_dtype": "bfloat16",
57
+ "use_cache": true,
58
+ "vocab_size": 32064
59
+ },
60
+ "tie_word_embeddings": false,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.50.0",
63
+ "use_image_newline_parameter": true,
64
+ "vision_config": {
65
+ "attention_dropout": 0.0,
66
+ "hidden_act": "quick_gelu",
67
+ "hidden_size": 1024,
68
+ "image_size": 336,
69
+ "initializer_factor": 1.0,
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 4096,
72
+ "layer_norm_eps": 1e-05,
73
+ "model_type": "clip_vision_model",
74
+ "num_attention_heads": 16,
75
+ "num_channels": 3,
76
+ "num_hidden_layers": 24,
77
+ "patch_size": 14,
78
+ "projection_dim": 768,
79
+ "torch_dtype": "bfloat16",
80
+ "vocab_size": 32000
81
+ },
82
+ "vision_feature_layer": -2,
83
+ "vision_feature_select_strategy": "default",
84
+ "vocab_size": 32064
85
+ }
slice_14649/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
slice_14649/processor_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_token": "<image>",
3
+ "num_additional_image_tokens": 1,
4
+ "patch_size": 14,
5
+ "processor_class": "LlavaNextProcessor",
6
+ "vision_feature_select_strategy": "default"
7
+ }
slice_14649/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f1fd230f3f01a6098c034d16a364fa803331cdd2e99039e7000a0f3a9a002b4
3
+ size 15133733934
slice_14649/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "image_token": "<image>",
17
+ "pad_token": {
18
+ "content": "<pad>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "unk_token": {
25
+ "content": "<unk>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
slice_14649/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
slice_14649/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
slice_14649/tokenizer_config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 4096,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "tokenizer_class": "LlamaTokenizer",
67
+ "unk_token": "<unk>",
68
+ "use_default_system_prompt": false
69
+ }
slice_19532/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
slice_19532/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
3
+ }
slice_19532/config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "ignore_index": -100,
9
+ "image_grid_pinpoints": [
10
+ [
11
+ 336,
12
+ 672
13
+ ],
14
+ [
15
+ 672,
16
+ 336
17
+ ],
18
+ [
19
+ 672,
20
+ 672
21
+ ],
22
+ [
23
+ 1008,
24
+ 336
25
+ ],
26
+ [
27
+ 336,
28
+ 1008
29
+ ]
30
+ ],
31
+ "image_seq_length": 576,
32
+ "image_token_index": 32000,
33
+ "model_type": "llava_next",
34
+ "multimodal_projector_bias": true,
35
+ "pad_token_id": 32001,
36
+ "projector_hidden_act": "gelu",
37
+ "text_config": {
38
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
39
+ "architectures": [
40
+ "MistralForCausalLM"
41
+ ],
42
+ "attention_dropout": 0.0,
43
+ "head_dim": 128,
44
+ "hidden_act": "silu",
45
+ "hidden_size": 4096,
46
+ "initializer_range": 0.02,
47
+ "intermediate_size": 14336,
48
+ "max_position_embeddings": 32768,
49
+ "model_type": "mistral",
50
+ "num_attention_heads": 32,
51
+ "num_hidden_layers": 32,
52
+ "num_key_value_heads": 8,
53
+ "rms_norm_eps": 1e-05,
54
+ "rope_theta": 1000000.0,
55
+ "sliding_window": null,
56
+ "torch_dtype": "bfloat16",
57
+ "use_cache": true,
58
+ "vocab_size": 32064
59
+ },
60
+ "tie_word_embeddings": false,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.50.0",
63
+ "use_image_newline_parameter": true,
64
+ "vision_config": {
65
+ "attention_dropout": 0.0,
66
+ "hidden_act": "quick_gelu",
67
+ "hidden_size": 1024,
68
+ "image_size": 336,
69
+ "initializer_factor": 1.0,
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 4096,
72
+ "layer_norm_eps": 1e-05,
73
+ "model_type": "clip_vision_model",
74
+ "num_attention_heads": 16,
75
+ "num_channels": 3,
76
+ "num_hidden_layers": 24,
77
+ "patch_size": 14,
78
+ "projection_dim": 768,
79
+ "torch_dtype": "bfloat16",
80
+ "vocab_size": 32000
81
+ },
82
+ "vision_feature_layer": -2,
83
+ "vision_feature_select_strategy": "default",
84
+ "vocab_size": 32064
85
+ }
slice_19532/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
slice_19532/processor_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_token": "<image>",
3
+ "num_additional_image_tokens": 1,
4
+ "patch_size": 14,
5
+ "processor_class": "LlavaNextProcessor",
6
+ "vision_feature_select_strategy": "default"
7
+ }
slice_19532/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ab661a9d25841f2346911001f2fff525ab0473e4430f7d52b42addaf867699ca
3
+ size 15133733934
slice_19532/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "image_token": "<image>",
17
+ "pad_token": {
18
+ "content": "<pad>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "unk_token": {
25
+ "content": "<unk>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
slice_19532/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
slice_19532/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
slice_19532/tokenizer_config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 4096,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "tokenizer_class": "LlamaTokenizer",
67
+ "unk_token": "<unk>",
68
+ "use_default_system_prompt": false
69
+ }
slice_24415/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
slice_24415/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
3
+ }
slice_24415/config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "ignore_index": -100,
9
+ "image_grid_pinpoints": [
10
+ [
11
+ 336,
12
+ 672
13
+ ],
14
+ [
15
+ 672,
16
+ 336
17
+ ],
18
+ [
19
+ 672,
20
+ 672
21
+ ],
22
+ [
23
+ 1008,
24
+ 336
25
+ ],
26
+ [
27
+ 336,
28
+ 1008
29
+ ]
30
+ ],
31
+ "image_seq_length": 576,
32
+ "image_token_index": 32000,
33
+ "model_type": "llava_next",
34
+ "multimodal_projector_bias": true,
35
+ "pad_token_id": 32001,
36
+ "projector_hidden_act": "gelu",
37
+ "text_config": {
38
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
39
+ "architectures": [
40
+ "MistralForCausalLM"
41
+ ],
42
+ "attention_dropout": 0.0,
43
+ "head_dim": 128,
44
+ "hidden_act": "silu",
45
+ "hidden_size": 4096,
46
+ "initializer_range": 0.02,
47
+ "intermediate_size": 14336,
48
+ "max_position_embeddings": 32768,
49
+ "model_type": "mistral",
50
+ "num_attention_heads": 32,
51
+ "num_hidden_layers": 32,
52
+ "num_key_value_heads": 8,
53
+ "rms_norm_eps": 1e-05,
54
+ "rope_theta": 1000000.0,
55
+ "sliding_window": null,
56
+ "torch_dtype": "bfloat16",
57
+ "use_cache": true,
58
+ "vocab_size": 32064
59
+ },
60
+ "tie_word_embeddings": false,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.50.0",
63
+ "use_image_newline_parameter": true,
64
+ "vision_config": {
65
+ "attention_dropout": 0.0,
66
+ "hidden_act": "quick_gelu",
67
+ "hidden_size": 1024,
68
+ "image_size": 336,
69
+ "initializer_factor": 1.0,
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 4096,
72
+ "layer_norm_eps": 1e-05,
73
+ "model_type": "clip_vision_model",
74
+ "num_attention_heads": 16,
75
+ "num_channels": 3,
76
+ "num_hidden_layers": 24,
77
+ "patch_size": 14,
78
+ "projection_dim": 768,
79
+ "torch_dtype": "bfloat16",
80
+ "vocab_size": 32000
81
+ },
82
+ "vision_feature_layer": -2,
83
+ "vision_feature_select_strategy": "default",
84
+ "vocab_size": 32064
85
+ }
slice_24415/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
slice_24415/processor_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_token": "<image>",
3
+ "num_additional_image_tokens": 1,
4
+ "patch_size": 14,
5
+ "processor_class": "LlavaNextProcessor",
6
+ "vision_feature_select_strategy": "default"
7
+ }
slice_24415/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1067de6e6a2dd60514331929e5e8fa8bdea95b178e2d56f3e10d53cb1565a34
3
+ size 15133733934
slice_24415/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "image_token": "<image>",
17
+ "pad_token": {
18
+ "content": "<pad>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "unk_token": {
25
+ "content": "<unk>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
slice_24415/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
slice_24415/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
slice_24415/tokenizer_config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 4096,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "tokenizer_class": "LlamaTokenizer",
67
+ "unk_token": "<unk>",
68
+ "use_default_system_prompt": false
69
+ }
slice_29298/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
slice_29298/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
3
+ }
slice_29298/config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "ignore_index": -100,
9
+ "image_grid_pinpoints": [
10
+ [
11
+ 336,
12
+ 672
13
+ ],
14
+ [
15
+ 672,
16
+ 336
17
+ ],
18
+ [
19
+ 672,
20
+ 672
21
+ ],
22
+ [
23
+ 1008,
24
+ 336
25
+ ],
26
+ [
27
+ 336,
28
+ 1008
29
+ ]
30
+ ],
31
+ "image_seq_length": 576,
32
+ "image_token_index": 32000,
33
+ "model_type": "llava_next",
34
+ "multimodal_projector_bias": true,
35
+ "pad_token_id": 32001,
36
+ "projector_hidden_act": "gelu",
37
+ "text_config": {
38
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
39
+ "architectures": [
40
+ "MistralForCausalLM"
41
+ ],
42
+ "attention_dropout": 0.0,
43
+ "head_dim": 128,
44
+ "hidden_act": "silu",
45
+ "hidden_size": 4096,
46
+ "initializer_range": 0.02,
47
+ "intermediate_size": 14336,
48
+ "max_position_embeddings": 32768,
49
+ "model_type": "mistral",
50
+ "num_attention_heads": 32,
51
+ "num_hidden_layers": 32,
52
+ "num_key_value_heads": 8,
53
+ "rms_norm_eps": 1e-05,
54
+ "rope_theta": 1000000.0,
55
+ "sliding_window": null,
56
+ "torch_dtype": "bfloat16",
57
+ "use_cache": true,
58
+ "vocab_size": 32064
59
+ },
60
+ "tie_word_embeddings": false,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.50.0",
63
+ "use_image_newline_parameter": true,
64
+ "vision_config": {
65
+ "attention_dropout": 0.0,
66
+ "hidden_act": "quick_gelu",
67
+ "hidden_size": 1024,
68
+ "image_size": 336,
69
+ "initializer_factor": 1.0,
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 4096,
72
+ "layer_norm_eps": 1e-05,
73
+ "model_type": "clip_vision_model",
74
+ "num_attention_heads": 16,
75
+ "num_channels": 3,
76
+ "num_hidden_layers": 24,
77
+ "patch_size": 14,
78
+ "projection_dim": 768,
79
+ "torch_dtype": "bfloat16",
80
+ "vocab_size": 32000
81
+ },
82
+ "vision_feature_layer": -2,
83
+ "vision_feature_select_strategy": "default",
84
+ "vocab_size": 32064
85
+ }
slice_29298/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
slice_29298/processor_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_token": "<image>",
3
+ "num_additional_image_tokens": 1,
4
+ "patch_size": 14,
5
+ "processor_class": "LlavaNextProcessor",
6
+ "vision_feature_select_strategy": "default"
7
+ }
slice_29298/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:10aecfb74b294ff71cad7db9a38b54042f26f379a4339cb36323dcb353204ef5
3
+ size 15133733934
slice_29298/special_tokens_map.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "image_token": "<image>",
17
+ "pad_token": {
18
+ "content": "<pad>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ },
24
+ "unk_token": {
25
+ "content": "<unk>",
26
+ "lstrip": false,
27
+ "normalized": false,
28
+ "rstrip": false,
29
+ "single_word": false
30
+ }
31
+ }
slice_29298/tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
slice_29298/tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
slice_29298/tokenizer_config.json ADDED
@@ -0,0 +1,69 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ },
30
+ "32000": {
31
+ "content": "<image>",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false,
36
+ "special": true
37
+ },
38
+ "32001": {
39
+ "content": "<pad>",
40
+ "lstrip": false,
41
+ "normalized": false,
42
+ "rstrip": false,
43
+ "single_word": false,
44
+ "special": true
45
+ }
46
+ },
47
+ "additional_special_tokens": [],
48
+ "bos_token": "<s>",
49
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
50
+ "clean_up_tokenization_spaces": false,
51
+ "eos_token": "</s>",
52
+ "extra_special_tokens": {
53
+ "image_token": "<image>"
54
+ },
55
+ "image_token": "<image>",
56
+ "legacy": true,
57
+ "max_length": null,
58
+ "model_max_length": 4096,
59
+ "pad_to_multiple_of": null,
60
+ "pad_token": "<pad>",
61
+ "pad_token_type_id": 0,
62
+ "padding_side": "right",
63
+ "processor_class": "LlavaNextProcessor",
64
+ "sp_model_kwargs": {},
65
+ "spaces_between_special_tokens": false,
66
+ "tokenizer_class": "LlamaTokenizer",
67
+ "unk_token": "<unk>",
68
+ "use_default_system_prompt": false
69
+ }
slice_4883/added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
slice_4883/chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
3
+ }
slice_4883/config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "architectures": [
4
+ "LlavaNextForConditionalGeneration"
5
+ ],
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "ignore_index": -100,
9
+ "image_grid_pinpoints": [
10
+ [
11
+ 336,
12
+ 672
13
+ ],
14
+ [
15
+ 672,
16
+ 336
17
+ ],
18
+ [
19
+ 672,
20
+ 672
21
+ ],
22
+ [
23
+ 1008,
24
+ 336
25
+ ],
26
+ [
27
+ 336,
28
+ 1008
29
+ ]
30
+ ],
31
+ "image_seq_length": 576,
32
+ "image_token_index": 32000,
33
+ "model_type": "llava_next",
34
+ "multimodal_projector_bias": true,
35
+ "pad_token_id": 32001,
36
+ "projector_hidden_act": "gelu",
37
+ "text_config": {
38
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
39
+ "architectures": [
40
+ "MistralForCausalLM"
41
+ ],
42
+ "attention_dropout": 0.0,
43
+ "head_dim": 128,
44
+ "hidden_act": "silu",
45
+ "hidden_size": 4096,
46
+ "initializer_range": 0.02,
47
+ "intermediate_size": 14336,
48
+ "max_position_embeddings": 32768,
49
+ "model_type": "mistral",
50
+ "num_attention_heads": 32,
51
+ "num_hidden_layers": 32,
52
+ "num_key_value_heads": 8,
53
+ "rms_norm_eps": 1e-05,
54
+ "rope_theta": 1000000.0,
55
+ "sliding_window": null,
56
+ "torch_dtype": "bfloat16",
57
+ "use_cache": true,
58
+ "vocab_size": 32064
59
+ },
60
+ "tie_word_embeddings": false,
61
+ "torch_dtype": "bfloat16",
62
+ "transformers_version": "4.50.0",
63
+ "use_image_newline_parameter": true,
64
+ "vision_config": {
65
+ "attention_dropout": 0.0,
66
+ "hidden_act": "quick_gelu",
67
+ "hidden_size": 1024,
68
+ "image_size": 336,
69
+ "initializer_factor": 1.0,
70
+ "initializer_range": 0.02,
71
+ "intermediate_size": 4096,
72
+ "layer_norm_eps": 1e-05,
73
+ "model_type": "clip_vision_model",
74
+ "num_attention_heads": 16,
75
+ "num_channels": 3,
76
+ "num_hidden_layers": 24,
77
+ "patch_size": 14,
78
+ "projection_dim": 768,
79
+ "torch_dtype": "bfloat16",
80
+ "vocab_size": 32000
81
+ },
82
+ "vision_feature_layer": -2,
83
+ "vision_feature_select_strategy": "default",
84
+ "vocab_size": 32064
85
+ }
slice_4883/preprocessor_config.json ADDED
@@ -0,0 +1,52 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aspect_ratio_setting": "anyres",
3
+ "crop_size": {
4
+ "height": 336,
5
+ "width": 336
6
+ },
7
+ "do_center_crop": true,
8
+ "do_convert_rgb": true,
9
+ "do_normalize": true,
10
+ "do_pad": true,
11
+ "do_rescale": true,
12
+ "do_resize": true,
13
+ "image_grid_pinpoints": [
14
+ [
15
+ 336,
16
+ 672
17
+ ],
18
+ [
19
+ 672,
20
+ 336
21
+ ],
22
+ [
23
+ 672,
24
+ 672
25
+ ],
26
+ [
27
+ 1008,
28
+ 336
29
+ ],
30
+ [
31
+ 336,
32
+ 1008
33
+ ]
34
+ ],
35
+ "image_mean": [
36
+ 0.48145466,
37
+ 0.4578275,
38
+ 0.40821073
39
+ ],
40
+ "image_processor_type": "LlavaNextImageProcessor",
41
+ "image_std": [
42
+ 0.26862954,
43
+ 0.26130258,
44
+ 0.27577711
45
+ ],
46
+ "processor_class": "LlavaNextProcessor",
47
+ "resample": 3,
48
+ "rescale_factor": 0.00392156862745098,
49
+ "size": {
50
+ "shortest_edge": 336
51
+ }
52
+ }
slice_4883/processor_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "image_token": "<image>",
3
+ "num_additional_image_tokens": 1,
4
+ "patch_size": 14,
5
+ "processor_class": "LlavaNextProcessor",
6
+ "vision_feature_select_strategy": "default"
7
+ }
slice_4883/pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:37c1986000ea6b0eb2357c5c441485335ede15f5532aa9809e5a838f1ba2e70f
3
+ size 15133733934