Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +2 -0
- arguments.yaml +11 -9
- environ.txt +21 -59
- script.sh +23 -26
- slice_14649/added_tokens.json +4 -0
- slice_14649/chat_template.json +3 -0
- slice_14649/config.json +85 -0
- slice_14649/preprocessor_config.json +52 -0
- slice_14649/processor_config.json +7 -0
- slice_14649/pytorch_model.bin +3 -0
- slice_14649/special_tokens_map.json +31 -0
- slice_14649/tokenizer.json +0 -0
- slice_14649/tokenizer.model +3 -0
- slice_14649/tokenizer_config.json +69 -0
- slice_19532/added_tokens.json +4 -0
- slice_19532/chat_template.json +3 -0
- slice_19532/config.json +85 -0
- slice_19532/preprocessor_config.json +52 -0
- slice_19532/processor_config.json +7 -0
- slice_19532/pytorch_model.bin +3 -0
- slice_19532/special_tokens_map.json +31 -0
- slice_19532/tokenizer.json +0 -0
- slice_19532/tokenizer.model +3 -0
- slice_19532/tokenizer_config.json +69 -0
- slice_24415/added_tokens.json +4 -0
- slice_24415/chat_template.json +3 -0
- slice_24415/config.json +85 -0
- slice_24415/preprocessor_config.json +52 -0
- slice_24415/processor_config.json +7 -0
- slice_24415/pytorch_model.bin +3 -0
- slice_24415/special_tokens_map.json +31 -0
- slice_24415/tokenizer.json +0 -0
- slice_24415/tokenizer.model +3 -0
- slice_24415/tokenizer_config.json +69 -0
- slice_29298/added_tokens.json +4 -0
- slice_29298/chat_template.json +3 -0
- slice_29298/config.json +85 -0
- slice_29298/preprocessor_config.json +52 -0
- slice_29298/processor_config.json +7 -0
- slice_29298/pytorch_model.bin +3 -0
- slice_29298/special_tokens_map.json +31 -0
- slice_29298/tokenizer.json +0 -0
- slice_29298/tokenizer.model +3 -0
- slice_29298/tokenizer_config.json +69 -0
- slice_4883/added_tokens.json +4 -0
- slice_4883/chat_template.json +3 -0
- slice_4883/config.json +85 -0
- slice_4883/preprocessor_config.json +52 -0
- slice_4883/processor_config.json +7 -0
- slice_4883/pytorch_model.bin +3 -0
.gitattributes
CHANGED
|
@@ -34,3 +34,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
wandb/run-20250331_180548-xid0f47f/run-xid0f47f.wandb filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
wandb/run-20250331_180548-xid0f47f/run-xid0f47f.wandb filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
wandb/run-20250329_055156-k1abll4c/run-k1abll4c.wandb filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
wandb/run-20250330_120229-onfejhxh/run-onfejhxh.wandb filter=lfs diff=lfs merge=lfs -text
|
arguments.yaml
CHANGED
|
@@ -1,25 +1,27 @@
|
|
| 1 |
data_cfgs:
|
| 2 |
eval_data_files: {}
|
| 3 |
eval_datasets: {}
|
|
|
|
| 4 |
eval_optional_args: []
|
| 5 |
eval_size: {}
|
| 6 |
eval_split: {}
|
| 7 |
eval_subset: {}
|
| 8 |
eval_template: {}
|
|
|
|
| 9 |
train_data_files: {}
|
| 10 |
-
train_datasets: /aifs4su/yaodong/hantao/datasets/
|
| 11 |
train_name: text-image-to-text
|
| 12 |
train_optional_args: []
|
| 13 |
train_size: {}
|
| 14 |
train_split: train
|
| 15 |
-
train_template:
|
| 16 |
logger_cfgs:
|
| 17 |
cache_dir: {}
|
| 18 |
log_project: align-anything
|
| 19 |
-
log_run_name:
|
| 20 |
log_type: wandb
|
| 21 |
-
output_dir: ../outputs/
|
| 22 |
-
save_total_limit:
|
| 23 |
model_cfgs:
|
| 24 |
model_max_length: 4096
|
| 25 |
model_name_or_path: /aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf
|
|
@@ -29,6 +31,7 @@ train_cfgs:
|
|
| 29 |
adam_betas:
|
| 30 |
- 0.9
|
| 31 |
- 0.95
|
|
|
|
| 32 |
bf16: true
|
| 33 |
ds_cfgs: ds_z3_config.json
|
| 34 |
epochs: 3
|
|
@@ -38,16 +41,15 @@ train_cfgs:
|
|
| 38 |
freeze_language_model: false
|
| 39 |
freeze_mm_proj: false
|
| 40 |
freeze_vision_tower: true
|
| 41 |
-
gradient_accumulation_steps:
|
| 42 |
gradient_checkpointing: true
|
| 43 |
-
learning_rate:
|
| 44 |
load_checkpoint: false
|
| 45 |
lr_scheduler_type: cosine
|
| 46 |
lr_warmup_ratio: 0.03
|
|
|
|
| 47 |
per_device_eval_batch_size: 1
|
| 48 |
per_device_train_batch_size: 1
|
| 49 |
-
regularization: 0.001
|
| 50 |
save_checkpoint: false
|
| 51 |
-
scale_coeff: 0.1
|
| 52 |
seed: 42
|
| 53 |
weight_decay: 0.0
|
|
|
|
| 1 |
data_cfgs:
|
| 2 |
eval_data_files: {}
|
| 3 |
eval_datasets: {}
|
| 4 |
+
eval_name: {}
|
| 5 |
eval_optional_args: []
|
| 6 |
eval_size: {}
|
| 7 |
eval_split: {}
|
| 8 |
eval_subset: {}
|
| 9 |
eval_template: {}
|
| 10 |
+
load_multi_datasets: false
|
| 11 |
train_data_files: {}
|
| 12 |
+
train_datasets: /aifs4su/yaodong/hantao/datasets/MMInstruct-GPT4V_mistral-7b_cosi_cut/merged/top1-40
|
| 13 |
train_name: text-image-to-text
|
| 14 |
train_optional_args: []
|
| 15 |
train_size: {}
|
| 16 |
train_split: train
|
| 17 |
+
train_template: MM_TI2T_LLAVA
|
| 18 |
logger_cfgs:
|
| 19 |
cache_dir: {}
|
| 20 |
log_project: align-anything
|
| 21 |
+
log_run_name: sft
|
| 22 |
log_type: wandb
|
| 23 |
+
output_dir: ../outputs/LLAVA_7B_cosi/top1-40
|
| 24 |
+
save_total_limit: 6
|
| 25 |
model_cfgs:
|
| 26 |
model_max_length: 4096
|
| 27 |
model_name_or_path: /aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf
|
|
|
|
| 31 |
adam_betas:
|
| 32 |
- 0.9
|
| 33 |
- 0.95
|
| 34 |
+
adam_epsilon: 1.0e-08
|
| 35 |
bf16: true
|
| 36 |
ds_cfgs: ds_z3_config.json
|
| 37 |
epochs: 3
|
|
|
|
| 41 |
freeze_language_model: false
|
| 42 |
freeze_mm_proj: false
|
| 43 |
freeze_vision_tower: true
|
| 44 |
+
gradient_accumulation_steps: 16
|
| 45 |
gradient_checkpointing: true
|
| 46 |
+
learning_rate: 2.0e-05
|
| 47 |
load_checkpoint: false
|
| 48 |
lr_scheduler_type: cosine
|
| 49 |
lr_warmup_ratio: 0.03
|
| 50 |
+
max_grad_norm: 1.0
|
| 51 |
per_device_eval_batch_size: 1
|
| 52 |
per_device_train_batch_size: 1
|
|
|
|
| 53 |
save_checkpoint: false
|
|
|
|
| 54 |
seed: 42
|
| 55 |
weight_decay: 0.0
|
environ.txt
CHANGED
|
@@ -84,75 +84,34 @@ BASH_FUNC_switchml%%=() { typeset swfound=1;
|
|
| 84 |
BUILD=x86_64-conda-linux-gnu
|
| 85 |
CC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
|
| 86 |
CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
|
| 87 |
-
CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/
|
| 88 |
CMAKE_ARGS=-DCMAKE_AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ar -DCMAKE_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib -DCMAKE_LINKER=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld -DCMAKE_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip -DCMAKE_BUILD_TYPE=Release
|
| 89 |
CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama:/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot/usr
|
| 90 |
CMD_WLM_CLUSTER_NAME=slurm
|
| 91 |
-
CONDA_BACKUP_ADDR2LINE=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-addr2line
|
| 92 |
-
CONDA_BACKUP_AR=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ar
|
| 93 |
-
CONDA_BACKUP_AS=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-as
|
| 94 |
-
CONDA_BACKUP_BUILD=x86_64-conda-linux-gnu
|
| 95 |
-
CONDA_BACKUP_CC=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cc
|
| 96 |
-
CONDA_BACKUP_CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cc
|
| 97 |
-
CONDA_BACKUP_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 98 |
-
CONDA_BACKUP_CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_anything:/aifs4su/yaodong/miniconda3/envs/hantao_anything/x86_64-conda-linux-gnu/sysroot/usr
|
| 99 |
-
CONDA_BACKUP_CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/x86_64-conda-linux-gnu/sysroot
|
| 100 |
-
CONDA_BACKUP_CPP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cpp
|
| 101 |
-
CONDA_BACKUP_CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 102 |
-
CONDA_BACKUP_CXX=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++
|
| 103 |
-
CONDA_BACKUP_CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++filt
|
| 104 |
-
CONDA_BACKUP_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 105 |
-
CONDA_BACKUP_CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++
|
| 106 |
-
CONDA_BACKUP_DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 107 |
-
CONDA_BACKUP_DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 108 |
-
CONDA_BACKUP_DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 109 |
-
CONDA_BACKUP_ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-elfedit
|
| 110 |
-
CONDA_BACKUP_GCC=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc
|
| 111 |
-
CONDA_BACKUP_GCC_AR=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-ar
|
| 112 |
-
CONDA_BACKUP_GCC_NM=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-nm
|
| 113 |
-
CONDA_BACKUP_GCC_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-ranlib
|
| 114 |
-
CONDA_BACKUP_GPROF=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gprof
|
| 115 |
-
CONDA_BACKUP_GXX=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-g++
|
| 116 |
-
CONDA_BACKUP_HOST=x86_64-conda-linux-gnu
|
| 117 |
-
CONDA_BACKUP_LD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ld
|
| 118 |
-
CONDA_BACKUP_LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib
|
| 119 |
-
CONDA_BACKUP_LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ld.gold
|
| 120 |
-
CONDA_BACKUP_NM=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-nm
|
| 121 |
-
CONDA_BACKUP_OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-objcopy
|
| 122 |
-
CONDA_BACKUP_OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-objdump
|
| 123 |
-
CONDA_BACKUP_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ranlib
|
| 124 |
-
CONDA_BACKUP_READELF=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-readelf
|
| 125 |
-
CONDA_BACKUP_SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-size
|
| 126 |
-
CONDA_BACKUP_STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-strings
|
| 127 |
-
CONDA_BACKUP_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-strip
|
| 128 |
-
CONDA_BACKUP__CONDA_PYTHON_SYSCONFIGDATA_NAME=_sysconfigdata_x86_64_conda_cos7_linux_gnu
|
| 129 |
-
CONDA_BACKUP_build_alias=x86_64-conda-linux-gnu
|
| 130 |
-
CONDA_BACKUP_host_alias=x86_64-conda-linux-gnu
|
| 131 |
CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot
|
| 132 |
CONDA_DEFAULT_ENV=hantao_llama
|
| 133 |
CONDA_EXE=/aifs4su/yaodong/miniconda3/bin/conda
|
| 134 |
CONDA_PREFIX=/aifs4su/yaodong/miniconda3/envs/hantao_llama
|
| 135 |
CONDA_PREFIX_1=/aifs4su/yaodong/miniconda3
|
| 136 |
-
CONDA_PREFIX_2=/aifs4su/yaodong/miniconda3/envs/hantao_anything
|
| 137 |
CONDA_PROMPT_MODIFIER=(hantao_llama)
|
| 138 |
CONDA_PYTHON_EXE=/aifs4su/yaodong/miniconda3/bin/python
|
| 139 |
-
CONDA_SHLVL=
|
| 140 |
CPATH=/cm/shared/apps/slurm/current/include
|
| 141 |
CPATH_modshare=/cm/shared/apps/slurm/current/include:1
|
| 142 |
CPP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cpp
|
| 143 |
-
CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/
|
| 144 |
CROSS_RANK=0
|
| 145 |
CROSS_SIZE=1
|
| 146 |
CUDA_MODULE_LOADING=LAZY
|
| 147 |
CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
|
| 148 |
CXX=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 149 |
CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++filt
|
| 150 |
-
CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/
|
| 151 |
CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 152 |
DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1028/bus
|
| 153 |
-
DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/
|
| 154 |
-
DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/
|
| 155 |
-
DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/
|
| 156 |
ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-elfedit
|
| 157 |
ENABLE_LMOD=0
|
| 158 |
GCC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc
|
|
@@ -172,7 +131,7 @@ KMP_DUPLICATE_LIB_OK=True
|
|
| 172 |
KMP_INIT_AT_FORK=FALSE
|
| 173 |
LANG=C.UTF-8
|
| 174 |
LD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld
|
| 175 |
-
LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/
|
| 176 |
LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld.gold
|
| 177 |
LD_LIBRARY_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/../../lib64:/usr/mpi/gcc/openmpi-4.1.7a1/lib:/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64
|
| 178 |
LD_LIBRARY_PATH_modshare=/cm/shared/apps/slurm/current/lib64:1:/usr/mpi/gcc/openmpi-4.1.7a1/lib:1:/cm/shared/apps/slurm/current/lib64/slurm:1
|
|
@@ -192,11 +151,12 @@ LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd
|
|
| 192 |
MANPATH=/usr/mpi/gcc/openmpi-4.1.7a1/share/man:/cm/shared/apps/slurm/current/man:/cm/local/apps/environment-modules/4.5.3/share/man:/usr/local/man:/usr/local/share/man:/usr/share/man:/cm/local/apps/environment-modules/current/share/man:/cm/local/apps/environment-modules/current/share/man
|
| 193 |
MANPATH_modshare=/usr/local/share/man:1:/usr/mpi/gcc/openmpi-4.1.7a1/share/man:1:/cm/local/apps/environment-modules/current/share/man:1:/cm/local/apps/environment-modules/4.5.3/share/man:1:/usr/local/man:1:/usr/share/man:1:/cm/shared/apps/slurm/current/man:1
|
| 194 |
MASTER_ADDR=127.0.0.1
|
| 195 |
-
MASTER_PORT=
|
| 196 |
MIG_PARTED_CHECKPOINT_FILE=/var/lib/nvidia-mig-manager/checkpoint.json
|
| 197 |
MIG_PARTED_CONFIG_FILE=/etc/nvidia-mig-manager/config.yaml
|
| 198 |
MIG_PARTED_HOOKS_FILE=/etc/nvidia-mig-manager/hooks.yaml
|
| 199 |
MODULEPATH=/cm/local/modulefiles:/cm/shared/modulefiles
|
|
|
|
| 200 |
MODULESHOME=/cm/local/apps/environment-modules/4.5.3
|
| 201 |
MODULES_CMD=/cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl
|
| 202 |
MODULES_SET_SHELL_STARTUP=0
|
|
@@ -209,12 +169,14 @@ NVCC_PREPEND_FLAGS_BACKUP= -ccbin=/aifs4su/yaodong/miniconda3/bin/x86_64-conda-l
|
|
| 209 |
NVITOP_MONITOR_MODE=colorful
|
| 210 |
OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objcopy
|
| 211 |
OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objdump
|
| 212 |
-
OLDPWD=/
|
| 213 |
PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin:/aifs4su/yaodong/miniconda3/condabin:/usr/mpi/gcc/openmpi-4.1.7a1/bin:/usr/lpp/mmfs/bin:/usr/local/cuda/bin:/opt/bin:/usr/lpp/mmfs/bin:/cm/shared/apps/slurm/current/sbin:/cm/shared/apps/slurm/current/bin:/usr/local/cuda/bin:/opt/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/sbin:/usr/sbin:/cm/local/apps/environment-modules/4.5.3/bin
|
| 214 |
PATH_modshare=/usr/mpi/gcc/openmpi-4.1.7a1/bin:1:/opt/bin/:1:/usr/bin:1:/usr/local/bin:1:/cm/shared/apps/slurm/current/bin:1:/cm/shared/apps/slurm/current/sbin:1:/bin:1:/snap/bin:1:/sbin:1:/usr/sbin:1:/cm/local/apps/environment-modules/4.5.3/bin:1:/usr/games:1:/usr/local/sbin:1:/usr/lpp/mmfs/bin:1:/usr/local/cuda/bin:1:/usr/local/games:1
|
| 215 |
PWD=/aifs4su/yaodong/hantao/align-anything/scripts
|
| 216 |
PYTHONHASHSEED=42
|
| 217 |
-
PYTHONPATH=/aifs4su/yaodong/hantao/align-anything/scripts
|
|
|
|
|
|
|
| 218 |
RANK=0
|
| 219 |
RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib
|
| 220 |
READELF=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-readelf
|
|
@@ -222,24 +184,24 @@ SHELL=/bin/bash
|
|
| 222 |
SHLVL=3
|
| 223 |
SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-size
|
| 224 |
SLURM_CONF=/cm/shared/apps/slurm/var/etc/slurm/slurm.conf
|
| 225 |
-
SSH_CLIENT=10.33.4.
|
| 226 |
-
SSH_CONNECTION=10.33.4.51
|
| 227 |
-
SSH_TTY=/dev/pts/
|
| 228 |
STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strings
|
| 229 |
STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip
|
| 230 |
TERM=screen
|
| 231 |
TERM_PROGRAM=tmux
|
| 232 |
TERM_PROGRAM_VERSION=3.2a
|
| 233 |
-
TMUX=/tmp/tmux-1028/default,
|
| 234 |
-
TMUX_PANE=%
|
| 235 |
USER=yangyaodong
|
| 236 |
WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
|
| 237 |
-
WANDB_SERVICE=2-
|
| 238 |
WORLD_SIZE=8
|
| 239 |
XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
|
| 240 |
XDG_RUNTIME_DIR=/run/user/1028
|
| 241 |
XDG_SESSION_CLASS=user
|
| 242 |
-
XDG_SESSION_ID=
|
| 243 |
XDG_SESSION_TYPE=tty
|
| 244 |
ZERO_STAGE=3
|
| 245 |
_=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/deepspeed
|
|
|
|
| 84 |
BUILD=x86_64-conda-linux-gnu
|
| 85 |
CC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
|
| 86 |
CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
|
| 87 |
+
CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 88 |
CMAKE_ARGS=-DCMAKE_AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ar -DCMAKE_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib -DCMAKE_LINKER=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld -DCMAKE_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip -DCMAKE_BUILD_TYPE=Release
|
| 89 |
CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama:/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot/usr
|
| 90 |
CMD_WLM_CLUSTER_NAME=slurm
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 91 |
CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot
|
| 92 |
CONDA_DEFAULT_ENV=hantao_llama
|
| 93 |
CONDA_EXE=/aifs4su/yaodong/miniconda3/bin/conda
|
| 94 |
CONDA_PREFIX=/aifs4su/yaodong/miniconda3/envs/hantao_llama
|
| 95 |
CONDA_PREFIX_1=/aifs4su/yaodong/miniconda3
|
|
|
|
| 96 |
CONDA_PROMPT_MODIFIER=(hantao_llama)
|
| 97 |
CONDA_PYTHON_EXE=/aifs4su/yaodong/miniconda3/bin/python
|
| 98 |
+
CONDA_SHLVL=2
|
| 99 |
CPATH=/cm/shared/apps/slurm/current/include
|
| 100 |
CPATH_modshare=/cm/shared/apps/slurm/current/include:1
|
| 101 |
CPP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cpp
|
| 102 |
+
CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 103 |
CROSS_RANK=0
|
| 104 |
CROSS_SIZE=1
|
| 105 |
CUDA_MODULE_LOADING=LAZY
|
| 106 |
CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
|
| 107 |
CXX=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 108 |
CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++filt
|
| 109 |
+
CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 110 |
CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 111 |
DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1028/bus
|
| 112 |
+
DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include
|
| 113 |
+
DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include
|
| 114 |
+
DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_llama/include
|
| 115 |
ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-elfedit
|
| 116 |
ENABLE_LMOD=0
|
| 117 |
GCC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc
|
|
|
|
| 131 |
KMP_INIT_AT_FORK=FALSE
|
| 132 |
LANG=C.UTF-8
|
| 133 |
LD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld
|
| 134 |
+
LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 135 |
LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld.gold
|
| 136 |
LD_LIBRARY_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/../../lib64:/usr/mpi/gcc/openmpi-4.1.7a1/lib:/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64
|
| 137 |
LD_LIBRARY_PATH_modshare=/cm/shared/apps/slurm/current/lib64:1:/usr/mpi/gcc/openmpi-4.1.7a1/lib:1:/cm/shared/apps/slurm/current/lib64/slurm:1
|
|
|
|
| 151 |
MANPATH=/usr/mpi/gcc/openmpi-4.1.7a1/share/man:/cm/shared/apps/slurm/current/man:/cm/local/apps/environment-modules/4.5.3/share/man:/usr/local/man:/usr/local/share/man:/usr/share/man:/cm/local/apps/environment-modules/current/share/man:/cm/local/apps/environment-modules/current/share/man
|
| 152 |
MANPATH_modshare=/usr/local/share/man:1:/usr/mpi/gcc/openmpi-4.1.7a1/share/man:1:/cm/local/apps/environment-modules/current/share/man:1:/cm/local/apps/environment-modules/4.5.3/share/man:1:/usr/local/man:1:/usr/share/man:1:/cm/shared/apps/slurm/current/man:1
|
| 153 |
MASTER_ADDR=127.0.0.1
|
| 154 |
+
MASTER_PORT=46590
|
| 155 |
MIG_PARTED_CHECKPOINT_FILE=/var/lib/nvidia-mig-manager/checkpoint.json
|
| 156 |
MIG_PARTED_CONFIG_FILE=/etc/nvidia-mig-manager/config.yaml
|
| 157 |
MIG_PARTED_HOOKS_FILE=/etc/nvidia-mig-manager/hooks.yaml
|
| 158 |
MODULEPATH=/cm/local/modulefiles:/cm/shared/modulefiles
|
| 159 |
+
MODULEPATH_modshare=/cm/local/modulefiles:1:/cm/shared/modulefiles:3
|
| 160 |
MODULESHOME=/cm/local/apps/environment-modules/4.5.3
|
| 161 |
MODULES_CMD=/cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl
|
| 162 |
MODULES_SET_SHELL_STARTUP=0
|
|
|
|
| 169 |
NVITOP_MONITOR_MODE=colorful
|
| 170 |
OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objcopy
|
| 171 |
OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objdump
|
| 172 |
+
OLDPWD=/home/yangyaodong
|
| 173 |
PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin:/aifs4su/yaodong/miniconda3/condabin:/usr/mpi/gcc/openmpi-4.1.7a1/bin:/usr/lpp/mmfs/bin:/usr/local/cuda/bin:/opt/bin:/usr/lpp/mmfs/bin:/cm/shared/apps/slurm/current/sbin:/cm/shared/apps/slurm/current/bin:/usr/local/cuda/bin:/opt/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/sbin:/usr/sbin:/cm/local/apps/environment-modules/4.5.3/bin
|
| 174 |
PATH_modshare=/usr/mpi/gcc/openmpi-4.1.7a1/bin:1:/opt/bin/:1:/usr/bin:1:/usr/local/bin:1:/cm/shared/apps/slurm/current/bin:1:/cm/shared/apps/slurm/current/sbin:1:/bin:1:/snap/bin:1:/sbin:1:/usr/sbin:1:/cm/local/apps/environment-modules/4.5.3/bin:1:/usr/games:1:/usr/local/sbin:1:/usr/lpp/mmfs/bin:1:/usr/local/cuda/bin:1:/usr/local/games:1
|
| 175 |
PWD=/aifs4su/yaodong/hantao/align-anything/scripts
|
| 176 |
PYTHONHASHSEED=42
|
| 177 |
+
PYTHONPATH=/aifs4su/yaodong/hantao/align-anything/scripts
|
| 178 |
+
QT_QPA_FONTDIR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/qt/fonts
|
| 179 |
+
QT_QPA_PLATFORM_PLUGIN_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/qt/plugins
|
| 180 |
RANK=0
|
| 181 |
RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib
|
| 182 |
READELF=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-readelf
|
|
|
|
| 184 |
SHLVL=3
|
| 185 |
SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-size
|
| 186 |
SLURM_CONF=/cm/shared/apps/slurm/var/etc/slurm/slurm.conf
|
| 187 |
+
SSH_CLIENT=10.33.4.51 34180 22
|
| 188 |
+
SSH_CONNECTION=10.33.4.51 34180 10.33.4.231 22
|
| 189 |
+
SSH_TTY=/dev/pts/8
|
| 190 |
STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strings
|
| 191 |
STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip
|
| 192 |
TERM=screen
|
| 193 |
TERM_PROGRAM=tmux
|
| 194 |
TERM_PROGRAM_VERSION=3.2a
|
| 195 |
+
TMUX=/tmp/tmux-1028/default,1588807,0
|
| 196 |
+
TMUX_PANE=%0
|
| 197 |
USER=yangyaodong
|
| 198 |
WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
|
| 199 |
+
WANDB_SERVICE=2-1593022-tcp-localhost-46867
|
| 200 |
WORLD_SIZE=8
|
| 201 |
XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
|
| 202 |
XDG_RUNTIME_DIR=/run/user/1028
|
| 203 |
XDG_SESSION_CLASS=user
|
| 204 |
+
XDG_SESSION_ID=1733
|
| 205 |
XDG_SESSION_TYPE=tty
|
| 206 |
ZERO_STAGE=3
|
| 207 |
_=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/deepspeed
|
script.sh
CHANGED
|
@@ -15,38 +15,35 @@
|
|
| 15 |
# limitations under the License.
|
| 16 |
# ==============================================================================
|
| 17 |
|
| 18 |
-
DATASETS_NAME=("top1-
|
| 19 |
-
# DATASETS_NAME=("top1-10")
|
| 20 |
-
#
|
| 21 |
-
MODEL_NAME_OR_PATH="/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf" # model path
|
| 22 |
|
| 23 |
-
|
| 24 |
|
| 25 |
-
for DATASET_NAME in ${DATASETS_NAME[@]}; do
|
| 26 |
-
TRAIN_DATASETS="/aifs4su/yaodong/hantao/datasets/
|
| 27 |
-
TRAIN_TEMPLATE="
|
| 28 |
TRAIN_NAME="text-image-to-text" # dataset name
|
| 29 |
TRAIN_SPLIT="train" # split the dataset
|
| 30 |
|
| 31 |
-
OUTPUT_DIR="../outputs/
|
| 32 |
|
| 33 |
-
|
| 34 |
-
|
| 35 |
|
| 36 |
-
|
| 37 |
-
|
| 38 |
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
| 51 |
-
|
| 52 |
done
|
|
|
|
| 15 |
# limitations under the License.
|
| 16 |
# ==============================================================================
|
| 17 |
|
| 18 |
+
DATASETS_NAME=("top1-40")
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
+
MODEL_NAME_OR_PATH="/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf" # model path
|
| 21 |
|
| 22 |
+
for DATASET_NAME in "${DATASETS_NAME[@]}"; do
|
| 23 |
+
TRAIN_DATASETS="/aifs4su/yaodong/hantao/datasets/MMInstruct-GPT4V_mistral-7b_cosi_cut/merged/${DATASET_NAME}" # dataset path
|
| 24 |
+
TRAIN_TEMPLATE="MM_TI2T_LLAVA" # dataset template
|
| 25 |
TRAIN_NAME="text-image-to-text" # dataset name
|
| 26 |
TRAIN_SPLIT="train" # split the dataset
|
| 27 |
|
| 28 |
+
OUTPUT_DIR="../outputs/LLAVA_7B_cosi/${DATASET_NAME}" # output dir
|
| 29 |
|
| 30 |
+
# For wandb online logging
|
| 31 |
+
export WANDB_API_KEY="7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33"
|
| 32 |
|
| 33 |
+
# Source the setup script
|
| 34 |
+
source ./setup.sh
|
| 35 |
|
| 36 |
+
# Execute deepspeed command
|
| 37 |
+
deepspeed \
|
| 38 |
+
--master_port ${MASTER_PORT} \
|
| 39 |
+
--module align_anything.trainers.text_image_to_text.sft \
|
| 40 |
+
--model_name_or_path ${MODEL_NAME_OR_PATH} \
|
| 41 |
+
--train_datasets ${TRAIN_DATASETS} \
|
| 42 |
+
--train_template ${TRAIN_TEMPLATE} \
|
| 43 |
+
--train_split ${TRAIN_SPLIT} \
|
| 44 |
+
--train_name ${TRAIN_NAME} \
|
| 45 |
+
--output_dir ${OUTPUT_DIR} \
|
| 46 |
+
--save_total_limit 6 \
|
| 47 |
+
--train_batch_size 16 \
|
| 48 |
+
--epochs 3
|
| 49 |
done
|
slice_14649/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<image>": 32000,
|
| 3 |
+
"<pad>": 32001
|
| 4 |
+
}
|
slice_14649/chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
|
| 3 |
+
}
|
slice_14649/config.json
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlavaNextForConditionalGeneration"
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"ignore_index": -100,
|
| 9 |
+
"image_grid_pinpoints": [
|
| 10 |
+
[
|
| 11 |
+
336,
|
| 12 |
+
672
|
| 13 |
+
],
|
| 14 |
+
[
|
| 15 |
+
672,
|
| 16 |
+
336
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
672
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
1008,
|
| 24 |
+
336
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
336,
|
| 28 |
+
1008
|
| 29 |
+
]
|
| 30 |
+
],
|
| 31 |
+
"image_seq_length": 576,
|
| 32 |
+
"image_token_index": 32000,
|
| 33 |
+
"model_type": "llava_next",
|
| 34 |
+
"multimodal_projector_bias": true,
|
| 35 |
+
"pad_token_id": 32001,
|
| 36 |
+
"projector_hidden_act": "gelu",
|
| 37 |
+
"text_config": {
|
| 38 |
+
"_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 39 |
+
"architectures": [
|
| 40 |
+
"MistralForCausalLM"
|
| 41 |
+
],
|
| 42 |
+
"attention_dropout": 0.0,
|
| 43 |
+
"head_dim": 128,
|
| 44 |
+
"hidden_act": "silu",
|
| 45 |
+
"hidden_size": 4096,
|
| 46 |
+
"initializer_range": 0.02,
|
| 47 |
+
"intermediate_size": 14336,
|
| 48 |
+
"max_position_embeddings": 32768,
|
| 49 |
+
"model_type": "mistral",
|
| 50 |
+
"num_attention_heads": 32,
|
| 51 |
+
"num_hidden_layers": 32,
|
| 52 |
+
"num_key_value_heads": 8,
|
| 53 |
+
"rms_norm_eps": 1e-05,
|
| 54 |
+
"rope_theta": 1000000.0,
|
| 55 |
+
"sliding_window": null,
|
| 56 |
+
"torch_dtype": "bfloat16",
|
| 57 |
+
"use_cache": true,
|
| 58 |
+
"vocab_size": 32064
|
| 59 |
+
},
|
| 60 |
+
"tie_word_embeddings": false,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.50.0",
|
| 63 |
+
"use_image_newline_parameter": true,
|
| 64 |
+
"vision_config": {
|
| 65 |
+
"attention_dropout": 0.0,
|
| 66 |
+
"hidden_act": "quick_gelu",
|
| 67 |
+
"hidden_size": 1024,
|
| 68 |
+
"image_size": 336,
|
| 69 |
+
"initializer_factor": 1.0,
|
| 70 |
+
"initializer_range": 0.02,
|
| 71 |
+
"intermediate_size": 4096,
|
| 72 |
+
"layer_norm_eps": 1e-05,
|
| 73 |
+
"model_type": "clip_vision_model",
|
| 74 |
+
"num_attention_heads": 16,
|
| 75 |
+
"num_channels": 3,
|
| 76 |
+
"num_hidden_layers": 24,
|
| 77 |
+
"patch_size": 14,
|
| 78 |
+
"projection_dim": 768,
|
| 79 |
+
"torch_dtype": "bfloat16",
|
| 80 |
+
"vocab_size": 32000
|
| 81 |
+
},
|
| 82 |
+
"vision_feature_layer": -2,
|
| 83 |
+
"vision_feature_select_strategy": "default",
|
| 84 |
+
"vocab_size": 32064
|
| 85 |
+
}
|
slice_14649/preprocessor_config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"aspect_ratio_setting": "anyres",
|
| 3 |
+
"crop_size": {
|
| 4 |
+
"height": 336,
|
| 5 |
+
"width": 336
|
| 6 |
+
},
|
| 7 |
+
"do_center_crop": true,
|
| 8 |
+
"do_convert_rgb": true,
|
| 9 |
+
"do_normalize": true,
|
| 10 |
+
"do_pad": true,
|
| 11 |
+
"do_rescale": true,
|
| 12 |
+
"do_resize": true,
|
| 13 |
+
"image_grid_pinpoints": [
|
| 14 |
+
[
|
| 15 |
+
336,
|
| 16 |
+
672
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
336
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
672,
|
| 24 |
+
672
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
1008,
|
| 28 |
+
336
|
| 29 |
+
],
|
| 30 |
+
[
|
| 31 |
+
336,
|
| 32 |
+
1008
|
| 33 |
+
]
|
| 34 |
+
],
|
| 35 |
+
"image_mean": [
|
| 36 |
+
0.48145466,
|
| 37 |
+
0.4578275,
|
| 38 |
+
0.40821073
|
| 39 |
+
],
|
| 40 |
+
"image_processor_type": "LlavaNextImageProcessor",
|
| 41 |
+
"image_std": [
|
| 42 |
+
0.26862954,
|
| 43 |
+
0.26130258,
|
| 44 |
+
0.27577711
|
| 45 |
+
],
|
| 46 |
+
"processor_class": "LlavaNextProcessor",
|
| 47 |
+
"resample": 3,
|
| 48 |
+
"rescale_factor": 0.00392156862745098,
|
| 49 |
+
"size": {
|
| 50 |
+
"shortest_edge": 336
|
| 51 |
+
}
|
| 52 |
+
}
|
slice_14649/processor_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"image_token": "<image>",
|
| 3 |
+
"num_additional_image_tokens": 1,
|
| 4 |
+
"patch_size": 14,
|
| 5 |
+
"processor_class": "LlavaNextProcessor",
|
| 6 |
+
"vision_feature_select_strategy": "default"
|
| 7 |
+
}
|
slice_14649/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2f1fd230f3f01a6098c034d16a364fa803331cdd2e99039e7000a0f3a9a002b4
|
| 3 |
+
size 15133733934
|
slice_14649/special_tokens_map.json
ADDED
|
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"image_token": "<image>",
|
| 17 |
+
"pad_token": {
|
| 18 |
+
"content": "<pad>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
},
|
| 24 |
+
"unk_token": {
|
| 25 |
+
"content": "<unk>",
|
| 26 |
+
"lstrip": false,
|
| 27 |
+
"normalized": false,
|
| 28 |
+
"rstrip": false,
|
| 29 |
+
"single_word": false
|
| 30 |
+
}
|
| 31 |
+
}
|
slice_14649/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_14649/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
slice_14649/tokenizer_config.json
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": false,
|
| 4 |
+
"add_prefix_space": null,
|
| 5 |
+
"added_tokens_decoder": {
|
| 6 |
+
"0": {
|
| 7 |
+
"content": "<unk>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false,
|
| 12 |
+
"special": true
|
| 13 |
+
},
|
| 14 |
+
"1": {
|
| 15 |
+
"content": "<s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false,
|
| 20 |
+
"special": true
|
| 21 |
+
},
|
| 22 |
+
"2": {
|
| 23 |
+
"content": "</s>",
|
| 24 |
+
"lstrip": false,
|
| 25 |
+
"normalized": false,
|
| 26 |
+
"rstrip": false,
|
| 27 |
+
"single_word": false,
|
| 28 |
+
"special": true
|
| 29 |
+
},
|
| 30 |
+
"32000": {
|
| 31 |
+
"content": "<image>",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false,
|
| 36 |
+
"special": true
|
| 37 |
+
},
|
| 38 |
+
"32001": {
|
| 39 |
+
"content": "<pad>",
|
| 40 |
+
"lstrip": false,
|
| 41 |
+
"normalized": false,
|
| 42 |
+
"rstrip": false,
|
| 43 |
+
"single_word": false,
|
| 44 |
+
"special": true
|
| 45 |
+
}
|
| 46 |
+
},
|
| 47 |
+
"additional_special_tokens": [],
|
| 48 |
+
"bos_token": "<s>",
|
| 49 |
+
"chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
| 50 |
+
"clean_up_tokenization_spaces": false,
|
| 51 |
+
"eos_token": "</s>",
|
| 52 |
+
"extra_special_tokens": {
|
| 53 |
+
"image_token": "<image>"
|
| 54 |
+
},
|
| 55 |
+
"image_token": "<image>",
|
| 56 |
+
"legacy": true,
|
| 57 |
+
"max_length": null,
|
| 58 |
+
"model_max_length": 4096,
|
| 59 |
+
"pad_to_multiple_of": null,
|
| 60 |
+
"pad_token": "<pad>",
|
| 61 |
+
"pad_token_type_id": 0,
|
| 62 |
+
"padding_side": "right",
|
| 63 |
+
"processor_class": "LlavaNextProcessor",
|
| 64 |
+
"sp_model_kwargs": {},
|
| 65 |
+
"spaces_between_special_tokens": false,
|
| 66 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 67 |
+
"unk_token": "<unk>",
|
| 68 |
+
"use_default_system_prompt": false
|
| 69 |
+
}
|
slice_19532/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<image>": 32000,
|
| 3 |
+
"<pad>": 32001
|
| 4 |
+
}
|
slice_19532/chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
|
| 3 |
+
}
|
slice_19532/config.json
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlavaNextForConditionalGeneration"
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"ignore_index": -100,
|
| 9 |
+
"image_grid_pinpoints": [
|
| 10 |
+
[
|
| 11 |
+
336,
|
| 12 |
+
672
|
| 13 |
+
],
|
| 14 |
+
[
|
| 15 |
+
672,
|
| 16 |
+
336
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
672
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
1008,
|
| 24 |
+
336
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
336,
|
| 28 |
+
1008
|
| 29 |
+
]
|
| 30 |
+
],
|
| 31 |
+
"image_seq_length": 576,
|
| 32 |
+
"image_token_index": 32000,
|
| 33 |
+
"model_type": "llava_next",
|
| 34 |
+
"multimodal_projector_bias": true,
|
| 35 |
+
"pad_token_id": 32001,
|
| 36 |
+
"projector_hidden_act": "gelu",
|
| 37 |
+
"text_config": {
|
| 38 |
+
"_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 39 |
+
"architectures": [
|
| 40 |
+
"MistralForCausalLM"
|
| 41 |
+
],
|
| 42 |
+
"attention_dropout": 0.0,
|
| 43 |
+
"head_dim": 128,
|
| 44 |
+
"hidden_act": "silu",
|
| 45 |
+
"hidden_size": 4096,
|
| 46 |
+
"initializer_range": 0.02,
|
| 47 |
+
"intermediate_size": 14336,
|
| 48 |
+
"max_position_embeddings": 32768,
|
| 49 |
+
"model_type": "mistral",
|
| 50 |
+
"num_attention_heads": 32,
|
| 51 |
+
"num_hidden_layers": 32,
|
| 52 |
+
"num_key_value_heads": 8,
|
| 53 |
+
"rms_norm_eps": 1e-05,
|
| 54 |
+
"rope_theta": 1000000.0,
|
| 55 |
+
"sliding_window": null,
|
| 56 |
+
"torch_dtype": "bfloat16",
|
| 57 |
+
"use_cache": true,
|
| 58 |
+
"vocab_size": 32064
|
| 59 |
+
},
|
| 60 |
+
"tie_word_embeddings": false,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.50.0",
|
| 63 |
+
"use_image_newline_parameter": true,
|
| 64 |
+
"vision_config": {
|
| 65 |
+
"attention_dropout": 0.0,
|
| 66 |
+
"hidden_act": "quick_gelu",
|
| 67 |
+
"hidden_size": 1024,
|
| 68 |
+
"image_size": 336,
|
| 69 |
+
"initializer_factor": 1.0,
|
| 70 |
+
"initializer_range": 0.02,
|
| 71 |
+
"intermediate_size": 4096,
|
| 72 |
+
"layer_norm_eps": 1e-05,
|
| 73 |
+
"model_type": "clip_vision_model",
|
| 74 |
+
"num_attention_heads": 16,
|
| 75 |
+
"num_channels": 3,
|
| 76 |
+
"num_hidden_layers": 24,
|
| 77 |
+
"patch_size": 14,
|
| 78 |
+
"projection_dim": 768,
|
| 79 |
+
"torch_dtype": "bfloat16",
|
| 80 |
+
"vocab_size": 32000
|
| 81 |
+
},
|
| 82 |
+
"vision_feature_layer": -2,
|
| 83 |
+
"vision_feature_select_strategy": "default",
|
| 84 |
+
"vocab_size": 32064
|
| 85 |
+
}
|
slice_19532/preprocessor_config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"aspect_ratio_setting": "anyres",
|
| 3 |
+
"crop_size": {
|
| 4 |
+
"height": 336,
|
| 5 |
+
"width": 336
|
| 6 |
+
},
|
| 7 |
+
"do_center_crop": true,
|
| 8 |
+
"do_convert_rgb": true,
|
| 9 |
+
"do_normalize": true,
|
| 10 |
+
"do_pad": true,
|
| 11 |
+
"do_rescale": true,
|
| 12 |
+
"do_resize": true,
|
| 13 |
+
"image_grid_pinpoints": [
|
| 14 |
+
[
|
| 15 |
+
336,
|
| 16 |
+
672
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
336
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
672,
|
| 24 |
+
672
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
1008,
|
| 28 |
+
336
|
| 29 |
+
],
|
| 30 |
+
[
|
| 31 |
+
336,
|
| 32 |
+
1008
|
| 33 |
+
]
|
| 34 |
+
],
|
| 35 |
+
"image_mean": [
|
| 36 |
+
0.48145466,
|
| 37 |
+
0.4578275,
|
| 38 |
+
0.40821073
|
| 39 |
+
],
|
| 40 |
+
"image_processor_type": "LlavaNextImageProcessor",
|
| 41 |
+
"image_std": [
|
| 42 |
+
0.26862954,
|
| 43 |
+
0.26130258,
|
| 44 |
+
0.27577711
|
| 45 |
+
],
|
| 46 |
+
"processor_class": "LlavaNextProcessor",
|
| 47 |
+
"resample": 3,
|
| 48 |
+
"rescale_factor": 0.00392156862745098,
|
| 49 |
+
"size": {
|
| 50 |
+
"shortest_edge": 336
|
| 51 |
+
}
|
| 52 |
+
}
|
slice_19532/processor_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"image_token": "<image>",
|
| 3 |
+
"num_additional_image_tokens": 1,
|
| 4 |
+
"patch_size": 14,
|
| 5 |
+
"processor_class": "LlavaNextProcessor",
|
| 6 |
+
"vision_feature_select_strategy": "default"
|
| 7 |
+
}
|
slice_19532/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:ab661a9d25841f2346911001f2fff525ab0473e4430f7d52b42addaf867699ca
|
| 3 |
+
size 15133733934
|
slice_19532/special_tokens_map.json
ADDED
|
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"image_token": "<image>",
|
| 17 |
+
"pad_token": {
|
| 18 |
+
"content": "<pad>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
},
|
| 24 |
+
"unk_token": {
|
| 25 |
+
"content": "<unk>",
|
| 26 |
+
"lstrip": false,
|
| 27 |
+
"normalized": false,
|
| 28 |
+
"rstrip": false,
|
| 29 |
+
"single_word": false
|
| 30 |
+
}
|
| 31 |
+
}
|
slice_19532/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_19532/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
slice_19532/tokenizer_config.json
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": false,
|
| 4 |
+
"add_prefix_space": null,
|
| 5 |
+
"added_tokens_decoder": {
|
| 6 |
+
"0": {
|
| 7 |
+
"content": "<unk>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false,
|
| 12 |
+
"special": true
|
| 13 |
+
},
|
| 14 |
+
"1": {
|
| 15 |
+
"content": "<s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false,
|
| 20 |
+
"special": true
|
| 21 |
+
},
|
| 22 |
+
"2": {
|
| 23 |
+
"content": "</s>",
|
| 24 |
+
"lstrip": false,
|
| 25 |
+
"normalized": false,
|
| 26 |
+
"rstrip": false,
|
| 27 |
+
"single_word": false,
|
| 28 |
+
"special": true
|
| 29 |
+
},
|
| 30 |
+
"32000": {
|
| 31 |
+
"content": "<image>",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false,
|
| 36 |
+
"special": true
|
| 37 |
+
},
|
| 38 |
+
"32001": {
|
| 39 |
+
"content": "<pad>",
|
| 40 |
+
"lstrip": false,
|
| 41 |
+
"normalized": false,
|
| 42 |
+
"rstrip": false,
|
| 43 |
+
"single_word": false,
|
| 44 |
+
"special": true
|
| 45 |
+
}
|
| 46 |
+
},
|
| 47 |
+
"additional_special_tokens": [],
|
| 48 |
+
"bos_token": "<s>",
|
| 49 |
+
"chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
| 50 |
+
"clean_up_tokenization_spaces": false,
|
| 51 |
+
"eos_token": "</s>",
|
| 52 |
+
"extra_special_tokens": {
|
| 53 |
+
"image_token": "<image>"
|
| 54 |
+
},
|
| 55 |
+
"image_token": "<image>",
|
| 56 |
+
"legacy": true,
|
| 57 |
+
"max_length": null,
|
| 58 |
+
"model_max_length": 4096,
|
| 59 |
+
"pad_to_multiple_of": null,
|
| 60 |
+
"pad_token": "<pad>",
|
| 61 |
+
"pad_token_type_id": 0,
|
| 62 |
+
"padding_side": "right",
|
| 63 |
+
"processor_class": "LlavaNextProcessor",
|
| 64 |
+
"sp_model_kwargs": {},
|
| 65 |
+
"spaces_between_special_tokens": false,
|
| 66 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 67 |
+
"unk_token": "<unk>",
|
| 68 |
+
"use_default_system_prompt": false
|
| 69 |
+
}
|
slice_24415/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<image>": 32000,
|
| 3 |
+
"<pad>": 32001
|
| 4 |
+
}
|
slice_24415/chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
|
| 3 |
+
}
|
slice_24415/config.json
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlavaNextForConditionalGeneration"
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"ignore_index": -100,
|
| 9 |
+
"image_grid_pinpoints": [
|
| 10 |
+
[
|
| 11 |
+
336,
|
| 12 |
+
672
|
| 13 |
+
],
|
| 14 |
+
[
|
| 15 |
+
672,
|
| 16 |
+
336
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
672
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
1008,
|
| 24 |
+
336
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
336,
|
| 28 |
+
1008
|
| 29 |
+
]
|
| 30 |
+
],
|
| 31 |
+
"image_seq_length": 576,
|
| 32 |
+
"image_token_index": 32000,
|
| 33 |
+
"model_type": "llava_next",
|
| 34 |
+
"multimodal_projector_bias": true,
|
| 35 |
+
"pad_token_id": 32001,
|
| 36 |
+
"projector_hidden_act": "gelu",
|
| 37 |
+
"text_config": {
|
| 38 |
+
"_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 39 |
+
"architectures": [
|
| 40 |
+
"MistralForCausalLM"
|
| 41 |
+
],
|
| 42 |
+
"attention_dropout": 0.0,
|
| 43 |
+
"head_dim": 128,
|
| 44 |
+
"hidden_act": "silu",
|
| 45 |
+
"hidden_size": 4096,
|
| 46 |
+
"initializer_range": 0.02,
|
| 47 |
+
"intermediate_size": 14336,
|
| 48 |
+
"max_position_embeddings": 32768,
|
| 49 |
+
"model_type": "mistral",
|
| 50 |
+
"num_attention_heads": 32,
|
| 51 |
+
"num_hidden_layers": 32,
|
| 52 |
+
"num_key_value_heads": 8,
|
| 53 |
+
"rms_norm_eps": 1e-05,
|
| 54 |
+
"rope_theta": 1000000.0,
|
| 55 |
+
"sliding_window": null,
|
| 56 |
+
"torch_dtype": "bfloat16",
|
| 57 |
+
"use_cache": true,
|
| 58 |
+
"vocab_size": 32064
|
| 59 |
+
},
|
| 60 |
+
"tie_word_embeddings": false,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.50.0",
|
| 63 |
+
"use_image_newline_parameter": true,
|
| 64 |
+
"vision_config": {
|
| 65 |
+
"attention_dropout": 0.0,
|
| 66 |
+
"hidden_act": "quick_gelu",
|
| 67 |
+
"hidden_size": 1024,
|
| 68 |
+
"image_size": 336,
|
| 69 |
+
"initializer_factor": 1.0,
|
| 70 |
+
"initializer_range": 0.02,
|
| 71 |
+
"intermediate_size": 4096,
|
| 72 |
+
"layer_norm_eps": 1e-05,
|
| 73 |
+
"model_type": "clip_vision_model",
|
| 74 |
+
"num_attention_heads": 16,
|
| 75 |
+
"num_channels": 3,
|
| 76 |
+
"num_hidden_layers": 24,
|
| 77 |
+
"patch_size": 14,
|
| 78 |
+
"projection_dim": 768,
|
| 79 |
+
"torch_dtype": "bfloat16",
|
| 80 |
+
"vocab_size": 32000
|
| 81 |
+
},
|
| 82 |
+
"vision_feature_layer": -2,
|
| 83 |
+
"vision_feature_select_strategy": "default",
|
| 84 |
+
"vocab_size": 32064
|
| 85 |
+
}
|
slice_24415/preprocessor_config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"aspect_ratio_setting": "anyres",
|
| 3 |
+
"crop_size": {
|
| 4 |
+
"height": 336,
|
| 5 |
+
"width": 336
|
| 6 |
+
},
|
| 7 |
+
"do_center_crop": true,
|
| 8 |
+
"do_convert_rgb": true,
|
| 9 |
+
"do_normalize": true,
|
| 10 |
+
"do_pad": true,
|
| 11 |
+
"do_rescale": true,
|
| 12 |
+
"do_resize": true,
|
| 13 |
+
"image_grid_pinpoints": [
|
| 14 |
+
[
|
| 15 |
+
336,
|
| 16 |
+
672
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
336
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
672,
|
| 24 |
+
672
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
1008,
|
| 28 |
+
336
|
| 29 |
+
],
|
| 30 |
+
[
|
| 31 |
+
336,
|
| 32 |
+
1008
|
| 33 |
+
]
|
| 34 |
+
],
|
| 35 |
+
"image_mean": [
|
| 36 |
+
0.48145466,
|
| 37 |
+
0.4578275,
|
| 38 |
+
0.40821073
|
| 39 |
+
],
|
| 40 |
+
"image_processor_type": "LlavaNextImageProcessor",
|
| 41 |
+
"image_std": [
|
| 42 |
+
0.26862954,
|
| 43 |
+
0.26130258,
|
| 44 |
+
0.27577711
|
| 45 |
+
],
|
| 46 |
+
"processor_class": "LlavaNextProcessor",
|
| 47 |
+
"resample": 3,
|
| 48 |
+
"rescale_factor": 0.00392156862745098,
|
| 49 |
+
"size": {
|
| 50 |
+
"shortest_edge": 336
|
| 51 |
+
}
|
| 52 |
+
}
|
slice_24415/processor_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"image_token": "<image>",
|
| 3 |
+
"num_additional_image_tokens": 1,
|
| 4 |
+
"patch_size": 14,
|
| 5 |
+
"processor_class": "LlavaNextProcessor",
|
| 6 |
+
"vision_feature_select_strategy": "default"
|
| 7 |
+
}
|
slice_24415/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a1067de6e6a2dd60514331929e5e8fa8bdea95b178e2d56f3e10d53cb1565a34
|
| 3 |
+
size 15133733934
|
slice_24415/special_tokens_map.json
ADDED
|
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"image_token": "<image>",
|
| 17 |
+
"pad_token": {
|
| 18 |
+
"content": "<pad>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
},
|
| 24 |
+
"unk_token": {
|
| 25 |
+
"content": "<unk>",
|
| 26 |
+
"lstrip": false,
|
| 27 |
+
"normalized": false,
|
| 28 |
+
"rstrip": false,
|
| 29 |
+
"single_word": false
|
| 30 |
+
}
|
| 31 |
+
}
|
slice_24415/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_24415/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
slice_24415/tokenizer_config.json
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": false,
|
| 4 |
+
"add_prefix_space": null,
|
| 5 |
+
"added_tokens_decoder": {
|
| 6 |
+
"0": {
|
| 7 |
+
"content": "<unk>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false,
|
| 12 |
+
"special": true
|
| 13 |
+
},
|
| 14 |
+
"1": {
|
| 15 |
+
"content": "<s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false,
|
| 20 |
+
"special": true
|
| 21 |
+
},
|
| 22 |
+
"2": {
|
| 23 |
+
"content": "</s>",
|
| 24 |
+
"lstrip": false,
|
| 25 |
+
"normalized": false,
|
| 26 |
+
"rstrip": false,
|
| 27 |
+
"single_word": false,
|
| 28 |
+
"special": true
|
| 29 |
+
},
|
| 30 |
+
"32000": {
|
| 31 |
+
"content": "<image>",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false,
|
| 36 |
+
"special": true
|
| 37 |
+
},
|
| 38 |
+
"32001": {
|
| 39 |
+
"content": "<pad>",
|
| 40 |
+
"lstrip": false,
|
| 41 |
+
"normalized": false,
|
| 42 |
+
"rstrip": false,
|
| 43 |
+
"single_word": false,
|
| 44 |
+
"special": true
|
| 45 |
+
}
|
| 46 |
+
},
|
| 47 |
+
"additional_special_tokens": [],
|
| 48 |
+
"bos_token": "<s>",
|
| 49 |
+
"chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
| 50 |
+
"clean_up_tokenization_spaces": false,
|
| 51 |
+
"eos_token": "</s>",
|
| 52 |
+
"extra_special_tokens": {
|
| 53 |
+
"image_token": "<image>"
|
| 54 |
+
},
|
| 55 |
+
"image_token": "<image>",
|
| 56 |
+
"legacy": true,
|
| 57 |
+
"max_length": null,
|
| 58 |
+
"model_max_length": 4096,
|
| 59 |
+
"pad_to_multiple_of": null,
|
| 60 |
+
"pad_token": "<pad>",
|
| 61 |
+
"pad_token_type_id": 0,
|
| 62 |
+
"padding_side": "right",
|
| 63 |
+
"processor_class": "LlavaNextProcessor",
|
| 64 |
+
"sp_model_kwargs": {},
|
| 65 |
+
"spaces_between_special_tokens": false,
|
| 66 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 67 |
+
"unk_token": "<unk>",
|
| 68 |
+
"use_default_system_prompt": false
|
| 69 |
+
}
|
slice_29298/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<image>": 32000,
|
| 3 |
+
"<pad>": 32001
|
| 4 |
+
}
|
slice_29298/chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
|
| 3 |
+
}
|
slice_29298/config.json
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlavaNextForConditionalGeneration"
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"ignore_index": -100,
|
| 9 |
+
"image_grid_pinpoints": [
|
| 10 |
+
[
|
| 11 |
+
336,
|
| 12 |
+
672
|
| 13 |
+
],
|
| 14 |
+
[
|
| 15 |
+
672,
|
| 16 |
+
336
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
672
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
1008,
|
| 24 |
+
336
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
336,
|
| 28 |
+
1008
|
| 29 |
+
]
|
| 30 |
+
],
|
| 31 |
+
"image_seq_length": 576,
|
| 32 |
+
"image_token_index": 32000,
|
| 33 |
+
"model_type": "llava_next",
|
| 34 |
+
"multimodal_projector_bias": true,
|
| 35 |
+
"pad_token_id": 32001,
|
| 36 |
+
"projector_hidden_act": "gelu",
|
| 37 |
+
"text_config": {
|
| 38 |
+
"_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 39 |
+
"architectures": [
|
| 40 |
+
"MistralForCausalLM"
|
| 41 |
+
],
|
| 42 |
+
"attention_dropout": 0.0,
|
| 43 |
+
"head_dim": 128,
|
| 44 |
+
"hidden_act": "silu",
|
| 45 |
+
"hidden_size": 4096,
|
| 46 |
+
"initializer_range": 0.02,
|
| 47 |
+
"intermediate_size": 14336,
|
| 48 |
+
"max_position_embeddings": 32768,
|
| 49 |
+
"model_type": "mistral",
|
| 50 |
+
"num_attention_heads": 32,
|
| 51 |
+
"num_hidden_layers": 32,
|
| 52 |
+
"num_key_value_heads": 8,
|
| 53 |
+
"rms_norm_eps": 1e-05,
|
| 54 |
+
"rope_theta": 1000000.0,
|
| 55 |
+
"sliding_window": null,
|
| 56 |
+
"torch_dtype": "bfloat16",
|
| 57 |
+
"use_cache": true,
|
| 58 |
+
"vocab_size": 32064
|
| 59 |
+
},
|
| 60 |
+
"tie_word_embeddings": false,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.50.0",
|
| 63 |
+
"use_image_newline_parameter": true,
|
| 64 |
+
"vision_config": {
|
| 65 |
+
"attention_dropout": 0.0,
|
| 66 |
+
"hidden_act": "quick_gelu",
|
| 67 |
+
"hidden_size": 1024,
|
| 68 |
+
"image_size": 336,
|
| 69 |
+
"initializer_factor": 1.0,
|
| 70 |
+
"initializer_range": 0.02,
|
| 71 |
+
"intermediate_size": 4096,
|
| 72 |
+
"layer_norm_eps": 1e-05,
|
| 73 |
+
"model_type": "clip_vision_model",
|
| 74 |
+
"num_attention_heads": 16,
|
| 75 |
+
"num_channels": 3,
|
| 76 |
+
"num_hidden_layers": 24,
|
| 77 |
+
"patch_size": 14,
|
| 78 |
+
"projection_dim": 768,
|
| 79 |
+
"torch_dtype": "bfloat16",
|
| 80 |
+
"vocab_size": 32000
|
| 81 |
+
},
|
| 82 |
+
"vision_feature_layer": -2,
|
| 83 |
+
"vision_feature_select_strategy": "default",
|
| 84 |
+
"vocab_size": 32064
|
| 85 |
+
}
|
slice_29298/preprocessor_config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"aspect_ratio_setting": "anyres",
|
| 3 |
+
"crop_size": {
|
| 4 |
+
"height": 336,
|
| 5 |
+
"width": 336
|
| 6 |
+
},
|
| 7 |
+
"do_center_crop": true,
|
| 8 |
+
"do_convert_rgb": true,
|
| 9 |
+
"do_normalize": true,
|
| 10 |
+
"do_pad": true,
|
| 11 |
+
"do_rescale": true,
|
| 12 |
+
"do_resize": true,
|
| 13 |
+
"image_grid_pinpoints": [
|
| 14 |
+
[
|
| 15 |
+
336,
|
| 16 |
+
672
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
336
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
672,
|
| 24 |
+
672
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
1008,
|
| 28 |
+
336
|
| 29 |
+
],
|
| 30 |
+
[
|
| 31 |
+
336,
|
| 32 |
+
1008
|
| 33 |
+
]
|
| 34 |
+
],
|
| 35 |
+
"image_mean": [
|
| 36 |
+
0.48145466,
|
| 37 |
+
0.4578275,
|
| 38 |
+
0.40821073
|
| 39 |
+
],
|
| 40 |
+
"image_processor_type": "LlavaNextImageProcessor",
|
| 41 |
+
"image_std": [
|
| 42 |
+
0.26862954,
|
| 43 |
+
0.26130258,
|
| 44 |
+
0.27577711
|
| 45 |
+
],
|
| 46 |
+
"processor_class": "LlavaNextProcessor",
|
| 47 |
+
"resample": 3,
|
| 48 |
+
"rescale_factor": 0.00392156862745098,
|
| 49 |
+
"size": {
|
| 50 |
+
"shortest_edge": 336
|
| 51 |
+
}
|
| 52 |
+
}
|
slice_29298/processor_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"image_token": "<image>",
|
| 3 |
+
"num_additional_image_tokens": 1,
|
| 4 |
+
"patch_size": 14,
|
| 5 |
+
"processor_class": "LlavaNextProcessor",
|
| 6 |
+
"vision_feature_select_strategy": "default"
|
| 7 |
+
}
|
slice_29298/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:10aecfb74b294ff71cad7db9a38b54042f26f379a4339cb36323dcb353204ef5
|
| 3 |
+
size 15133733934
|
slice_29298/special_tokens_map.json
ADDED
|
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"image_token": "<image>",
|
| 17 |
+
"pad_token": {
|
| 18 |
+
"content": "<pad>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
},
|
| 24 |
+
"unk_token": {
|
| 25 |
+
"content": "<unk>",
|
| 26 |
+
"lstrip": false,
|
| 27 |
+
"normalized": false,
|
| 28 |
+
"rstrip": false,
|
| 29 |
+
"single_word": false
|
| 30 |
+
}
|
| 31 |
+
}
|
slice_29298/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_29298/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
slice_29298/tokenizer_config.json
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": false,
|
| 4 |
+
"add_prefix_space": null,
|
| 5 |
+
"added_tokens_decoder": {
|
| 6 |
+
"0": {
|
| 7 |
+
"content": "<unk>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false,
|
| 12 |
+
"special": true
|
| 13 |
+
},
|
| 14 |
+
"1": {
|
| 15 |
+
"content": "<s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false,
|
| 20 |
+
"special": true
|
| 21 |
+
},
|
| 22 |
+
"2": {
|
| 23 |
+
"content": "</s>",
|
| 24 |
+
"lstrip": false,
|
| 25 |
+
"normalized": false,
|
| 26 |
+
"rstrip": false,
|
| 27 |
+
"single_word": false,
|
| 28 |
+
"special": true
|
| 29 |
+
},
|
| 30 |
+
"32000": {
|
| 31 |
+
"content": "<image>",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false,
|
| 36 |
+
"special": true
|
| 37 |
+
},
|
| 38 |
+
"32001": {
|
| 39 |
+
"content": "<pad>",
|
| 40 |
+
"lstrip": false,
|
| 41 |
+
"normalized": false,
|
| 42 |
+
"rstrip": false,
|
| 43 |
+
"single_word": false,
|
| 44 |
+
"special": true
|
| 45 |
+
}
|
| 46 |
+
},
|
| 47 |
+
"additional_special_tokens": [],
|
| 48 |
+
"bos_token": "<s>",
|
| 49 |
+
"chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
| 50 |
+
"clean_up_tokenization_spaces": false,
|
| 51 |
+
"eos_token": "</s>",
|
| 52 |
+
"extra_special_tokens": {
|
| 53 |
+
"image_token": "<image>"
|
| 54 |
+
},
|
| 55 |
+
"image_token": "<image>",
|
| 56 |
+
"legacy": true,
|
| 57 |
+
"max_length": null,
|
| 58 |
+
"model_max_length": 4096,
|
| 59 |
+
"pad_to_multiple_of": null,
|
| 60 |
+
"pad_token": "<pad>",
|
| 61 |
+
"pad_token_type_id": 0,
|
| 62 |
+
"padding_side": "right",
|
| 63 |
+
"processor_class": "LlavaNextProcessor",
|
| 64 |
+
"sp_model_kwargs": {},
|
| 65 |
+
"spaces_between_special_tokens": false,
|
| 66 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 67 |
+
"unk_token": "<unk>",
|
| 68 |
+
"use_default_system_prompt": false
|
| 69 |
+
}
|
slice_4883/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<image>": 32000,
|
| 3 |
+
"<pad>": 32001
|
| 4 |
+
}
|
slice_4883/chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
|
| 3 |
+
}
|
slice_4883/config.json
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlavaNextForConditionalGeneration"
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"ignore_index": -100,
|
| 9 |
+
"image_grid_pinpoints": [
|
| 10 |
+
[
|
| 11 |
+
336,
|
| 12 |
+
672
|
| 13 |
+
],
|
| 14 |
+
[
|
| 15 |
+
672,
|
| 16 |
+
336
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
672
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
1008,
|
| 24 |
+
336
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
336,
|
| 28 |
+
1008
|
| 29 |
+
]
|
| 30 |
+
],
|
| 31 |
+
"image_seq_length": 576,
|
| 32 |
+
"image_token_index": 32000,
|
| 33 |
+
"model_type": "llava_next",
|
| 34 |
+
"multimodal_projector_bias": true,
|
| 35 |
+
"pad_token_id": 32001,
|
| 36 |
+
"projector_hidden_act": "gelu",
|
| 37 |
+
"text_config": {
|
| 38 |
+
"_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 39 |
+
"architectures": [
|
| 40 |
+
"MistralForCausalLM"
|
| 41 |
+
],
|
| 42 |
+
"attention_dropout": 0.0,
|
| 43 |
+
"head_dim": 128,
|
| 44 |
+
"hidden_act": "silu",
|
| 45 |
+
"hidden_size": 4096,
|
| 46 |
+
"initializer_range": 0.02,
|
| 47 |
+
"intermediate_size": 14336,
|
| 48 |
+
"max_position_embeddings": 32768,
|
| 49 |
+
"model_type": "mistral",
|
| 50 |
+
"num_attention_heads": 32,
|
| 51 |
+
"num_hidden_layers": 32,
|
| 52 |
+
"num_key_value_heads": 8,
|
| 53 |
+
"rms_norm_eps": 1e-05,
|
| 54 |
+
"rope_theta": 1000000.0,
|
| 55 |
+
"sliding_window": null,
|
| 56 |
+
"torch_dtype": "bfloat16",
|
| 57 |
+
"use_cache": true,
|
| 58 |
+
"vocab_size": 32064
|
| 59 |
+
},
|
| 60 |
+
"tie_word_embeddings": false,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.50.0",
|
| 63 |
+
"use_image_newline_parameter": true,
|
| 64 |
+
"vision_config": {
|
| 65 |
+
"attention_dropout": 0.0,
|
| 66 |
+
"hidden_act": "quick_gelu",
|
| 67 |
+
"hidden_size": 1024,
|
| 68 |
+
"image_size": 336,
|
| 69 |
+
"initializer_factor": 1.0,
|
| 70 |
+
"initializer_range": 0.02,
|
| 71 |
+
"intermediate_size": 4096,
|
| 72 |
+
"layer_norm_eps": 1e-05,
|
| 73 |
+
"model_type": "clip_vision_model",
|
| 74 |
+
"num_attention_heads": 16,
|
| 75 |
+
"num_channels": 3,
|
| 76 |
+
"num_hidden_layers": 24,
|
| 77 |
+
"patch_size": 14,
|
| 78 |
+
"projection_dim": 768,
|
| 79 |
+
"torch_dtype": "bfloat16",
|
| 80 |
+
"vocab_size": 32000
|
| 81 |
+
},
|
| 82 |
+
"vision_feature_layer": -2,
|
| 83 |
+
"vision_feature_select_strategy": "default",
|
| 84 |
+
"vocab_size": 32064
|
| 85 |
+
}
|
slice_4883/preprocessor_config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"aspect_ratio_setting": "anyres",
|
| 3 |
+
"crop_size": {
|
| 4 |
+
"height": 336,
|
| 5 |
+
"width": 336
|
| 6 |
+
},
|
| 7 |
+
"do_center_crop": true,
|
| 8 |
+
"do_convert_rgb": true,
|
| 9 |
+
"do_normalize": true,
|
| 10 |
+
"do_pad": true,
|
| 11 |
+
"do_rescale": true,
|
| 12 |
+
"do_resize": true,
|
| 13 |
+
"image_grid_pinpoints": [
|
| 14 |
+
[
|
| 15 |
+
336,
|
| 16 |
+
672
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
336
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
672,
|
| 24 |
+
672
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
1008,
|
| 28 |
+
336
|
| 29 |
+
],
|
| 30 |
+
[
|
| 31 |
+
336,
|
| 32 |
+
1008
|
| 33 |
+
]
|
| 34 |
+
],
|
| 35 |
+
"image_mean": [
|
| 36 |
+
0.48145466,
|
| 37 |
+
0.4578275,
|
| 38 |
+
0.40821073
|
| 39 |
+
],
|
| 40 |
+
"image_processor_type": "LlavaNextImageProcessor",
|
| 41 |
+
"image_std": [
|
| 42 |
+
0.26862954,
|
| 43 |
+
0.26130258,
|
| 44 |
+
0.27577711
|
| 45 |
+
],
|
| 46 |
+
"processor_class": "LlavaNextProcessor",
|
| 47 |
+
"resample": 3,
|
| 48 |
+
"rescale_factor": 0.00392156862745098,
|
| 49 |
+
"size": {
|
| 50 |
+
"shortest_edge": 336
|
| 51 |
+
}
|
| 52 |
+
}
|
slice_4883/processor_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"image_token": "<image>",
|
| 3 |
+
"num_additional_image_tokens": 1,
|
| 4 |
+
"patch_size": 14,
|
| 5 |
+
"processor_class": "LlavaNextProcessor",
|
| 6 |
+
"vision_feature_select_strategy": "default"
|
| 7 |
+
}
|
slice_4883/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:37c1986000ea6b0eb2357c5c441485335ede15f5532aa9809e5a838f1ba2e70f
|
| 3 |
+
size 15133733934
|