Upload folder using huggingface_hub
Browse files- .gitattributes +1 -0
- arguments.yaml +53 -0
- environ.txt +252 -0
- script.sh +52 -0
- slice_end/17434368370.0/complex.json +0 -0
- slice_end/17434368370.0/conv.json +212 -0
- slice_end/17434368370.0/detail.json +0 -0
- slice_end/Evaluation.log +23 -0
- slice_end/added_tokens.json +4 -0
- slice_end/chat_template.json +3 -0
- slice_end/config.json +85 -0
- slice_end/preprocessor_config.json +52 -0
- slice_end/processor_config.json +7 -0
- slice_end/pytorch_model.bin +3 -0
- slice_end/raw_outputs__aifs4su_yaodong_hantao_align-anything_outputs_llava_1.6_mistral_7b_val_top1-10_slice_end.pkl +3 -0
- slice_end/special_tokens_map.json +31 -0
- slice_end/tokenizer.json +0 -0
- slice_end/tokenizer.model +3 -0
- slice_end/tokenizer_config.json +69 -0
- wandb/debug-internal.log +16 -0
- wandb/debug.log +47 -0
- wandb/run-20250331_173510-7rsczbti/files/output.log +16 -0
- wandb/run-20250331_173510-7rsczbti/files/requirements.txt +258 -0
- wandb/run-20250331_173510-7rsczbti/files/wandb-metadata.json +106 -0
- wandb/run-20250331_173510-7rsczbti/logs/debug-core.log +7 -0
- wandb/run-20250331_173510-7rsczbti/logs/debug-internal.log +7 -0
- wandb/run-20250331_173510-7rsczbti/logs/debug.log +22 -0
- wandb/run-20250331_173510-7rsczbti/run-7rsczbti.wandb +0 -0
- wandb/run-20250331_173726-4v7gx39g/files/output.log +16 -0
- wandb/run-20250331_173726-4v7gx39g/files/requirements.txt +258 -0
- wandb/run-20250331_173726-4v7gx39g/files/wandb-metadata.json +106 -0
- wandb/run-20250331_173726-4v7gx39g/logs/debug-core.log +7 -0
- wandb/run-20250331_173726-4v7gx39g/logs/debug-internal.log +7 -0
- wandb/run-20250331_173726-4v7gx39g/logs/debug.log +22 -0
- wandb/run-20250331_173726-4v7gx39g/run-4v7gx39g.wandb +0 -0
- wandb/run-20250331_173929-oflhglry/files/config.yaml +95 -0
- wandb/run-20250331_173929-oflhglry/files/output.log +58 -0
- wandb/run-20250331_173929-oflhglry/files/requirements.txt +258 -0
- wandb/run-20250331_173929-oflhglry/files/wandb-metadata.json +106 -0
- wandb/run-20250331_173929-oflhglry/files/wandb-summary.json +1 -0
- wandb/run-20250331_173929-oflhglry/logs/debug-core.log +14 -0
- wandb/run-20250331_173929-oflhglry/logs/debug-internal.log +16 -0
- wandb/run-20250331_173929-oflhglry/logs/debug.log +47 -0
- wandb/run-20250331_173929-oflhglry/run-oflhglry.wandb +3 -0
.gitattributes
CHANGED
|
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
wandb/run-20250331_173929-oflhglry/run-oflhglry.wandb filter=lfs diff=lfs merge=lfs -text
|
arguments.yaml
ADDED
|
@@ -0,0 +1,53 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
data_cfgs:
|
| 2 |
+
eval_data_files: {}
|
| 3 |
+
eval_datasets: {}
|
| 4 |
+
eval_optional_args: []
|
| 5 |
+
eval_size: {}
|
| 6 |
+
eval_split: {}
|
| 7 |
+
eval_subset: {}
|
| 8 |
+
eval_template: {}
|
| 9 |
+
train_data_files: {}
|
| 10 |
+
train_datasets: /aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10
|
| 11 |
+
train_name: text-image-to-text
|
| 12 |
+
train_optional_args: []
|
| 13 |
+
train_size: {}
|
| 14 |
+
train_split: train
|
| 15 |
+
train_template: AA_TI2T_LLAVA
|
| 16 |
+
logger_cfgs:
|
| 17 |
+
cache_dir: {}
|
| 18 |
+
log_project: align-anything
|
| 19 |
+
log_run_name: dpo
|
| 20 |
+
log_type: wandb
|
| 21 |
+
output_dir: ../outputs/llava_1.6_mistral_7b_val/top1-10
|
| 22 |
+
save_total_limit: 3
|
| 23 |
+
model_cfgs:
|
| 24 |
+
model_max_length: 4096
|
| 25 |
+
model_name_or_path: /aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf
|
| 26 |
+
trust_remote_code: true
|
| 27 |
+
special_tokens: {}
|
| 28 |
+
train_cfgs:
|
| 29 |
+
adam_betas:
|
| 30 |
+
- 0.9
|
| 31 |
+
- 0.95
|
| 32 |
+
bf16: true
|
| 33 |
+
ds_cfgs: ds_z3_config.json
|
| 34 |
+
epochs: 3
|
| 35 |
+
eval_interval: 10
|
| 36 |
+
eval_strategy: epoch
|
| 37 |
+
fp16: false
|
| 38 |
+
freeze_language_model: false
|
| 39 |
+
freeze_mm_proj: false
|
| 40 |
+
freeze_vision_tower: true
|
| 41 |
+
gradient_accumulation_steps: 1
|
| 42 |
+
gradient_checkpointing: true
|
| 43 |
+
learning_rate: 1.0e-06
|
| 44 |
+
load_checkpoint: false
|
| 45 |
+
lr_scheduler_type: cosine
|
| 46 |
+
lr_warmup_ratio: 0.03
|
| 47 |
+
per_device_eval_batch_size: 1
|
| 48 |
+
per_device_train_batch_size: 1
|
| 49 |
+
regularization: 0.001
|
| 50 |
+
save_checkpoint: false
|
| 51 |
+
scale_coeff: 0.1
|
| 52 |
+
seed: 42
|
| 53 |
+
weight_decay: 0.0
|
environ.txt
ADDED
|
@@ -0,0 +1,252 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
ADDR2LINE=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-addr2line
|
| 2 |
+
AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ar
|
| 3 |
+
AS=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-as
|
| 4 |
+
BASH_FUNC__module_raw%%=() { unset _mlshdbg;
|
| 5 |
+
if [ "${MODULES_SILENT_SHELL_DEBUG:-0}" = '1' ]; then
|
| 6 |
+
case "$-" in
|
| 7 |
+
*v*x*)
|
| 8 |
+
set +vx;
|
| 9 |
+
_mlshdbg='vx'
|
| 10 |
+
;;
|
| 11 |
+
*v*)
|
| 12 |
+
set +v;
|
| 13 |
+
_mlshdbg='v'
|
| 14 |
+
;;
|
| 15 |
+
*x*)
|
| 16 |
+
set +x;
|
| 17 |
+
_mlshdbg='x'
|
| 18 |
+
;;
|
| 19 |
+
*)
|
| 20 |
+
_mlshdbg=''
|
| 21 |
+
;;
|
| 22 |
+
esac;
|
| 23 |
+
fi;
|
| 24 |
+
unset _mlre _mlIFS;
|
| 25 |
+
if [ -n "${IFS+x}" ]; then
|
| 26 |
+
_mlIFS=$IFS;
|
| 27 |
+
fi;
|
| 28 |
+
IFS=' ';
|
| 29 |
+
for _mlv in ${MODULES_RUN_QUARANTINE:-};
|
| 30 |
+
do
|
| 31 |
+
if [ "${_mlv}" = "${_mlv##*[!A-Za-z0-9_]}" -a "${_mlv}" = "${_mlv#[0-9]}" ]; then
|
| 32 |
+
if [ -n "`eval 'echo ${'$_mlv'+x}'`" ]; then
|
| 33 |
+
_mlre="${_mlre:-}${_mlv}_modquar='`eval 'echo ${'$_mlv'}'`' ";
|
| 34 |
+
fi;
|
| 35 |
+
_mlrv="MODULES_RUNENV_${_mlv}";
|
| 36 |
+
_mlre="${_mlre:-}${_mlv}='`eval 'echo ${'$_mlrv':-}'`' ";
|
| 37 |
+
fi;
|
| 38 |
+
done;
|
| 39 |
+
if [ -n "${_mlre:-}" ]; then
|
| 40 |
+
eval `eval ${_mlre} /usr/bin/tclsh /cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl bash '"$@"'`;
|
| 41 |
+
else
|
| 42 |
+
eval `/usr/bin/tclsh /cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl bash "$@"`;
|
| 43 |
+
fi;
|
| 44 |
+
_mlstatus=$?;
|
| 45 |
+
if [ -n "${_mlIFS+x}" ]; then
|
| 46 |
+
IFS=$_mlIFS;
|
| 47 |
+
else
|
| 48 |
+
unset IFS;
|
| 49 |
+
fi;
|
| 50 |
+
unset _mlre _mlv _mlrv _mlIFS;
|
| 51 |
+
if [ -n "${_mlshdbg:-}" ]; then
|
| 52 |
+
set -$_mlshdbg;
|
| 53 |
+
fi;
|
| 54 |
+
unset _mlshdbg;
|
| 55 |
+
return $_mlstatus
|
| 56 |
+
}
|
| 57 |
+
BASH_FUNC_ml%%=() { module ml "$@"
|
| 58 |
+
}
|
| 59 |
+
BASH_FUNC_module%%=() { _module_raw "$@" 2>&1
|
| 60 |
+
}
|
| 61 |
+
BASH_FUNC_switchml%%=() { typeset swfound=1;
|
| 62 |
+
if [ "${MODULES_USE_COMPAT_VERSION:-0}" = '1' ]; then
|
| 63 |
+
typeset swname='main';
|
| 64 |
+
if [ -e /cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl ]; then
|
| 65 |
+
typeset swfound=0;
|
| 66 |
+
unset MODULES_USE_COMPAT_VERSION;
|
| 67 |
+
fi;
|
| 68 |
+
else
|
| 69 |
+
typeset swname='compatibility';
|
| 70 |
+
if [ -e /cm/local/apps/environment-modules/4.5.3/libexec/modulecmd-compat ]; then
|
| 71 |
+
typeset swfound=0;
|
| 72 |
+
MODULES_USE_COMPAT_VERSION=1;
|
| 73 |
+
export MODULES_USE_COMPAT_VERSION;
|
| 74 |
+
fi;
|
| 75 |
+
fi;
|
| 76 |
+
if [ $swfound -eq 0 ]; then
|
| 77 |
+
echo "Switching to Modules $swname version";
|
| 78 |
+
source /cm/local/apps/environment-modules/4.5.3/init/bash;
|
| 79 |
+
else
|
| 80 |
+
echo "Cannot switch to Modules $swname version, command not found";
|
| 81 |
+
return 1;
|
| 82 |
+
fi
|
| 83 |
+
}
|
| 84 |
+
BUILD=x86_64-conda-linux-gnu
|
| 85 |
+
CC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
|
| 86 |
+
CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cc
|
| 87 |
+
CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 88 |
+
CMAKE_ARGS=-DCMAKE_AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ar -DCMAKE_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib -DCMAKE_LINKER=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld -DCMAKE_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip -DCMAKE_BUILD_TYPE=Release
|
| 89 |
+
CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama:/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot/usr
|
| 90 |
+
CMD_WLM_CLUSTER_NAME=slurm
|
| 91 |
+
CONDA_BACKUP_ADDR2LINE=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-addr2line
|
| 92 |
+
CONDA_BACKUP_AR=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ar
|
| 93 |
+
CONDA_BACKUP_AS=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-as
|
| 94 |
+
CONDA_BACKUP_BUILD=x86_64-conda-linux-gnu
|
| 95 |
+
CONDA_BACKUP_CC=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cc
|
| 96 |
+
CONDA_BACKUP_CC_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cc
|
| 97 |
+
CONDA_BACKUP_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 98 |
+
CONDA_BACKUP_CMAKE_PREFIX_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_anything:/aifs4su/yaodong/miniconda3/envs/hantao_anything/x86_64-conda-linux-gnu/sysroot/usr
|
| 99 |
+
CONDA_BACKUP_CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/x86_64-conda-linux-gnu/sysroot
|
| 100 |
+
CONDA_BACKUP_CPP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-cpp
|
| 101 |
+
CONDA_BACKUP_CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 102 |
+
CONDA_BACKUP_CXX=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++
|
| 103 |
+
CONDA_BACKUP_CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++filt
|
| 104 |
+
CONDA_BACKUP_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 105 |
+
CONDA_BACKUP_CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-c++
|
| 106 |
+
CONDA_BACKUP_DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 107 |
+
CONDA_BACKUP_DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 108 |
+
CONDA_BACKUP_DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 109 |
+
CONDA_BACKUP_ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-elfedit
|
| 110 |
+
CONDA_BACKUP_GCC=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc
|
| 111 |
+
CONDA_BACKUP_GCC_AR=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-ar
|
| 112 |
+
CONDA_BACKUP_GCC_NM=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-nm
|
| 113 |
+
CONDA_BACKUP_GCC_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gcc-ranlib
|
| 114 |
+
CONDA_BACKUP_GPROF=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-gprof
|
| 115 |
+
CONDA_BACKUP_GXX=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-g++
|
| 116 |
+
CONDA_BACKUP_HOST=x86_64-conda-linux-gnu
|
| 117 |
+
CONDA_BACKUP_LD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ld
|
| 118 |
+
CONDA_BACKUP_LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib
|
| 119 |
+
CONDA_BACKUP_LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ld.gold
|
| 120 |
+
CONDA_BACKUP_NM=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-nm
|
| 121 |
+
CONDA_BACKUP_OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-objcopy
|
| 122 |
+
CONDA_BACKUP_OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-objdump
|
| 123 |
+
CONDA_BACKUP_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-ranlib
|
| 124 |
+
CONDA_BACKUP_READELF=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-readelf
|
| 125 |
+
CONDA_BACKUP_SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-size
|
| 126 |
+
CONDA_BACKUP_STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-strings
|
| 127 |
+
CONDA_BACKUP_STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_anything/bin/x86_64-conda-linux-gnu-strip
|
| 128 |
+
CONDA_BACKUP__CONDA_PYTHON_SYSCONFIGDATA_NAME=_sysconfigdata_x86_64_conda_cos7_linux_gnu
|
| 129 |
+
CONDA_BACKUP_build_alias=x86_64-conda-linux-gnu
|
| 130 |
+
CONDA_BACKUP_host_alias=x86_64-conda-linux-gnu
|
| 131 |
+
CONDA_BUILD_SYSROOT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/x86_64-conda-linux-gnu/sysroot
|
| 132 |
+
CONDA_DEFAULT_ENV=hantao_llama
|
| 133 |
+
CONDA_EXE=/aifs4su/yaodong/miniconda3/bin/conda
|
| 134 |
+
CONDA_PREFIX=/aifs4su/yaodong/miniconda3/envs/hantao_llama
|
| 135 |
+
CONDA_PREFIX_1=/aifs4su/yaodong/miniconda3
|
| 136 |
+
CONDA_PREFIX_2=/aifs4su/yaodong/miniconda3/envs/hantao_anything
|
| 137 |
+
CONDA_PROMPT_MODIFIER=(hantao_llama)
|
| 138 |
+
CONDA_PYTHON_EXE=/aifs4su/yaodong/miniconda3/bin/python
|
| 139 |
+
CONDA_SHLVL=3
|
| 140 |
+
CPATH=/cm/shared/apps/slurm/current/include
|
| 141 |
+
CPATH_modshare=/cm/shared/apps/slurm/current/include:1
|
| 142 |
+
CPP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-cpp
|
| 143 |
+
CPPFLAGS=-DNDEBUG -D_FORTIFY_SOURCE=2 -O2 -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 144 |
+
CROSS_RANK=0
|
| 145 |
+
CROSS_SIZE=1
|
| 146 |
+
CUDA_MODULE_LOADING=LAZY
|
| 147 |
+
CUDA_VISIBLE_DEVICES=0,1,2,3,4,5,6,7
|
| 148 |
+
CXX=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 149 |
+
CXXFILT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++filt
|
| 150 |
+
CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-strong -fno-plt -O2 -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include -I/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/include -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 151 |
+
CXX_FOR_BUILD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 152 |
+
DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1028/bus
|
| 153 |
+
DEBUG_CFLAGS=-march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 154 |
+
DEBUG_CPPFLAGS=-D_DEBUG -D_FORTIFY_SOURCE=2 -Og -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 155 |
+
DEBUG_CXXFLAGS=-fvisibility-inlines-hidden -std=c++17 -fmessage-length=0 -march=nocona -mtune=haswell -ftree-vectorize -fPIC -fstack-protector-all -fno-plt -Og -g -Wall -Wextra -fvar-tracking-assignments -ffunction-sections -pipe -isystem /aifs4su/yaodong/miniconda3/envs/hantao_anything/include
|
| 156 |
+
ELFEDIT=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-elfedit
|
| 157 |
+
ENABLE_LMOD=0
|
| 158 |
+
GCC=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc
|
| 159 |
+
GCC_AR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc-ar
|
| 160 |
+
GCC_NM=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc-nm
|
| 161 |
+
GCC_RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gcc-ranlib
|
| 162 |
+
GPROF=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-gprof
|
| 163 |
+
GSETTINGS_SCHEMA_DIR=/aifs4su/yaodong/miniconda3/envs/hantao_llama/share/glib-2.0/schemas
|
| 164 |
+
GSETTINGS_SCHEMA_DIR_CONDA_BACKUP=
|
| 165 |
+
GXX=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-g++
|
| 166 |
+
HF_DATASETS_CACHE=/aifs4su/yaodong/.cache/huggingface/datasets
|
| 167 |
+
HF_HOME=/aifs4su/yaodong/.cache/huggingface
|
| 168 |
+
HISTTIMEFORMAT=%y/%m/%d %T
|
| 169 |
+
HOME=/home/yangyaodong
|
| 170 |
+
HOST=x86_64-conda-linux-gnu
|
| 171 |
+
KMP_DUPLICATE_LIB_OK=True
|
| 172 |
+
KMP_INIT_AT_FORK=FALSE
|
| 173 |
+
LANG=C.UTF-8
|
| 174 |
+
LD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld
|
| 175 |
+
LDFLAGS=-Wl,-O2 -Wl,--sort-common -Wl,--as-needed -Wl,-z,relro -Wl,-z,now -Wl,--disable-new-dtags -Wl,--gc-sections -Wl,-rpath,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -Wl,-rpath-link,/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_anything/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib -L/aifs4su/yaodong/miniconda3/envs/hantao_llama/targets/x86_64-linux/lib/stubs
|
| 176 |
+
LD_GOLD=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ld.gold
|
| 177 |
+
LD_LIBRARY_PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/cv2/../../lib64:/usr/mpi/gcc/openmpi-4.1.7a1/lib:/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64
|
| 178 |
+
LD_LIBRARY_PATH_modshare=/cm/shared/apps/slurm/current/lib64:1:/usr/mpi/gcc/openmpi-4.1.7a1/lib:1:/cm/shared/apps/slurm/current/lib64/slurm:1
|
| 179 |
+
LD_RUN_PATH=/usr/mpi/gcc/openmpi-4.1.7a1/lib
|
| 180 |
+
LD_RUN_PATH_modshare=/usr/mpi/gcc/openmpi-4.1.7a1/lib:1
|
| 181 |
+
LESSCLOSE=/usr/bin/lesspipe %s %s
|
| 182 |
+
LESSOPEN=| /usr/bin/lesspipe %s
|
| 183 |
+
LIBRARY_PATH=/cm/shared/apps/slurm/current/lib64/slurm:/cm/shared/apps/slurm/current/lib64
|
| 184 |
+
LIBRARY_PATH_modshare=/cm/shared/apps/slurm/current/lib64:1:/cm/shared/apps/slurm/current/lib64/slurm:1
|
| 185 |
+
LOADEDMODULES=slurm/slurm/23.02.6:gcc/64/4.1.7a1
|
| 186 |
+
LOADEDMODULES_modshare=slurm/slurm/23.02.6:1:gcc/64/4.1.7a1:1
|
| 187 |
+
LOCAL_RANK=0
|
| 188 |
+
LOCAL_SIZE=8
|
| 189 |
+
LOGLEVEL=WARNING
|
| 190 |
+
LOGNAME=yangyaodong
|
| 191 |
+
LS_COLORS=rs=0:di=01;34:ln=01;36:mh=00:pi=40;33:so=01;35:do=01;35:bd=40;33;01:cd=40;33;01:or=40;31;01:mi=00:su=37;41:sg=30;43:ca=30;41:tw=30;42:ow=34;42:st=37;44:ex=01;32:*.tar=01;31:*.tgz=01;31:*.arc=01;31:*.arj=01;31:*.taz=01;31:*.lha=01;31:*.lz4=01;31:*.lzh=01;31:*.lzma=01;31:*.tlz=01;31:*.txz=01;31:*.tzo=01;31:*.t7z=01;31:*.zip=01;31:*.z=01;31:*.dz=01;31:*.gz=01;31:*.lrz=01;31:*.lz=01;31:*.lzo=01;31:*.xz=01;31:*.zst=01;31:*.tzst=01;31:*.bz2=01;31:*.bz=01;31:*.tbz=01;31:*.tbz2=01;31:*.tz=01;31:*.deb=01;31:*.rpm=01;31:*.jar=01;31:*.war=01;31:*.ear=01;31:*.sar=01;31:*.rar=01;31:*.alz=01;31:*.ace=01;31:*.zoo=01;31:*.cpio=01;31:*.7z=01;31:*.rz=01;31:*.cab=01;31:*.wim=01;31:*.swm=01;31:*.dwm=01;31:*.esd=01;31:*.jpg=01;35:*.jpeg=01;35:*.mjpg=01;35:*.mjpeg=01;35:*.gif=01;35:*.bmp=01;35:*.pbm=01;35:*.pgm=01;35:*.ppm=01;35:*.tga=01;35:*.xbm=01;35:*.xpm=01;35:*.tif=01;35:*.tiff=01;35:*.png=01;35:*.svg=01;35:*.svgz=01;35:*.mng=01;35:*.pcx=01;35:*.mov=01;35:*.mpg=01;35:*.mpeg=01;35:*.m2v=01;35:*.mkv=01;35:*.webm=01;35:*.webp=01;35:*.ogm=01;35:*.mp4=01;35:*.m4v=01;35:*.mp4v=01;35:*.vob=01;35:*.qt=01;35:*.nuv=01;35:*.wmv=01;35:*.asf=01;35:*.rm=01;35:*.rmvb=01;35:*.flc=01;35:*.avi=01;35:*.fli=01;35:*.flv=01;35:*.gl=01;35:*.dl=01;35:*.xcf=01;35:*.xwd=01;35:*.yuv=01;35:*.cgm=01;35:*.emf=01;35:*.ogv=01;35:*.ogx=01;35:*.aac=00;36:*.au=00;36:*.flac=00;36:*.m4a=00;36:*.mid=00;36:*.midi=00;36:*.mka=00;36:*.mp3=00;36:*.mpc=00;36:*.ogg=00;36:*.ra=00;36:*.wav=00;36:*.oga=00;36:*.opus=00;36:*.spx=00;36:*.xspf=00;36:
|
| 192 |
+
MANPATH=/usr/mpi/gcc/openmpi-4.1.7a1/share/man:/cm/shared/apps/slurm/current/man:/cm/local/apps/environment-modules/4.5.3/share/man:/usr/local/man:/usr/local/share/man:/usr/share/man:/cm/local/apps/environment-modules/current/share/man:/cm/local/apps/environment-modules/current/share/man
|
| 193 |
+
MANPATH_modshare=/usr/local/share/man:1:/usr/mpi/gcc/openmpi-4.1.7a1/share/man:1:/cm/local/apps/environment-modules/current/share/man:1:/cm/local/apps/environment-modules/4.5.3/share/man:1:/usr/local/man:1:/usr/share/man:1:/cm/shared/apps/slurm/current/man:1
|
| 194 |
+
MASTER_ADDR=127.0.0.1
|
| 195 |
+
MASTER_PORT=65009
|
| 196 |
+
MIG_PARTED_CHECKPOINT_FILE=/var/lib/nvidia-mig-manager/checkpoint.json
|
| 197 |
+
MIG_PARTED_CONFIG_FILE=/etc/nvidia-mig-manager/config.yaml
|
| 198 |
+
MIG_PARTED_HOOKS_FILE=/etc/nvidia-mig-manager/hooks.yaml
|
| 199 |
+
MODULEPATH=/cm/local/modulefiles:/cm/shared/modulefiles
|
| 200 |
+
MODULESHOME=/cm/local/apps/environment-modules/4.5.3
|
| 201 |
+
MODULES_CMD=/cm/local/apps/environment-modules/4.5.3/libexec/modulecmd.tcl
|
| 202 |
+
MODULES_SET_SHELL_STARTUP=0
|
| 203 |
+
MOTD_SHOWN=pam
|
| 204 |
+
MPI_HOME=/usr/mpi/gcc/openmpi-4.1.7a1
|
| 205 |
+
MPI_RUN=/usr/mpi/gcc/openmpi-4.1.7a1/bin/mpirun
|
| 206 |
+
NM=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-nm
|
| 207 |
+
NVCC_PREPEND_FLAGS= -ccbin=/aifs4su/yaodong/miniconda3/bin/x86_64-conda-linux-gnu-c++ -ccbin=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-c++
|
| 208 |
+
NVCC_PREPEND_FLAGS_BACKUP= -ccbin=/aifs4su/yaodong/miniconda3/bin/x86_64-conda-linux-gnu-c++
|
| 209 |
+
NVITOP_MONITOR_MODE=colorful
|
| 210 |
+
OBJCOPY=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objcopy
|
| 211 |
+
OBJDUMP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-objdump
|
| 212 |
+
OLDPWD=/aifs4su/yaodong/hantao/align-anything/scripts/llava
|
| 213 |
+
PATH=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin:/aifs4su/yaodong/miniconda3/condabin:/usr/mpi/gcc/openmpi-4.1.7a1/bin:/usr/lpp/mmfs/bin:/usr/local/cuda/bin:/opt/bin:/usr/lpp/mmfs/bin:/cm/shared/apps/slurm/current/sbin:/cm/shared/apps/slurm/current/bin:/usr/local/cuda/bin:/opt/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/sbin:/usr/sbin:/cm/local/apps/environment-modules/4.5.3/bin
|
| 214 |
+
PATH_modshare=/usr/mpi/gcc/openmpi-4.1.7a1/bin:1:/opt/bin/:1:/usr/bin:1:/usr/local/bin:1:/cm/shared/apps/slurm/current/bin:1:/cm/shared/apps/slurm/current/sbin:1:/bin:1:/snap/bin:1:/sbin:1:/usr/sbin:1:/cm/local/apps/environment-modules/4.5.3/bin:1:/usr/games:1:/usr/local/sbin:1:/usr/lpp/mmfs/bin:1:/usr/local/cuda/bin:1:/usr/local/games:1
|
| 215 |
+
PWD=/aifs4su/yaodong/hantao/align-anything/scripts
|
| 216 |
+
PYTHONHASHSEED=42
|
| 217 |
+
PYTHONPATH=/aifs4su/yaodong/hantao/align-anything/scripts
|
| 218 |
+
RANK=0
|
| 219 |
+
RANLIB=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-ranlib
|
| 220 |
+
READELF=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-readelf
|
| 221 |
+
SHELL=/bin/bash
|
| 222 |
+
SHLVL=3
|
| 223 |
+
SIZE=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-size
|
| 224 |
+
SLURM_CONF=/cm/shared/apps/slurm/var/etc/slurm/slurm.conf
|
| 225 |
+
SSH_CLIENT=10.33.4.229 44942 22
|
| 226 |
+
SSH_CONNECTION=10.33.4.51 48576 10.33.4.230 22
|
| 227 |
+
SSH_TTY=/dev/pts/2
|
| 228 |
+
STRINGS=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strings
|
| 229 |
+
STRIP=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/x86_64-conda-linux-gnu-strip
|
| 230 |
+
TERM=screen
|
| 231 |
+
TERM_PROGRAM=tmux
|
| 232 |
+
TERM_PROGRAM_VERSION=3.2a
|
| 233 |
+
TMUX=/tmp/tmux-1028/default,2764504,18
|
| 234 |
+
TMUX_PANE=%18
|
| 235 |
+
USER=yangyaodong
|
| 236 |
+
WANDB_API_KEY=7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33
|
| 237 |
+
WANDB_SERVICE=2-1814399-tcp-localhost-39655
|
| 238 |
+
WORLD_SIZE=8
|
| 239 |
+
XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop
|
| 240 |
+
XDG_RUNTIME_DIR=/run/user/1028
|
| 241 |
+
XDG_SESSION_CLASS=user
|
| 242 |
+
XDG_SESSION_ID=63754
|
| 243 |
+
XDG_SESSION_TYPE=tty
|
| 244 |
+
ZERO_STAGE=3
|
| 245 |
+
_=/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/deepspeed
|
| 246 |
+
_CE_CONDA=
|
| 247 |
+
_CE_M=
|
| 248 |
+
_CONDA_PYTHON_SYSCONFIGDATA_NAME=_sysconfigdata_x86_64_conda_cos7_linux_gnu
|
| 249 |
+
_LMFILES_=/cm/local/modulefiles/slurm/slurm/23.02.6:/cm/local/modulefiles/gcc/64/4.1.7a1
|
| 250 |
+
_LMFILES__modshare=/cm/local/modulefiles/slurm/slurm/23.02.6:1:/cm/local/modulefiles/gcc/64/4.1.7a1:1
|
| 251 |
+
build_alias=x86_64-conda-linux-gnu
|
| 252 |
+
host_alias=x86_64-conda-linux-gnu
|
script.sh
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
#!/usr/bin/env bash
|
| 2 |
+
#
|
| 3 |
+
# Copyright 2025 PKU-Alignment Team. All Rights Reserved.
|
| 4 |
+
#
|
| 5 |
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
| 6 |
+
# you may not use this file except in compliance with the License.
|
| 7 |
+
# You may obtain a copy of the License at
|
| 8 |
+
#
|
| 9 |
+
# http://www.apache.org/licenses/LICENSE-2.0
|
| 10 |
+
#
|
| 11 |
+
# Unless required by applicable law or agreed to in writing, software
|
| 12 |
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
| 13 |
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
| 14 |
+
# See the License for the specific language governing permissions and
|
| 15 |
+
# limitations under the License.
|
| 16 |
+
# ==============================================================================
|
| 17 |
+
|
| 18 |
+
DATASETS_NAME=("top1-10" "top1-20" "top1-30" "top1-40" "top1-50" "top1-60" "top1-70" "top1-80" "top1-90" "top1-100")
|
| 19 |
+
# DATASETS_NAME=("top1-10")
|
| 20 |
+
#
|
| 21 |
+
MODEL_NAME_OR_PATH="/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf" # model path
|
| 22 |
+
|
| 23 |
+
# HOSTFILE="/aifs4su/yaodong/hantao/align-anything/scripts/.hostfile"
|
| 24 |
+
|
| 25 |
+
for DATASET_NAME in ${DATASETS_NAME[@]}; do
|
| 26 |
+
TRAIN_DATASETS="/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/${DATASET_NAME}" # dataset path
|
| 27 |
+
TRAIN_TEMPLATE="AA_TI2T_LLAVA" # dataset template
|
| 28 |
+
TRAIN_NAME="text-image-to-text" # dataset name
|
| 29 |
+
TRAIN_SPLIT="train" # split the dataset
|
| 30 |
+
|
| 31 |
+
OUTPUT_DIR="../outputs/llava_1.6_mistral_7b_val/${DATASET_NAME}" # output dir
|
| 32 |
+
|
| 33 |
+
# For wandb online logging
|
| 34 |
+
export WANDB_API_KEY="7e2dcc0c310ebcb7cdcafd5e9320d6be55cf1a33"
|
| 35 |
+
|
| 36 |
+
# Source the setup script
|
| 37 |
+
source ./setup.sh
|
| 38 |
+
|
| 39 |
+
# Execute deepspeed command
|
| 40 |
+
deepspeed \
|
| 41 |
+
--master_port ${MASTER_PORT} \
|
| 42 |
+
--module align_anything.trainers.text_image_to_text.dpo \
|
| 43 |
+
--model_name_or_path ${MODEL_NAME_OR_PATH} \
|
| 44 |
+
--train_datasets ${TRAIN_DATASETS} \
|
| 45 |
+
--train_template ${TRAIN_TEMPLATE} \
|
| 46 |
+
--train_split ${TRAIN_SPLIT} \
|
| 47 |
+
--train_name ${TRAIN_NAME} \
|
| 48 |
+
--output_dir ${OUTPUT_DIR} \
|
| 49 |
+
--save_total_limit 3 \
|
| 50 |
+
--train_batch_size 8 \
|
| 51 |
+
--epochs 3
|
| 52 |
+
done
|
slice_end/17434368370.0/complex.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_end/17434368370.0/conv.json
ADDED
|
@@ -0,0 +1,212 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
[
|
| 2 |
+
{
|
| 3 |
+
"question": "What is the position of the skateboard in the image?",
|
| 4 |
+
"correct_answer": "The skateboard in the image is in an upside-down position, with its wheels pointing up and laying on the ground.",
|
| 5 |
+
"response": "1. The skateboard is positioned vertically, with its wheels off the ground and its trucks (the metal axles that hold the wheels) resting on a surface.\n2. The skateboard is tilted to the side, with its front wheels raised higher than its rear wheels.\n3. The skateboard is leaning against a wall or ledge, which is acting as a support.\n4. The skateboarder is standing on top of the skateboard, with their feet positioned on the trucks.\n5. The skateboarder appears to be in a relaxed stance, possibly preparing to perform a trick or simply enjoying the moment.\n\nPlease note that this description is based on the visible elements in the image and does not include any speculative or imaginary content. ",
|
| 6 |
+
"score": false,
|
| 7 |
+
"gpt_response": "4 8\n\nAssistant 1's response, while imaginative and detailed, does not accurately address the specific question about the position of the skateboard in the image. The user's question implies a need for a straightforward description of the skateboard's orientation based on the context provided (\"a person with an upside down skateboard\"). Assistant 1's answer, however, offers multiple speculative scenarios that do not directly confirm or describe the skateboard's upside-down position. This response introduces unnecessary complexity and ambiguity, making it less helpful and relevant to the user's query. Therefore, it receives a lower score of 4 for not directly answering the question and for adding speculative elements that were not grounded in the provided context.\n\nAssistant 2's response directly addresses the user's question by stating that the skateboard is in an upside-down position with its wheels pointing up and laying on the ground. This answer is concise, relevant, and accurate, given the context provided by the user. It does not include unnecessary details or speculative content, making it more useful for someone seeking a straightforward description of the skateboard's position in the image. Therefore, Assistant 2 receives a higher score of 8 for its directness, relevance, and accuracy in answering the question."
|
| 8 |
+
},
|
| 9 |
+
{
|
| 10 |
+
"question": "What is the main focus of the image?",
|
| 11 |
+
"correct_answer": "The main focus of the image is a car parked on the side of the road, next to a parking meter.",
|
| 12 |
+
"response": "",
|
| 13 |
+
"score": false,
|
| 14 |
+
"gpt_response": "1 8\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on the criteria of helpfulness, relevance, accuracy, and level of detail. Therefore, it receives the lowest possible score of 1.\n\nAssistant 2, on the other hand, provided a direct answer to the question by stating that the main focus of the image is a car parked on the side of the road, next to a parking meter. This response is helpful as it directly addresses the user's query, is relevant to the context provided, and includes a level of detail (mention of the parking meter) that was not explicitly asked for but adds to the understanding of the scene. The accuracy of the response cannot be directly verified without the image or the descriptive sentences and bounding box coordinates mentioned in the system's note. However, based on the information given, Assistant 2's response seems plausible and informative. Therefore, it receives a score of 8, reflecting its relative performance in providing a useful and detailed answer within the constraints of the information provided."
|
| 15 |
+
},
|
| 16 |
+
{
|
| 17 |
+
"question": "What is the pattern on the little girl's umbrella and raincoat?",
|
| 18 |
+
"correct_answer": "The little girl's umbrella and raincoat both have a ladybug pattern on them.",
|
| 19 |
+
"response": "¡Hola! The little girl's umbrella and raincoat have a pattern featuring black spots with red outlines, reminiscent of ladybugs. It's a cute and playful design that matches well with the rainy weather. The pattern is consistent across both items, creating a coordinated look. ",
|
| 20 |
+
"score": true,
|
| 21 |
+
"gpt_response": "8 7\n\nAssistant 1 provided a more detailed and descriptive answer, elaborating on the pattern of the little girl's umbrella and raincoat by mentioning the black spots with red outlines, which effectively conveys the ladybug theme. This level of detail helps to paint a vivid picture for someone who might not have seen the image, enhancing the helpfulness and relevance of the response. The answer is accurate based on the context provided, and the additional details about the design being cute and playful and matching well with rainy weather add a nice touch, contributing to a higher score.\n\nAssistant 2, while accurate and directly answering the question, offered a much more concise response. It correctly identified the pattern as a ladybug pattern on both the umbrella and raincoat, which is relevant and accurate. However, it lacked the descriptive details that Assistant 1 provided, which could have enriched the response and made it more helpful for someone trying to visualize the scene without seeing the image. The brevity of the response, while efficient, missed an opportunity to engage more deeply with the question by not elaborating on the specifics of the pattern or how it contributes to the overall look of the raincoat and umbrella.\n\nBoth assistants provided correct and relevant information, but Assistant 1's response was more detailed and descriptive, offering a richer answer to the question asked. This is why Assistant 1 scores higher than Assistant 2. However, both responses were accurate and relevant to the question, hence the relatively close scores."
|
| 22 |
+
},
|
| 23 |
+
{
|
| 24 |
+
"question": "What is the cat doing on the red couch?",
|
| 25 |
+
"correct_answer": "The cat is curled up and sleeping on the red couch.",
|
| 26 |
+
"response": "",
|
| 27 |
+
"score": false,
|
| 28 |
+
"gpt_response": "0 10\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2 accurately and concisely answered the user's question by stating that the cat is \"curled up and sleeping on the red couch.\" This response directly addresses the user's query with relevant and accurate information based on the context provided. It is both helpful and precise, offering a clear and straightforward answer without unnecessary details. Thus, Assistant 2 receives a score of 10 for effectively fulfilling the user's request for information."
|
| 29 |
+
},
|
| 30 |
+
{
|
| 31 |
+
"question": "What is the main focus of the image?",
|
| 32 |
+
"correct_answer": "The main focus of the image is the head of a giraffe, shown in a close-up headshot.",
|
| 33 |
+
"response": "",
|
| 34 |
+
"score": false,
|
| 35 |
+
"gpt_response": "0 7\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, on the other hand, offered a specific answer by stating that the main focus of the image is the head of a giraffe in a close-up headshot. This response is helpful as it directly addresses the user's question about the main focus of the image. It is relevant because it specifies that the image centers on the giraffe's head, which is likely the most significant detail if the giraffe is chewing on leaves. The accuracy of the response cannot be directly verified without the image or the descriptive sentences and bounding box coordinates mentioned in the system's note. However, assuming the description of the giraffe chewing on leaves is accurate, focusing on the head is a reasonable interpretation. The level of detail is minimal but sufficient for the question asked. Therefore, Assistant 2 receives a score of 7 for providing a relevant and potentially accurate response with a reasonable level of detail given the context."
|
| 36 |
+
},
|
| 37 |
+
{
|
| 38 |
+
"question": "Where is the cart with luggage bags located?",
|
| 39 |
+
"correct_answer": "The cart with luggage bags is located in a lobby.",
|
| 40 |
+
"response": "1. The cart with luggage bags is located inside a building, likely a hotel lobby or a transportation hub like an airport or train station.\n2. It's positioned on a tiled floor, which suggests an indoor setting with a polished floor.\n3. The luggage bags are hanging from a metal frame, which is mounted on wheels, allowing for easy movement.\n4. There are various types of luggage, including suitcases, duffel bags, and backpacks, indicating that these belongings are likely those of travelers.\n5. The presence of other people in the background suggests that this is a public space where people are either arriving or departing.\n\nPlease note that without additional context or landmarks, it's not possible to determine the exact location of this cart. However, the setting is consistent with common areas where luggage carts are used for transporting personal belongings. ",
|
| 41 |
+
"score": true,
|
| 42 |
+
"gpt_response": "8 6\n\nAssistant 1 provided a detailed and imaginative response based on the hypothetical visual cues described in the context. The answer included several observations that could be inferred from an image of a cart filled with luggage, such as the location being indoors, likely a hotel lobby or transportation hub, and the presence of a tiled floor suggesting an indoor setting. The mention of various types of luggage and the presence of other people in the background further enriches the description, offering a vivid picture that aligns well with common scenarios where luggage carts are found. This level of detail and the effort to contextualize the information based on the visual cues given make the response helpful, relevant, and accurate within the hypothetical scenario. However, the response might have slightly overstepped by assuming specifics not directly provided in the question, such as the exact type of indoor setting and the floor's material, which is why it does not receive a perfect score.\n\nAssistant 2 provided a much more concise response, stating simply that the cart with luggage bags is located in a lobby. This answer is straightforward and likely accurate given the context of the question, which implies a setting where luggage carts are commonly found. However, the response lacks the detail and contextual analysis provided by Assistant 1, making it less informative and engaging. It does not leverage the visual cues that could have been inferred from the question, resulting in a less helpful and less immersive answer. Nonetheless, it directly addresses the question without making unfounded assumptions, which maintains its relevance and accuracy, justifying a slightly above-average score."
|
| 43 |
+
},
|
| 44 |
+
{
|
| 45 |
+
"question": "What objects are on the table in the image?",
|
| 46 |
+
"correct_answer": "The objects on the table in the image include a folded newspaper, sunglasses, books, an apple, and a teddy bear.",
|
| 47 |
+
"response": "1. A newspaper\n2. A pair of sunglasses\n3. A red apple\n4. A tie with a blue geometric pattern\n5. A book titled \"Federal Grants and Cooperative Agreements\"\n6. A book titled \"Administrative Law\"\n7. A book titled \"Administrative Practice\"\n8. A book titled \"Administrative Law and Practice\"\n9. A stuffed teddy bear wearing a baseball uniform\n\nThe teddy bear is positioned behind the books, and the apple is placed next to the newspaper. The tie is resting on top of the newspaper. ",
|
| 48 |
+
"score": false,
|
| 49 |
+
"gpt_response": "5 7\n\nAssistant 1 provided a detailed list of items supposedly on the table, including a newspaper, a pair of sunglasses, a red apple, a tie with a blue geometric pattern, several books with specific titles, and a stuffed teddy bear wearing a baseball uniform. However, the question context only mentions sunglasses laying on a folded newspaper, implying that the additional items listed by Assistant 1 (apple, tie, books, and teddy bear) are not supported by the given context. This introduces inaccuracies in the response, as it adds details not present in the provided context. Therefore, while the response is detailed, its accuracy is compromised, leading to a lower score.\n\nAssistant 2, on the other hand, provides a more general response that includes items that could be inferred from the context (a folded newspaper, sunglasses, books, an apple, and a teddy bear). Although this response is less detailed than that of Assistant 1, it does not explicitly contradict the provided context but rather offers a broad overview that could potentially align with the items mentioned in the context. However, it still includes items not specified in the context (books, an apple, and a teddy bear), which affects its accuracy. Despite this, Assistant 2's response is somewhat more aligned with the context since it does not delve into specifics that are unsupported, making it slightly more reliable in this scenario.\n\nBoth assistants failed to strictly adhere to the context provided, which only explicitly mentioned sunglasses and a newspaper. However, Assistant 2's approach of providing a broad overview without delving into unsupported specifics makes its response slightly more acceptable in this context, hence the slightly higher score. Both responses, however, could have benefited from a more cautious approach that strictly adhered to the information provided in the context."
|
| 50 |
+
},
|
| 51 |
+
{
|
| 52 |
+
"question": "Is the person riding a motorbike or scooter?",
|
| 53 |
+
"correct_answer": "The person is riding a motorbike, which can also be referred to as a motorcycle or a moped. In this case, the terms motorcycle, scooter, and moped are used interchangeably to describe the same vehicle the woman is riding.",
|
| 54 |
+
"response": "",
|
| 55 |
+
"score": false,
|
| 56 |
+
"gpt_response": "0 4\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, while providing a response, offered incorrect information by stating that the terms motorcycle, scooter, and moped can be used interchangeably to describe the same vehicle. This is inaccurate as motorcycles, scooters, and mopeds have distinct characteristics and are not the same. Motorcycles generally have larger engines and are built for longer distances and higher speeds, scooters have a step-through frame and an engine size typically ranging from 50cc to 250cc, and mopeds are similar to scooters but are designed for even lower power and are often pedal-assisted. The assistant's attempt to answer the question shows an effort to provide relevant information, but the inaccuracy of the content leads to a lower score. Therefore, Assistant 2 receives a score of 4 for attempting to provide a relevant response but failing to accurately distinguish between the different types of two-wheeled vehicles."
|
| 57 |
+
},
|
| 58 |
+
{
|
| 59 |
+
"question": "How many airplanes are visible in the image?",
|
| 60 |
+
"correct_answer": "There are two large passenger airplanes visible in the image.",
|
| 61 |
+
"response": "2 ",
|
| 62 |
+
"score": false,
|
| 63 |
+
"gpt_response": "7 8\n\nBoth Assistant 1 and Assistant 2 accurately answered the user's question, identifying that there are two airplanes visible in the image. However, Assistant 2 provided a slightly more detailed response by specifying that the airplanes are \"large passenger airplanes,\" which adds context and could be more helpful to the user depending on their intent or interest in asking. This additional detail about the type of airplanes enriches the response without assuming too much beyond the visible information, thus making it slightly more informative.\n\nAssistant 1's response was straightforward and directly answered the question, which is why it still receives a relatively high score. It is clear, concise, and correct, which are important qualities in a response. However, it lacks the additional context that Assistant 2 provided, which could enhance the user's understanding or satisfaction with the answer.\n\nThe scores reflect the accuracy and relevance of both responses to the user's question, with a slight preference for the additional detail provided by Assistant 2. Both assistants performed well in terms of accuracy, but the level of detail and the potential for added helpfulness in Assistant 2's response gives it a slight edge in the overall evaluation."
|
| 64 |
+
},
|
| 65 |
+
{
|
| 66 |
+
"question": "What is the man doing in the image?",
|
| 67 |
+
"correct_answer": "In the image, the man is holding up clothes on hangers, possibly in some hanger bags, and appears to be preparing for travel, as a suitcase is also in front of him.",
|
| 68 |
+
"response": "",
|
| 69 |
+
"score": false,
|
| 70 |
+
"gpt_response": "0 7\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance on the criteria of helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, on the other hand, offered a plausible interpretation of the scene based on the description provided in the question. The assistant inferred that the man is preparing for travel due to the presence of clothes on hangers (possibly in hanger bags) and a suitcase. This response is helpful as it provides a context for the man's actions, relevant to the question asked, and accurate based on the information given. The level of detail is appropriate for the amount of information available from the question, making assumptions that are logical and adding to the user's understanding of the scene. Therefore, Assistant 2 receives a score of 7. The deduction from a perfect score is due to the fact that without visual confirmation, we cannot be entirely sure of the accuracy of the details provided, such as the presence of a suitcase, which was inferred rather than directly observed from the question."
|
| 71 |
+
},
|
| 72 |
+
{
|
| 73 |
+
"question": "What is the man doing in the image?",
|
| 74 |
+
"correct_answer": "The man in the image is using his cell phone and giving a thumbs up, possibly posing for a picture.",
|
| 75 |
+
"response": "",
|
| 76 |
+
"score": false,
|
| 77 |
+
"gpt_response": "0 7\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance on the criteria of helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, on the other hand, offered a concise answer that directly addressed the user's question based on the provided context. The response was relevant and accurate, given the description of the man's actions in the image. Although the answer could have included more details or speculated on the reason behind the man's actions for a higher score in terms of detail, it still provided the essential information requested by the user. Therefore, Assistant 2 receives a score of 7 for its performance, reflecting its ability to provide a helpful and relevant answer, albeit with room for more detail."
|
| 78 |
+
},
|
| 79 |
+
{
|
| 80 |
+
"question": "How many giraffes are depicted in the image?",
|
| 81 |
+
"correct_answer": "There are four giraffes in the image.",
|
| 82 |
+
"response": "3 giraffes are depicted in the image. ",
|
| 83 |
+
"score": true,
|
| 84 |
+
"gpt_response": "0 0\n\nWithout access to the actual image or the descriptive sentences and bounding box coordinates mentioned in the system's note, it is impossible to accurately evaluate the responses of Assistant 1 and Assistant 2. Both assistants provide different answers to the question of how many giraffes are depicted in the image, with Assistant 1 stating there are 3 giraffes and Assistant 2 stating there are 4 giraffes. Without the visual content or a detailed description of the image, there is no way to determine which assistant, if either, provided the correct answer. Therefore, both assistants receive a score of 0, as there is insufficient information to assess the helpfulness, relevance, accuracy, or level of detail in their responses. This evaluation strictly adheres to the information provided and does not assume any external knowledge or bias towards either assistant's response."
|
| 85 |
+
},
|
| 86 |
+
{
|
| 87 |
+
"question": "What type of bird is pictured in the image?",
|
| 88 |
+
"correct_answer": "A white duck is pictured in the image.",
|
| 89 |
+
"response": "¡Hola! The bird in the image is a goose, specifically a white goose with an orange bill. It appears to be a domestic variety, possibly a breed such as a Toulouse-Lautrec goose, which is known for its distinctive coloration. These geese are often kept as pets or for ornamental purposes due to their striking appearance. ",
|
| 90 |
+
"score": true,
|
| 91 |
+
"gpt_response": "8 4\n\nAssistant 1 provided a detailed and relevant response by identifying the bird in the image as a goose, specifically mentioning its color and bill characteristics, which aligns with the context provided. The mention of a possible breed, the Toulouse-Lautrec goose, and the reasons for keeping such geese (as pets or for ornamental purposes) adds valuable information for the user, enhancing the helpfulness and level of detail in the response. However, without visual confirmation, the specific breed identification is speculative, which slightly affects the accuracy score.\n\nAssistant 2, on the other hand, incorrectly identifies the bird as a white duck. This response is less accurate given the context provided, which clearly states the bird is a goose. The lack of additional information or detail about the duck, such as breed or characteristics, makes the response less helpful and relevant compared to Assistant 1. The response is concise and directly addresses the question but does so inaccurately based on the provided context, leading to a lower overall score.\n\nBoth assistants could potentially improve by acknowledging the limitations of identifying specific breeds without more detailed visual information or by providing a broader range of possibilities with the caveat that exact identification might require more details. However, given the context and the information provided, Assistant 1's response is significantly more aligned with the user's query, hence the higher score. Assistant 2's response, while straightforward, suffers from a critical accuracy issue, impacting its overall helpfulness and relevance."
|
| 92 |
+
},
|
| 93 |
+
{
|
| 94 |
+
"question": "What is the main object featured in the image?",
|
| 95 |
+
"correct_answer": "The main object featured in the image is the clock mounted on a pole or a short tower, standing on the sidewalk next to a building.",
|
| 96 |
+
"response": "",
|
| 97 |
+
"score": false,
|
| 98 |
+
"gpt_response": "0 8\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, on the other hand, gave a clear and direct answer to the question, identifying the main object in the image as the clock mounted on a pole or a short tower. This response is relevant and accurate based on the context provided. It also includes a bit of detail by specifying the location of the clock (on the sidewalk next to a building), which helps in visualizing the scene. However, without additional information about the quality or style of the clock, the level of detail is somewhat limited, but this is understandable given the nature of the question. Thus, Assistant 2 receives a score of 8 for its effective communication and for providing a precise answer that aligns with the information given in the context."
|
| 99 |
+
},
|
| 100 |
+
{
|
| 101 |
+
"question": "What is the main focus of the image?",
|
| 102 |
+
"correct_answer": "The main focus of the image is a cat or kitten sitting on or next to a laptop computer.",
|
| 103 |
+
"response": "",
|
| 104 |
+
"score": false,
|
| 105 |
+
"gpt_response": "0 8\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance on the criteria of helpfulness, relevance, accuracy, and level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, on the other hand, provided a relevant and accurate response to the question based on the context given. The assistant correctly identified the main focus of the image as a cat or kitten sitting on or next to a laptop computer. This response is helpful as it directly addresses the user's query, is relevant to the context provided, and is accurate based on the description of the image. The level of detail is appropriate for the question asked, although it could potentially be improved with more specific information about the kitten's actions or the laptop's appearance if such details were available. Given these considerations, Assistant 2 receives a score of 8 for effectively communicating the main focus of the image in a clear and concise manner."
|
| 106 |
+
},
|
| 107 |
+
{
|
| 108 |
+
"question": "What is the main object in the image?",
|
| 109 |
+
"correct_answer": "The main object in the image is a colorful striped umbrella standing on the beach.",
|
| 110 |
+
"response": "",
|
| 111 |
+
"score": false,
|
| 112 |
+
"gpt_response": "0 8\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0, as it failed to fulfill the user's request for information.\n\nAssistant 2, on the other hand, gave a direct and relevant answer by identifying the \"colorful striped umbrella standing on the beach\" as the main object in the image. This response is helpful as it directly addresses the user's question, is relevant to the context provided, and is accurate based on the description of the image. The level of detail is appropriate for the question asked, identifying the umbrella's key characteristics (colorful and striped) and its location (on the beach). Therefore, Assistant 2 receives a score of 8. The deduction in points is due to the lack of additional information that could have been provided, such as more detailed description of the umbrella or the beach, which would have enhanced the response's richness. However, without access to the actual image and based solely on the provided context, Assistant 2's response is considered highly effective."
|
| 113 |
+
},
|
| 114 |
+
{
|
| 115 |
+
"question": "How many people are involved in taking apart the video game controllers?",
|
| 116 |
+
"correct_answer": "There are two people involved in taking apart the video game controllers.",
|
| 117 |
+
"response": "2 people are involved in taking apart the video game controllers. ",
|
| 118 |
+
"score": true,
|
| 119 |
+
"gpt_response": "10 10\n\nBoth Assistant 1 and Assistant 2 provided responses that are directly relevant, accurate, and concise, answering the user's question without any unnecessary information. The question was straightforward, asking for the number of people involved in taking apart video game controllers. Both assistants answered the question correctly and in a manner that is easy to understand, with Assistant 1 stating \"2 people are involved in taking apart the video game controllers,\" and Assistant 2 providing a nearly identical response, \"There are two people involved in taking apart the video game controllers.\"\n\nGiven the simplicity of the question and the directness required in the answer, both assistants performed optimally by providing the exact information requested without any extraneous details or errors. There was no additional context or complexity in the question that would necessitate a more detailed response or a differentiation in the quality of the answers provided by the two assistants. Therefore, both are rated equally with a score of 10, reflecting their accuracy, relevance, and the helpfulness of their responses in this specific scenario."
|
| 120 |
+
},
|
| 121 |
+
{
|
| 122 |
+
"question": "What is the color of the two suitcases in the image?",
|
| 123 |
+
"correct_answer": "The colors of the two suitcases in the image are black and brown with yellow details.",
|
| 124 |
+
"response": "1. The suitcase on top appears to be a vintage leather suitcase with a yellowish-brown fabric exterior and a dark brown leather trim.\n2. The suitcase on the bottom is a dark brown leather suitcase with a similar style of trim to the one above it.\n\nPlease note that due to the lighting and angle of the photo, the colors may appear slightly different than they actually are. The yellowish-brown fabric could be a faded yellow or a light brown, and the dark brown leather could be a deep brown or even black. The exact colors are difficult to determine with certainty from the image. ",
|
| 125 |
+
"score": true,
|
| 126 |
+
"gpt_response": "8 6\n\nAssistant 1 provided a more detailed and nuanced description of the suitcases, noting the potential for color variation due to lighting and the angle of the photo, which shows a deeper understanding of how images can be perceived differently. This level of detail, including the mention of fabric exterior, leather trim, and the possibility of color misinterpretation, is helpful for someone trying to get a comprehensive understanding of the suitcases' appearance without seeing the image. The caution about color accuracy due to lighting and angle adds a layer of reliability to the description, acknowledging the limitations of visual representation in a digital format.\n\nAssistant 2, while still accurate, offered a more simplified and less detailed description. The mention of \"black and brown with yellow details\" provides a basic understanding but lacks the depth and consideration of potential visual misinterpretation due to photo conditions mentioned by Assistant 1. This response might be easier to digest quickly but doesn't give as thorough an insight into the nuances of the suitcases' appearance.\n\nBoth assistants provided relevant and seemingly accurate descriptions based on the context given, but Assistant 1's response was more detailed and considered, hence the higher score. Assistant 2's response, while still useful, was more general and lacked the depth provided by Assistant 1, resulting in a slightly lower score. The evaluation focuses on the depth of description, consideration of potential visual misinterpretation, and overall usefulness of the information provided to someone who cannot see the image."
|
| 127 |
+
},
|
| 128 |
+
{
|
| 129 |
+
"question": "What are the main objects on the table in the image?",
|
| 130 |
+
"correct_answer": "The main objects on the table in the image are plates, glasses, eating utensils, and possibly a bottle of something.",
|
| 131 |
+
"response": "1. A plate with a fork resting on it.\n2. An empty glass.\n3. A bottle with a label that appears to be \"REBEL\" on it.\n4. A partially eaten dish on another plate.\n5. A spoon.\n6. A knife.\n7. A napkin.\n8. A piece of bread or pastry.\n9. A plate with some food remnants on it.\n10. A glass that looks like it contains a beverage.\n\nThe image has a warm, orange-toned filter applied to it, giving it a cozy or intimate ambiance. The lighting suggests it might be taken in a restaurant or a home setting during the evening or night. ",
|
| 132 |
+
"score": true,
|
| 133 |
+
"gpt_response": "9 6\n\nAssistant 1 provided a detailed and comprehensive list of the objects on the table, which aligns closely with the expected content based on the description of the image. The inclusion of specific details such as the label \"REBEL\" on a bottle and the description of the ambiance through the mention of a warm, orange-toned filter adds depth to the response, making it highly informative and engaging for the reader. This level of detail not only helps in visualizing the scene more vividly but also indicates a thorough examination of the image. The mention of the setting (restaurant or home) and the time of day (evening or night) based on the lighting further enhances the descriptive quality of the response, making it more immersive.\n\nAssistant 2, while accurately identifying the categories of objects present (plates, glasses, eating utensils, and possibly a bottle), provided a much less detailed response. This answer is correct but lacks the specificity and richness of detail that Assistant 1 offered. It gives a broad overview without delving into the specifics of each object, such as the state of the dishes (partially eaten, with remnants), the types of utensils, or the ambiance of the setting. This makes the response less engaging and informative compared to Assistant 1.\n\nBoth assistants provided relevant and accurate information based on the context given. However, Assistant 1's response was significantly more detailed and helpful for someone trying to get a comprehensive understanding of the scene depicted in the image. Assistant 2's response, while accurate, did not fully utilize the opportunity to provide a rich, detailed description, resulting in a lower score. The scores reflect the difference in the level of detail and engagement in their responses."
|
| 134 |
+
},
|
| 135 |
+
{
|
| 136 |
+
"question": "How many doughnuts are in the box?",
|
| 137 |
+
"correct_answer": "There are four doughnuts in the box.",
|
| 138 |
+
"response": "4 ",
|
| 139 |
+
"score": true,
|
| 140 |
+
"gpt_response": "5 5\n\nBoth Assistant 1 and Assistant 2 provided the same exact answer to the question, which was \"4\" or \"There are four doughnuts in the box,\" respectively. Given the context provided, there is no direct information about the accuracy of their responses because the actual number of doughnuts in the box, based on the image described, is not provided in the context. Therefore, evaluating the accuracy of their responses is not possible. \n\nHowever, both responses are equally relevant to the question asked, as they both attempt to answer the question directly. The level of detail is minimal in both responses, with Assistant 2 providing a slightly more complete sentence, which could be seen as marginally more helpful in a real-world context where complete sentences are generally preferred for clarity and professionalism. \n\nNeither assistant provided any information on how they arrived at the number four, such as describing the visual content or referencing the bounding box coordinates, which would have been helpful for understanding the basis of their answer. This lack of explanation and justification for their answers limits the helpfulness and detail of their responses.\n\nOverall, both assistants performed equally in terms of relevance and accuracy (with the assumption that \"4\" is correct), but both lacked in providing a detailed explanation or justification for their answer. The slight difference in the format of the answer does not significantly impact the overall score, hence the equal scoring."
|
| 141 |
+
},
|
| 142 |
+
{
|
| 143 |
+
"question": "What objects are next to the cellphone in the image?",
|
| 144 |
+
"correct_answer": "In the image, the cellphone is next to a pair of earbuds and two jewels.",
|
| 145 |
+
"response": "1. A pair of earbuds with a black case.\n2. A small, clear crystal with a pyramid shape.\n3. A larger, blue crystal with a pyramid shape.\n\nThese objects are placed on a dark surface, possibly a table or a mat. The cellphone appears to be a Samsung model, as indicated by the logo on its screen. The image has a shallow depth of field, with the cellphone in sharp focus and the background blurred. ",
|
| 146 |
+
"score": true,
|
| 147 |
+
"gpt_response": "8 6\n\nAssistant 1 provided a more detailed description of the objects next to the cellphone, including the color, shape, and possible material of the jewels, as well as the color and case of the earbuds. This level of detail could significantly enhance the reader's understanding and visualization of the scene, especially when the reader cannot see the image themselves. Mentioning the brand of the cellphone and the quality of the image (shallow depth of field, sharp focus on the cellphone, blurred background) adds context that might be relevant depending on the user's intent. However, the additional details about the surface and the specific model of the cellphone, while interesting, were not directly related to the question about what objects are next to the cellphone, which slightly detracts from the relevance of the response.\n\nAssistant 2, while less detailed, directly answered the question with a concise list of objects next to the cellphone. This response is straightforward and relevant but lacks the descriptive detail that might be necessary for a user who cannot see the image. The simplicity of the response makes it accessible but might leave the user wanting more information to fully understand the scene.\n\nBoth assistants provided accurate information based on the context given. However, Assistant 1's response was more helpful for someone trying to visualize the scene without seeing the image, hence the higher score. Assistant 2's response, while accurate and relevant, lacked the depth that could make the description more vivid and informative, resulting in a slightly lower score."
|
| 148 |
+
},
|
| 149 |
+
{
|
| 150 |
+
"question": "What breed is the dog in the image?",
|
| 151 |
+
"correct_answer": "The dog in the image is a husky.",
|
| 152 |
+
"response": "",
|
| 153 |
+
"score": true,
|
| 154 |
+
"gpt_response": "0 0\n\nBoth Assistant 1 and Assistant 2 cannot be accurately evaluated based on the information provided because there is a critical piece of context missing: the actual image or a detailed description of the image they were supposed to analyze. Assistant 1 did not provide any response, which automatically results in a score of 0 due to the lack of attempt to answer the question. Assistant 2 claimed the dog in the image is a husky without any visible means to verify this claim against the provided context, as the description of the visual content and bounding box coordinates mentioned in the system's note were not actually provided in the evaluation context. Therefore, there's no way to assess the relevance, accuracy, or level of detail in Assistant 2's response, leading to a score of 0 as well. The evaluation strictly adheres to the information provided, and without the necessary context or content, accurate scoring cannot be justified for either assistant."
|
| 155 |
+
},
|
| 156 |
+
{
|
| 157 |
+
"question": "What type of tree is outside the window with brick siding?",
|
| 158 |
+
"correct_answer": "There is an apple tree outside the window with brick siding.",
|
| 159 |
+
"response": "1. The tree outside the window with brick siding appears to be an apple tree. You can identify it by the characteristic red apple fruits hanging from the branches. The leaves are also typical of apple trees, with a rounded shape and a slightly serrated edge.\n\n2. The brick siding suggests a traditional or rustic architectural style, which is often associated with older buildings or those designed to mimic older architecture.\n\n3. The window itself has shutters, which are common in older buildings and add to the charm of the scene.\n\n4. The overall setting suggests a peaceful, rural, or suburban environment where apple trees are common.\n\n5. The presence of ripe apples indicates that it's likely late summer or early autumn, which is the harvest season for apples.\n\n6. The tree's health and the abundance of fruit suggest that it is well cared for and possibly even organic, as organic farming practices often lead to healthier fruit trees.\n\n7. The tree's location next to the building indicates that it might be part of a small orchard or garden associated with the property.\n\n8. The tree's proximity to the building also suggests that it provides shade and possibly some privacy for the occupants of the building.\n\n9. The combination of the apple tree, brick siding, and shuttered window creates a picturesque scene that evokes a sense of tranquility and connection to nature.\n\n10. The image captures a moment of natural beauty and simplicity, reminding us of the simple pleasures of life. ",
|
| 160 |
+
"score": true,
|
| 161 |
+
"gpt_response": "8 6\n\nAssistant 1 provided a detailed and imaginative response to the question, incorporating various elements that could be inferred from the presence of an apple tree outside a window with brick siding. The assistant's answer goes beyond merely identifying the tree as an apple tree; it elaborates on the architectural style suggested by brick siding, the season indicated by the presence of ripe apples, and the potential care of the tree. It also speculates on the setting (rural or suburban), the health of the tree, and the scene's aesthetic and emotional impact. This level of detail and the effort to connect the tree to broader themes of architecture, seasonality, and well-being are commendable. However, the response might have slightly overstepped by assuming too much about the setting and the care of the tree without explicit evidence, which is why it does not receive a perfect score.\n\nAssistant 2 provided a straightforward and accurate response to the question, identifying the tree as an apple tree. This response is relevant and directly addresses the question without adding unnecessary details or assumptions. However, it lacks the depth and richness of context that might enhance the user's understanding or appreciation of the scene. While brevity can be a strength, in this case, a bit more detail about how to identify an apple tree or why it fits well with brick siding could have made the response more informative and engaging. Therefore, Assistant 2 receives a slightly lower score, reflecting its accuracy and relevance but also its lack of detail and context compared to Assistant 1."
|
| 162 |
+
},
|
| 163 |
+
{
|
| 164 |
+
"question": "What is the man doing in the image?",
|
| 165 |
+
"correct_answer": "The man in the image is walking down the street while talking on his cell phone.",
|
| 166 |
+
"response": "",
|
| 167 |
+
"score": false,
|
| 168 |
+
"gpt_response": "0 5\n\nAssistant 1 did not provide any response to the question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0.\n\nAssistant 2, on the other hand, offered a specific action that the man in the image is performing, which is \"walking down the street while talking on his cell phone.\" This response is partially helpful as it addresses the question about what the man is doing. However, it does not mention the man's hair color, which was a significant detail provided in the context. This omission affects the relevance and accuracy of the response, as it does not fully acknowledge the provided context. Despite this, the assistant did provide some level of detail by including the action of talking on a cell phone, which adds to the scene's description. Therefore, Assistant 2 receives a score of 5 for at least contributing relevant information about the man's activity, even though it missed addressing the detail about the hair color."
|
| 169 |
+
},
|
| 170 |
+
{
|
| 171 |
+
"question": "What type of hats are the man and woman wearing?",
|
| 172 |
+
"correct_answer": "The man and woman are wearing umbrella hats, which are designed to provide protection from the rain.",
|
| 173 |
+
"response": "1. The person on the left appears to be wearing a baseball cap.\n2. The person on the right is wearing a bucket hat.\n\nBucket hats are characterized by their wide brim and rounded crown, which is designed to provide shade and protect the wearer's head from rain or sun. They are typically made from materials like cotton canvas or straw, and they have become a popular style of headwear for their practicality and casual appeal.\n\nBaseball caps, on the other hand, are a type of headwear with a flat crown and a curved bill that extends forward and slightly upward. They are typically made from materials like cotton twill or polyester, and they feature a six-panel design with a stiffened front panel that holds its shape. Baseball caps are often associated with sports teams due to their logo placement, but they are also widely worn as a fashion accessory.\n\nBoth types of hats are commonly worn for sun protection and as a fashion statement. ",
|
| 174 |
+
"score": false,
|
| 175 |
+
"gpt_response": "2 8\n\nAssistant 1 provided a detailed explanation of bucket hats and baseball caps, including their characteristics, materials, and common uses. However, this response is not relevant to the user's question about the type of hats being worn by the man and woman in the image, which was specified as \"umbrella hats.\" The detailed descriptions, while informative about bucket hats and baseball caps, do not address the actual query, leading to a low score for relevance and accuracy in the context of the question asked.\n\nAssistant 2 correctly identified the type of hats as \"umbrella hats,\" which directly answers the user's question. The response is concise and lacks the detailed descriptions found in Assistant 1's answer, but it is accurate and relevant to the question posed. Given the context that the question is about identifying a specific type of hat based on a visual observation, Assistant 2's response is more appropriate and helpful, despite its brevity.\n\nThe scores reflect the relevance and accuracy of each assistant's response to the specific question asked. Assistant 1's detailed but off-target answer receives a lower score due to its lack of relevance to the question. Assistant 2's correct but brief answer is rated higher because it directly addresses the user's query."
|
| 176 |
+
},
|
| 177 |
+
{
|
| 178 |
+
"question": "What are the main items of clothing visible in the image?",
|
| 179 |
+
"correct_answer": "The main items of clothing visible in the image are a white shirt, a red tie, and black shoes.",
|
| 180 |
+
"response": "1. A pair of black shoes\n2. A black suit jacket\n3. A white dress shirt\n4. A red necktie with a patterned design\n\nThese items are laid out on a bed, suggesting they are being prepared for an event or occasion. ",
|
| 181 |
+
"score": false,
|
| 182 |
+
"gpt_response": "6 8\n\nAssistant 1 provides a detailed description of the items, including a black suit jacket and a patterned design on the necktie, which were not mentioned in the user's context. This additional information suggests that Assistant 1 might be inferring or adding details not explicitly stated by the user, potentially leading to inaccuracies if the user did not intend for these items to be included in the description. The mention of the items being laid out on a bed for an event or occasion is speculative and not grounded in the user's provided context, which could mislead or confuse the user if their intention was merely to identify visible items without implying any specific setting or purpose.\n\nAssistant 2 directly addresses the user's question by listing the main items of clothing visible in the image as described in the user's context. This response is concise and directly relevant to the user's query, making no assumptions beyond the information provided by the user. However, Assistant 2's response lacks the detail and descriptive richness that might enhance the user's understanding or appreciation of the image. For instance, mentioning the style or specific features of the items (if visible in the image) could have made the response more informative.\n\nBoth assistants could improve by aligning their responses more closely with the user's context and avoiding assumptions or additions not supported by the user's input. Assistant 1's response, while detailed, risks adding potentially inaccurate or irrelevant information, thus receiving a lower score. Assistant 2's response is more aligned with the user's query but could benefit from additional detail to provide a fuller understanding of the items described, hence receiving a slightly higher score."
|
| 183 |
+
},
|
| 184 |
+
{
|
| 185 |
+
"question": "What stage is the house in regarding its construction?",
|
| 186 |
+
"correct_answer": "The house is in the beginning stages of its construction, with the wooden structure, studs, and staircase partially built but not yet completed.",
|
| 187 |
+
"response": "2021 Update: The image shows a house under construction. It appears to be in the framing stage, where the walls and structural elements are being built. The presence of exposed studs, the absence of drywall, and the unfinished state of the interior suggest that this is a building that is still in the process of being constructed.\n\nThe house features a combination of wooden studs and metal framing, which is common in modern construction. The floor appears to be laid out with plywood sheets, and there are insulation batts visible, indicating that the house is being insulated.\n\nThe presence of a large piece of lumber on the floor suggests that it may be a part of the construction process, possibly for framing or as a temporary support. The overall condition of the house indicates that it is still in the early stages of construction and is not yet ready for occupancy. ",
|
| 188 |
+
"score": true,
|
| 189 |
+
"gpt_response": "8 6\n\nAssistant 1 provided a detailed and comprehensive response, elaborating on the construction stage of the house by mentioning specific elements such as the framing stage, the presence of wooden studs and metal framing, plywood sheets for the floor, and insulation batts. This level of detail not only helps to accurately identify the construction stage but also offers insights into the materials and processes involved, which could be very informative for someone unfamiliar with construction. The mention of a large piece of lumber and its potential use adds to the depth of the analysis. The response is relevant and accurate based on the described image, making it highly helpful for someone trying to understand the construction stage of the house.\n\nAssistant 2, while accurate in identifying the house as being in the beginning stages of construction, provided a much less detailed response. It mentions the wooden structure, studs, and staircase but does not delve into the specifics of the construction process or the materials used, which makes the response less informative compared to Assistant 1. The brevity of the response might be sufficient for someone with a basic query, but it lacks the depth and detail that could provide a fuller understanding of the construction stage.\n\nBoth assistants provided relevant and accurate information based on the described image. However, Assistant 1's response was more detailed and informative, which justifies a higher score. Assistant 2's response, while still accurate and relevant, was less detailed, resulting in a slightly lower score. The scores reflect the level of detail and the amount of information provided by each assistant, which are important factors in assessing their performance in answering the user's question."
|
| 190 |
+
},
|
| 191 |
+
{
|
| 192 |
+
"question": "How many people are visible in the image?",
|
| 193 |
+
"correct_answer": "There are three people visible in the image - two women and one man.",
|
| 194 |
+
"response": "6 people are visible in the image. There are three individuals in the foreground who are smiling and holding wine glasses, and three others in the background who appear to be engaged in conversation. ",
|
| 195 |
+
"score": true,
|
| 196 |
+
"gpt_response": "5 3\n\nAssistant 1 provided a response indicating that six people are visible in the image, detailing that three individuals are in the foreground, smiling and holding wine glasses, and three others are in the background, appearing to be engaged in conversation. This response suggests a detailed observation of the scene, including both the foreground and background activities, which could be very helpful if the description matched the actual content of the image. However, without access to the image or its descriptive sentences and bounding box coordinates, it's impossible to verify the accuracy of Assistant 1's claim. Assuming the description provided by Assistant 1 accurately reflects the content of the image, the response would be highly relevant, accurate, and detailed, meriting a higher score. However, given the lack of verifiable information, a moderate score of 5 is assigned to reflect the potential for high-quality response tempered by the inability to confirm its accuracy.\n\nAssistant 2's response, indicating that three people are visible in the image (two women and one man), provides a straightforward answer but lacks the detail found in Assistant 1's response. This response could be seen as less helpful if indeed there are more individuals present in the image as suggested by Assistant 1. Without additional context or description of the scene beyond the number and gender of the visible individuals, Assistant 2's response is less informative. Given the lack of detail and the potential undercount of individuals present (based on Assistant 1's description), Assistant 2's response is rated lower, with a score of 3. This score reflects the response's potential relevance but acknowledges its limited detail and the possibility that it may not fully capture the scene's complexity.\n\nIn summary, the evaluation considers the potential helpfulness, relevance, accuracy, and level of detail in each response, with the understanding that without access to the image or its descriptive content, the accuracy of the assistants' responses cannot be directly verified. The scores reflect a balance between the perceived quality of the responses based on the information provided and the limitations imposed by the lack of direct access to the image content."
|
| 197 |
+
},
|
| 198 |
+
{
|
| 199 |
+
"question": "What is the color of the elephant in the image?",
|
| 200 |
+
"correct_answer": "The color of the elephant in the image is grey.",
|
| 201 |
+
"response": "",
|
| 202 |
+
"score": false,
|
| 203 |
+
"gpt_response": "0 5\n\nAssistant 1 did not provide any response to the user's question, which makes it impossible to evaluate its performance based on helpfulness, relevance, accuracy, or level of detail. Therefore, it receives a score of 0, as it failed to contribute any information or answer to the user's query.\n\nAssistant 2, on the other hand, provided a direct answer to the user's question by stating that the color of the elephant in the image is grey. This response is helpful and relevant to the user's question, assuming the typical coloration of elephants. However, without access to the image or the descriptive sentences and bounding box coordinates mentioned in the system's note, it's impossible to verify the accuracy of this statement. The response lacks detail, as it does not describe the context of the image, the elephant's surroundings, or any other characteristics that might have been visible. Given these considerations, Assistant 2 receives a score of 5 for providing a plausible answer but without any supporting details or acknowledgment of the limitations in its ability to visually verify the image's content."
|
| 204 |
+
},
|
| 205 |
+
{
|
| 206 |
+
"question": "What kind of objects are included in the art installation?",
|
| 207 |
+
"correct_answer": "The art installation includes a display of large artworks, such as an oversized bench, painted feline heads, and potentially other artistic sculptures or images.",
|
| 208 |
+
"response": "1. A large, colorful sculpture of a tiger's head.\n2. A wooden structure that resembles a bench or a piece of furniture with a weathered appearance.\n3. A large, blue sculpture with a face that appears to be a stylized animal or character.\n4. A small black car with a colorful roof rack.\n5. A person sitting on the edge of the sidewalk, facing away from the camera.\n6. A person standing on the sidewalk, facing towards the camera.\n7. A black car with a white license plate.\n8. A red car with a white license plate.\n9. A person standing on the street, facing away from the camera.\n10. A person standing on the street, facing towards the camera.\n\nThe setting appears to be an outdoor public space, possibly a pedestrian area or a street corner, with a mix of art installations and everyday life. The art installation seems to be a creative blend of organic and man-made elements, possibly commenting on themes of nature, urban life, and the relationship between humans and their environment. The presence of people and vehicles adds a dynamic element to the scene, suggesting a lively atmosphere despite the stillness of the artwork. ",
|
| 209 |
+
"score": true,
|
| 210 |
+
"gpt_response": "7 6\n\nAssistant 1 provided a detailed and imaginative description of the scene, including both the art installations and the surrounding environment, which includes people and vehicles. This response offers a vivid picture of the setting, suggesting a lively atmosphere and a blend of organic and man-made elements. However, the inclusion of people and vehicles as part of the art installation description could be misleading, as these are not typically considered elements of an art installation. This detracts slightly from the accuracy of the response in terms of directly answering the user's question about the objects included in the art installation.\n\nAssistant 2 offered a more concise response, focusing on the art installation itself without the additional context of the surrounding environment. This response is more directly aligned with the user's question, identifying key elements of the art installation. However, it lacks the vivid detail and imaginative description provided by Assistant 1, offering less insight into the nature and atmosphere of the installation. The response could benefit from more specific details about the artworks mentioned to enhance its helpfulness and relevance.\n\nBoth assistants could improve by focusing more precisely on the objects that are unambiguously part of the art installation, excluding elements like people and vehicles unless they are explicitly described as part of the installation. Assistant 1's response, while more detailed and engaging, could be more accurate by focusing on the art objects. Assistant 2's response, though more focused, could provide more detail to enhance its usefulness."
|
| 211 |
+
}
|
| 212 |
+
]
|
slice_end/17434368370.0/detail.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_end/Evaluation.log
ADDED
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-04-01 00:05:26,600 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 2 |
+
2025-04-01 00:05:26,603 - Evaluation - INFO - task: conv
|
| 3 |
+
2025-04-01 00:05:26,603 - Evaluation - INFO - model_id: 0,
|
| 4 |
+
2025-04-01 00:05:26,604 - Evaluation - INFO - average: 62.6,
|
| 5 |
+
2025-04-01 00:05:26,605 - Evaluation - INFO - question: 30,
|
| 6 |
+
2025-04-01 00:05:26,605 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 7 |
+
2025-04-01 00:06:17,335 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 8 |
+
2025-04-01 00:06:17,336 - Evaluation - INFO - task: detail
|
| 9 |
+
2025-04-01 00:06:17,336 - Evaluation - INFO - model_id: 0,
|
| 10 |
+
2025-04-01 00:06:17,337 - Evaluation - INFO - average: 97.3,
|
| 11 |
+
2025-04-01 00:06:17,338 - Evaluation - INFO - question: 30,
|
| 12 |
+
2025-04-01 00:06:17,338 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 13 |
+
2025-04-01 00:06:41,500 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 14 |
+
2025-04-01 00:06:41,501 - Evaluation - INFO - task: complex
|
| 15 |
+
2025-04-01 00:06:41,502 - Evaluation - INFO - model_id: 0,
|
| 16 |
+
2025-04-01 00:06:41,502 - Evaluation - INFO - average: 110.7,
|
| 17 |
+
2025-04-01 00:06:41,503 - Evaluation - INFO - question: 30,
|
| 18 |
+
2025-04-01 00:06:41,503 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 19 |
+
2025-04-01 00:06:41,505 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
| 20 |
+
2025-04-01 00:06:41,505 - Evaluation - INFO - model_id: 0,
|
| 21 |
+
2025-04-01 00:06:41,506 - Evaluation - INFO - total_average: 90.2,
|
| 22 |
+
2025-04-01 00:06:41,507 - Evaluation - INFO - total_question: 90,
|
| 23 |
+
2025-04-01 00:06:41,507 - Evaluation - INFO - +++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
|
slice_end/added_tokens.json
ADDED
|
@@ -0,0 +1,4 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"<image>": 32000,
|
| 3 |
+
"<pad>": 32001
|
| 4 |
+
}
|
slice_end/chat_template.json
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"chat_template": "{% for message in messages %}{% if message['role'] != 'system' %}{{ message['role'].upper() + ': '}}{% endif %}{# Render all images first #}{% for content in message['content'] | selectattr('type', 'equalto', 'image') %}{{ '<image>\n' }}{% endfor %}{# Render all text next #}{% if message['role'] != 'assistant' %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{{ content['text'] + ' '}}{% endfor %}{% else %}{% for content in message['content'] | selectattr('type', 'equalto', 'text') %}{% generation %}{{ content['text'] + ' '}}{% endgeneration %}{% endfor %}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ 'ASSISTANT:' }}{% endif %}"
|
| 3 |
+
}
|
slice_end/config.json
ADDED
|
@@ -0,0 +1,85 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
+
"architectures": [
|
| 4 |
+
"LlavaNextForConditionalGeneration"
|
| 5 |
+
],
|
| 6 |
+
"bos_token_id": 1,
|
| 7 |
+
"eos_token_id": 2,
|
| 8 |
+
"ignore_index": -100,
|
| 9 |
+
"image_grid_pinpoints": [
|
| 10 |
+
[
|
| 11 |
+
336,
|
| 12 |
+
672
|
| 13 |
+
],
|
| 14 |
+
[
|
| 15 |
+
672,
|
| 16 |
+
336
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
672
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
1008,
|
| 24 |
+
336
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
336,
|
| 28 |
+
1008
|
| 29 |
+
]
|
| 30 |
+
],
|
| 31 |
+
"image_seq_length": 576,
|
| 32 |
+
"image_token_index": 32000,
|
| 33 |
+
"model_type": "llava_next",
|
| 34 |
+
"multimodal_projector_bias": true,
|
| 35 |
+
"pad_token_id": 32001,
|
| 36 |
+
"projector_hidden_act": "gelu",
|
| 37 |
+
"text_config": {
|
| 38 |
+
"_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
|
| 39 |
+
"architectures": [
|
| 40 |
+
"MistralForCausalLM"
|
| 41 |
+
],
|
| 42 |
+
"attention_dropout": 0.0,
|
| 43 |
+
"head_dim": 128,
|
| 44 |
+
"hidden_act": "silu",
|
| 45 |
+
"hidden_size": 4096,
|
| 46 |
+
"initializer_range": 0.02,
|
| 47 |
+
"intermediate_size": 14336,
|
| 48 |
+
"max_position_embeddings": 32768,
|
| 49 |
+
"model_type": "mistral",
|
| 50 |
+
"num_attention_heads": 32,
|
| 51 |
+
"num_hidden_layers": 32,
|
| 52 |
+
"num_key_value_heads": 8,
|
| 53 |
+
"rms_norm_eps": 1e-05,
|
| 54 |
+
"rope_theta": 1000000.0,
|
| 55 |
+
"sliding_window": null,
|
| 56 |
+
"torch_dtype": "bfloat16",
|
| 57 |
+
"use_cache": true,
|
| 58 |
+
"vocab_size": 32064
|
| 59 |
+
},
|
| 60 |
+
"tie_word_embeddings": false,
|
| 61 |
+
"torch_dtype": "bfloat16",
|
| 62 |
+
"transformers_version": "4.50.0",
|
| 63 |
+
"use_image_newline_parameter": true,
|
| 64 |
+
"vision_config": {
|
| 65 |
+
"attention_dropout": 0.0,
|
| 66 |
+
"hidden_act": "quick_gelu",
|
| 67 |
+
"hidden_size": 1024,
|
| 68 |
+
"image_size": 336,
|
| 69 |
+
"initializer_factor": 1.0,
|
| 70 |
+
"initializer_range": 0.02,
|
| 71 |
+
"intermediate_size": 4096,
|
| 72 |
+
"layer_norm_eps": 1e-05,
|
| 73 |
+
"model_type": "clip_vision_model",
|
| 74 |
+
"num_attention_heads": 16,
|
| 75 |
+
"num_channels": 3,
|
| 76 |
+
"num_hidden_layers": 24,
|
| 77 |
+
"patch_size": 14,
|
| 78 |
+
"projection_dim": 768,
|
| 79 |
+
"torch_dtype": "bfloat16",
|
| 80 |
+
"vocab_size": 32000
|
| 81 |
+
},
|
| 82 |
+
"vision_feature_layer": -2,
|
| 83 |
+
"vision_feature_select_strategy": "default",
|
| 84 |
+
"vocab_size": 32064
|
| 85 |
+
}
|
slice_end/preprocessor_config.json
ADDED
|
@@ -0,0 +1,52 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"aspect_ratio_setting": "anyres",
|
| 3 |
+
"crop_size": {
|
| 4 |
+
"height": 336,
|
| 5 |
+
"width": 336
|
| 6 |
+
},
|
| 7 |
+
"do_center_crop": true,
|
| 8 |
+
"do_convert_rgb": true,
|
| 9 |
+
"do_normalize": true,
|
| 10 |
+
"do_pad": true,
|
| 11 |
+
"do_rescale": true,
|
| 12 |
+
"do_resize": true,
|
| 13 |
+
"image_grid_pinpoints": [
|
| 14 |
+
[
|
| 15 |
+
336,
|
| 16 |
+
672
|
| 17 |
+
],
|
| 18 |
+
[
|
| 19 |
+
672,
|
| 20 |
+
336
|
| 21 |
+
],
|
| 22 |
+
[
|
| 23 |
+
672,
|
| 24 |
+
672
|
| 25 |
+
],
|
| 26 |
+
[
|
| 27 |
+
1008,
|
| 28 |
+
336
|
| 29 |
+
],
|
| 30 |
+
[
|
| 31 |
+
336,
|
| 32 |
+
1008
|
| 33 |
+
]
|
| 34 |
+
],
|
| 35 |
+
"image_mean": [
|
| 36 |
+
0.48145466,
|
| 37 |
+
0.4578275,
|
| 38 |
+
0.40821073
|
| 39 |
+
],
|
| 40 |
+
"image_processor_type": "LlavaNextImageProcessor",
|
| 41 |
+
"image_std": [
|
| 42 |
+
0.26862954,
|
| 43 |
+
0.26130258,
|
| 44 |
+
0.27577711
|
| 45 |
+
],
|
| 46 |
+
"processor_class": "LlavaNextProcessor",
|
| 47 |
+
"resample": 3,
|
| 48 |
+
"rescale_factor": 0.00392156862745098,
|
| 49 |
+
"size": {
|
| 50 |
+
"shortest_edge": 336
|
| 51 |
+
}
|
| 52 |
+
}
|
slice_end/processor_config.json
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"image_token": "<image>",
|
| 3 |
+
"num_additional_image_tokens": 1,
|
| 4 |
+
"patch_size": 14,
|
| 5 |
+
"processor_class": "LlavaNextProcessor",
|
| 6 |
+
"vision_feature_select_strategy": "default"
|
| 7 |
+
}
|
slice_end/pytorch_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2943a2f4dfd6a161bafdc470a8f46f2a0d48f957837d839d49640c30743aa903
|
| 3 |
+
size 15133733934
|
slice_end/raw_outputs__aifs4su_yaodong_hantao_align-anything_outputs_llava_1.6_mistral_7b_val_top1-10_slice_end.pkl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c735dd9a0aa7797ce78fddd7c2013a94be47b7df5e7ed6027b218c6e3e3b4281
|
| 3 |
+
size 15661443
|
slice_end/special_tokens_map.json
ADDED
|
@@ -0,0 +1,31 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"bos_token": {
|
| 3 |
+
"content": "<s>",
|
| 4 |
+
"lstrip": false,
|
| 5 |
+
"normalized": false,
|
| 6 |
+
"rstrip": false,
|
| 7 |
+
"single_word": false
|
| 8 |
+
},
|
| 9 |
+
"eos_token": {
|
| 10 |
+
"content": "</s>",
|
| 11 |
+
"lstrip": false,
|
| 12 |
+
"normalized": false,
|
| 13 |
+
"rstrip": false,
|
| 14 |
+
"single_word": false
|
| 15 |
+
},
|
| 16 |
+
"image_token": "<image>",
|
| 17 |
+
"pad_token": {
|
| 18 |
+
"content": "<pad>",
|
| 19 |
+
"lstrip": false,
|
| 20 |
+
"normalized": false,
|
| 21 |
+
"rstrip": false,
|
| 22 |
+
"single_word": false
|
| 23 |
+
},
|
| 24 |
+
"unk_token": {
|
| 25 |
+
"content": "<unk>",
|
| 26 |
+
"lstrip": false,
|
| 27 |
+
"normalized": false,
|
| 28 |
+
"rstrip": false,
|
| 29 |
+
"single_word": false
|
| 30 |
+
}
|
| 31 |
+
}
|
slice_end/tokenizer.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
slice_end/tokenizer.model
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
| 3 |
+
size 493443
|
slice_end/tokenizer_config.json
ADDED
|
@@ -0,0 +1,69 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"add_bos_token": true,
|
| 3 |
+
"add_eos_token": false,
|
| 4 |
+
"add_prefix_space": null,
|
| 5 |
+
"added_tokens_decoder": {
|
| 6 |
+
"0": {
|
| 7 |
+
"content": "<unk>",
|
| 8 |
+
"lstrip": false,
|
| 9 |
+
"normalized": false,
|
| 10 |
+
"rstrip": false,
|
| 11 |
+
"single_word": false,
|
| 12 |
+
"special": true
|
| 13 |
+
},
|
| 14 |
+
"1": {
|
| 15 |
+
"content": "<s>",
|
| 16 |
+
"lstrip": false,
|
| 17 |
+
"normalized": false,
|
| 18 |
+
"rstrip": false,
|
| 19 |
+
"single_word": false,
|
| 20 |
+
"special": true
|
| 21 |
+
},
|
| 22 |
+
"2": {
|
| 23 |
+
"content": "</s>",
|
| 24 |
+
"lstrip": false,
|
| 25 |
+
"normalized": false,
|
| 26 |
+
"rstrip": false,
|
| 27 |
+
"single_word": false,
|
| 28 |
+
"special": true
|
| 29 |
+
},
|
| 30 |
+
"32000": {
|
| 31 |
+
"content": "<image>",
|
| 32 |
+
"lstrip": false,
|
| 33 |
+
"normalized": false,
|
| 34 |
+
"rstrip": false,
|
| 35 |
+
"single_word": false,
|
| 36 |
+
"special": true
|
| 37 |
+
},
|
| 38 |
+
"32001": {
|
| 39 |
+
"content": "<pad>",
|
| 40 |
+
"lstrip": false,
|
| 41 |
+
"normalized": false,
|
| 42 |
+
"rstrip": false,
|
| 43 |
+
"single_word": false,
|
| 44 |
+
"special": true
|
| 45 |
+
}
|
| 46 |
+
},
|
| 47 |
+
"additional_special_tokens": [],
|
| 48 |
+
"bos_token": "<s>",
|
| 49 |
+
"chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
| 50 |
+
"clean_up_tokenization_spaces": false,
|
| 51 |
+
"eos_token": "</s>",
|
| 52 |
+
"extra_special_tokens": {
|
| 53 |
+
"image_token": "<image>"
|
| 54 |
+
},
|
| 55 |
+
"image_token": "<image>",
|
| 56 |
+
"legacy": true,
|
| 57 |
+
"max_length": null,
|
| 58 |
+
"model_max_length": 4096,
|
| 59 |
+
"pad_to_multiple_of": null,
|
| 60 |
+
"pad_token": "<pad>",
|
| 61 |
+
"pad_token_type_id": 0,
|
| 62 |
+
"padding_side": "left",
|
| 63 |
+
"processor_class": "LlavaNextProcessor",
|
| 64 |
+
"sp_model_kwargs": {},
|
| 65 |
+
"spaces_between_special_tokens": false,
|
| 66 |
+
"tokenizer_class": "LlamaTokenizer",
|
| 67 |
+
"unk_token": "<unk>",
|
| 68 |
+
"use_default_system_prompt": false
|
| 69 |
+
}
|
wandb/debug-internal.log
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:39:29.357857467+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.8","symlink path":"../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173929-oflhglry/logs/debug-core.log"}
|
| 2 |
+
{"time":"2025-03-31T17:39:29.573098876+08:00","level":"INFO","msg":"created new stream","id":"oflhglry"}
|
| 3 |
+
{"time":"2025-03-31T17:39:29.573167483+08:00","level":"INFO","msg":"stream: started","id":"oflhglry"}
|
| 4 |
+
{"time":"2025-03-31T17:39:29.573190711+08:00","level":"INFO","msg":"handler: started","stream_id":"oflhglry"}
|
| 5 |
+
{"time":"2025-03-31T17:39:29.573196411+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"oflhglry"}
|
| 6 |
+
{"time":"2025-03-31T17:39:29.573195051+08:00","level":"INFO","msg":"sender: started","stream_id":"oflhglry"}
|
| 7 |
+
{"time":"2025-03-31T17:39:29.880478815+08:00","level":"INFO","msg":"Starting system monitor"}
|
| 8 |
+
{"time":"2025-03-31T17:44:01.38994239+08:00","level":"INFO","msg":"Stopping system monitor"}
|
| 9 |
+
{"time":"2025-03-31T17:44:01.39093428+08:00","level":"INFO","msg":"Stopped system monitor"}
|
| 10 |
+
{"time":"2025-03-31T17:44:02.852694686+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 11 |
+
{"time":"2025-03-31T17:44:02.902712341+08:00","level":"INFO","msg":"stream: closing","id":"oflhglry"}
|
| 12 |
+
{"time":"2025-03-31T17:44:02.902739889+08:00","level":"WARN","msg":"sender: received Exit record more than once, ignoring"}
|
| 13 |
+
{"time":"2025-03-31T17:44:03.099767177+08:00","level":"INFO","msg":"handler: closed","stream_id":"oflhglry"}
|
| 14 |
+
{"time":"2025-03-31T17:44:03.099816798+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":"oflhglry"}
|
| 15 |
+
{"time":"2025-03-31T17:44:03.099920876+08:00","level":"INFO","msg":"sender: closed","stream_id":"oflhglry"}
|
| 16 |
+
{"time":"2025-03-31T17:44:03.100069001+08:00","level":"INFO","msg":"stream: closed","id":"oflhglry"}
|
wandb/debug.log
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Current SDK version is 0.19.8
|
| 2 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Configure stats pid to 1814399
|
| 3 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Loading settings from /home/yangyaodong/.config/wandb/settings
|
| 4 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Loading settings from /aifs4su/yaodong/hantao/align-anything/scripts/wandb/settings
|
| 5 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Loading settings from environment variables
|
| 6 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_init.py:setup_run_log_directory():647] Logging user logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173929-oflhglry/logs/debug.log
|
| 7 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_init.py:setup_run_log_directory():648] Logging internal logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173929-oflhglry/logs/debug-internal.log
|
| 8 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_init.py:init():761] calling init triggers
|
| 9 |
+
2025-03-31 17:39:29,350 INFO MainThread:1814399 [wandb_init.py:init():766] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'train_cfgs': {'save_checkpoint': False, 'load_checkpoint': False, 'ds_cfgs': 'ds_z3_config.json', 'epochs': 3, 'seed': 42, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 1, 'gradient_checkpointing': True, 'learning_rate': 1e-06, 'lr_scheduler_type': 'cosine', 'lr_warmup_ratio': 0.03, 'weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'bf16': True, 'fp16': False, 'eval_strategy': 'epoch', 'eval_interval': 10, 'regularization': 0.001, 'scale_coeff': 0.1, 'freeze_mm_proj': False, 'freeze_vision_tower': True, 'freeze_language_model': False}, 'data_cfgs': {'train_datasets': '/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10', 'train_template': 'AA_TI2T_LLAVA', 'train_size': {}, 'train_split': 'train', 'train_name': 'text-image-to-text', 'train_data_files': {}, 'train_optional_args': [], 'eval_datasets': {}, 'eval_template': {}, 'eval_size': {}, 'eval_split': {}, 'eval_subset': {}, 'eval_data_files': {}, 'eval_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'dpo', 'output_dir': '../outputs/llava_1.6_mistral_7b_val/top1-10', 'cache_dir': {}, 'save_total_limit': 3}, 'model_cfgs': {'model_name_or_path': '/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf', 'trust_remote_code': True, 'model_max_length': 4096}, 'special_tokens': {}, '_wandb': {}}
|
| 11 |
+
2025-03-31 17:39:29,350 INFO MainThread:1814399 [wandb_init.py:init():784] starting backend
|
| 12 |
+
2025-03-31 17:39:29,350 INFO MainThread:1814399 [wandb_init.py:init():788] sending inform_init request
|
| 13 |
+
2025-03-31 17:39:29,355 INFO MainThread:1814399 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
| 14 |
+
2025-03-31 17:39:29,355 INFO MainThread:1814399 [wandb_init.py:init():798] backend started and connected
|
| 15 |
+
2025-03-31 17:39:29,357 INFO MainThread:1814399 [wandb_init.py:init():891] updated telemetry
|
| 16 |
+
2025-03-31 17:39:29,371 INFO MainThread:1814399 [wandb_init.py:init():915] communicating run to backend with 90.0 second timeout
|
| 17 |
+
2025-03-31 17:39:29,874 INFO MainThread:1814399 [wandb_init.py:init():990] starting run threads in backend
|
| 18 |
+
2025-03-31 17:39:30,115 INFO MainThread:1814399 [wandb_run.py:_console_start():2375] atexit reg
|
| 19 |
+
2025-03-31 17:39:30,115 INFO MainThread:1814399 [wandb_run.py:_redirect():2227] redirect: wrap_raw
|
| 20 |
+
2025-03-31 17:39:30,115 INFO MainThread:1814399 [wandb_run.py:_redirect():2292] Wrapping output streams.
|
| 21 |
+
2025-03-31 17:39:30,116 INFO MainThread:1814399 [wandb_run.py:_redirect():2315] Redirects installed.
|
| 22 |
+
2025-03-31 17:39:30,118 INFO MainThread:1814399 [wandb_init.py:init():1032] run started, returning control to user process
|
| 23 |
+
2025-03-31 17:44:01,349 INFO MainThread:1814399 [wandb_run.py:_finish():2112] finishing run htlou/align-anything/oflhglry
|
| 24 |
+
2025-03-31 17:44:01,350 INFO MainThread:1814399 [wandb_run.py:_atexit_cleanup():2340] got exitcode: 0
|
| 25 |
+
2025-03-31 17:44:01,350 INFO MainThread:1814399 [wandb_run.py:_restore():2322] restore
|
| 26 |
+
2025-03-31 17:44:01,351 INFO MainThread:1814399 [wandb_run.py:_restore():2328] restore done
|
| 27 |
+
2025-03-31 17:44:02,351 INFO MainThread:1814399 [wandb_run.py:_restore():2322] restore
|
| 28 |
+
2025-03-31 17:44:02,352 INFO MainThread:1814399 [wandb_run.py:_restore():2328] restore done
|
| 29 |
+
2025-03-31 17:44:02,352 ERROR MainThread:1814399 [wandb_run.py:_atexit_cleanup():2361] Problem finishing run
|
| 30 |
+
Traceback (most recent call last):
|
| 31 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2352, in _atexit_cleanup
|
| 32 |
+
self._on_finish()
|
| 33 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2609, in _on_finish
|
| 34 |
+
wait_with_progress(
|
| 35 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/mailbox/wait_with_progress.py", line 24, in wait_with_progress
|
| 36 |
+
return wait_all_with_progress(
|
| 37 |
+
^^^^^^^^^^^^^^^^^^^^^^^
|
| 38 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/mailbox/wait_with_progress.py", line 87, in wait_all_with_progress
|
| 39 |
+
return asyncio_compat.run(progress_loop_with_timeout)
|
| 40 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 41 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/lib/asyncio_compat.py", line 27, in run
|
| 42 |
+
future = executor.submit(runner.run, fn)
|
| 43 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 44 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/concurrent/futures/thread.py", line 169, in submit
|
| 45 |
+
raise RuntimeError('cannot schedule new futures after '
|
| 46 |
+
RuntimeError: cannot schedule new futures after interpreter shutdown
|
| 47 |
+
2025-03-31 17:44:02,902 INFO MsgRouterThr:1814399 [mailbox.py:close():129] Closing mailbox, abandoning 2 handles.
|
wandb/run-20250331_173510-7rsczbti/files/output.log
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
***** Running training *****
|
| 2 |
+
Resuming from checkpoint 1/3 epoch : 0%| | 0/72 [00:00<?, ?it/s]/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/torch/utils/checkpoint.py:87: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
|
| 3 |
+
warnings.warn(
|
| 4 |
+
`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`.
|
| 5 |
+
Training 1/3 epoch (loss 0.6172): 7%|████████▌ | 5/72 [00:19<03:24, 3.05s/it]Token indices sequence length is longer than the specified maximum sequence length for this model (4205 > 4096). Running this sequence through the model will result in indexing errors
|
| 6 |
+
Training 1/3 epoch (loss 0.8398): 12%|███████████████▌ | 9/72 [00:58<07:20, 6.99s/it]
|
| 7 |
+
Saving checkpoint at step 6 ...
|
| 8 |
+
Saving model to "../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6" ...
|
| 9 |
+
Saving 16-bit model...
|
| 10 |
+
[2025-03-31 17:35:42,883] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step6 is about to be saved!
|
| 11 |
+
[2025-03-31 17:35:42,884] [INFO] [engine.py:3831:save_16bit_model] Saving model weights to ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6/pytorch_model.bin, tag: global_step6
|
| 12 |
+
[2025-03-31 17:35:42,884] [INFO] [torch_checkpoint_engine.py:21:save] [Torch] Saving ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6/pytorch_model.bin...
|
| 13 |
+
[2025-03-31 17:36:03,605] [INFO] [torch_checkpoint_engine.py:23:save] [Torch] Saved ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6/pytorch_model.bin.
|
| 14 |
+
[2025-03-31 17:36:03,608] [INFO] [torch_checkpoint_engine.py:33:commit] [Torch] Checkpoint global_step6 is ready now!
|
| 15 |
+
Model saved!
|
| 16 |
+
Checkpoint saved.
|
wandb/run-20250331_173510-7rsczbti/files/requirements.txt
ADDED
|
@@ -0,0 +1,258 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
maskrcnn_benchmark==0.0.0
|
| 2 |
+
webdataset==0.2.111
|
| 3 |
+
websockets==15.0.1
|
| 4 |
+
typer==0.15.2
|
| 5 |
+
blobfile==3.0.0
|
| 6 |
+
pooch==1.8.2
|
| 7 |
+
filelock==3.18.0
|
| 8 |
+
referencing==0.36.2
|
| 9 |
+
matplotlib==3.10.1
|
| 10 |
+
cachetools==5.5.2
|
| 11 |
+
python-dateutil==2.9.0.post0
|
| 12 |
+
gmpy2==2.2.1
|
| 13 |
+
uvloop==0.21.0
|
| 14 |
+
nvidia-cusparselt-cu12==0.6.2
|
| 15 |
+
clip==0.2.0
|
| 16 |
+
httpcore==1.0.7
|
| 17 |
+
charset-normalizer==3.3.2
|
| 18 |
+
torchlibrosa==0.1.0
|
| 19 |
+
contourpy==1.3.1
|
| 20 |
+
multiprocess==0.70.16
|
| 21 |
+
nest-asyncio==1.6.0
|
| 22 |
+
Werkzeug==3.1.3
|
| 23 |
+
aiofiles==23.2.1
|
| 24 |
+
six==1.17.0
|
| 25 |
+
torch==2.6.0
|
| 26 |
+
sse-starlette==2.2.1
|
| 27 |
+
typing_extensions==4.12.2
|
| 28 |
+
xgrammar==0.1.16
|
| 29 |
+
psutil==7.0.0
|
| 30 |
+
kiwisolver==1.4.8
|
| 31 |
+
moviepy==2.1.2
|
| 32 |
+
frozenlist==1.5.0
|
| 33 |
+
jiter==0.9.0
|
| 34 |
+
einops==0.8.1
|
| 35 |
+
flash_attn==2.7.4.post1
|
| 36 |
+
PySocks==1.7.1
|
| 37 |
+
regex==2024.11.6
|
| 38 |
+
markdown-it-py==3.0.0
|
| 39 |
+
ruff==0.11.2
|
| 40 |
+
docker-pycreds==0.4.0
|
| 41 |
+
nvidia-nvtx-cu12==12.4.127
|
| 42 |
+
pyparsing==3.2.3
|
| 43 |
+
resampy==0.4.3
|
| 44 |
+
tokenizers==0.21.0
|
| 45 |
+
frechet-audio-distance==0.1.2
|
| 46 |
+
aiohappyeyeballs==2.6.1
|
| 47 |
+
llamafactory==0.9.3.dev0
|
| 48 |
+
msgspec==0.19.0
|
| 49 |
+
httpx==0.28.1
|
| 50 |
+
encodec==0.1.1
|
| 51 |
+
ffmpy==0.5.0
|
| 52 |
+
jsonschema==4.23.0
|
| 53 |
+
imageio-ffmpeg==0.6.0
|
| 54 |
+
mkl_random==1.2.8
|
| 55 |
+
fairscale==0.4.13
|
| 56 |
+
soxr==0.5.0.post1
|
| 57 |
+
lark==1.2.2
|
| 58 |
+
gradio==5.21.0
|
| 59 |
+
absl-py==2.2.1
|
| 60 |
+
dnspython==2.7.0
|
| 61 |
+
networkx==3.4.2
|
| 62 |
+
h5py==3.13.0
|
| 63 |
+
hjson==3.1.0
|
| 64 |
+
tensorboard==2.19.0
|
| 65 |
+
aiosignal==1.3.2
|
| 66 |
+
pip==25.0
|
| 67 |
+
nvidia-cublas-cu12==12.4.5.8
|
| 68 |
+
llguidance==0.7.11
|
| 69 |
+
zipp==3.21.0
|
| 70 |
+
ftfy==6.3.1
|
| 71 |
+
peft==0.15.0
|
| 72 |
+
attrs==25.3.0
|
| 73 |
+
trl==0.9.6
|
| 74 |
+
requests==2.32.3
|
| 75 |
+
progressbar==2.5
|
| 76 |
+
sniffio==1.3.1
|
| 77 |
+
pycountry==24.6.1
|
| 78 |
+
lxml==5.3.1
|
| 79 |
+
starlette==0.46.1
|
| 80 |
+
pytest==7.2.0
|
| 81 |
+
Markdown==3.7
|
| 82 |
+
mdurl==0.1.2
|
| 83 |
+
pyzmq==26.3.0
|
| 84 |
+
safetensors==0.5.3
|
| 85 |
+
opencv-python==4.6.0.66
|
| 86 |
+
prometheus-fastapi-instrumentator==7.1.0
|
| 87 |
+
shellingham==1.5.4
|
| 88 |
+
torchvision==0.21.0
|
| 89 |
+
pluggy==1.5.0
|
| 90 |
+
timm==1.0.15
|
| 91 |
+
multidict==6.2.0
|
| 92 |
+
semantic-version==2.10.0
|
| 93 |
+
airportsdata==20250224
|
| 94 |
+
numba==0.60.0
|
| 95 |
+
MarkupSafe==2.1.5
|
| 96 |
+
pydantic_core==2.33.0
|
| 97 |
+
imageio==2.37.0
|
| 98 |
+
nvidia-nccl-cu12==2.21.5
|
| 99 |
+
dill==0.3.8
|
| 100 |
+
msgpack==1.1.0
|
| 101 |
+
sentry-sdk==2.24.1
|
| 102 |
+
rpds-py==0.24.0
|
| 103 |
+
grpcio==1.71.0
|
| 104 |
+
fastrlock==0.8.3
|
| 105 |
+
python-json-logger==3.3.0
|
| 106 |
+
cffi==1.17.1
|
| 107 |
+
gradio_client==1.7.2
|
| 108 |
+
PyYAML==6.0.2
|
| 109 |
+
tensorboard-data-server==0.7.2
|
| 110 |
+
termcolor==2.5.0
|
| 111 |
+
torchaudio==2.6.0
|
| 112 |
+
triton==3.2.0
|
| 113 |
+
fastapi==0.115.12
|
| 114 |
+
clint==0.5.1
|
| 115 |
+
lazy_loader==0.4
|
| 116 |
+
depyf==0.18.0
|
| 117 |
+
mkl_fft==1.3.11
|
| 118 |
+
annotated-types==0.7.0
|
| 119 |
+
scikit-learn==1.6.1
|
| 120 |
+
wget==3.2
|
| 121 |
+
setuptools==75.8.0
|
| 122 |
+
args==0.1.0
|
| 123 |
+
certifi==2025.1.31
|
| 124 |
+
click==8.1.8
|
| 125 |
+
python-dotenv==1.1.0
|
| 126 |
+
laion_clap==1.1.5
|
| 127 |
+
Pygments==2.19.1
|
| 128 |
+
tomlkit==0.13.2
|
| 129 |
+
idna==3.7
|
| 130 |
+
propcache==0.3.1
|
| 131 |
+
platformdirs==4.3.7
|
| 132 |
+
align-anything==0.0.1.dev0
|
| 133 |
+
ray==2.44.1
|
| 134 |
+
cloudpickle==3.1.1
|
| 135 |
+
deepspeed==0.16.5
|
| 136 |
+
smmap==5.0.2
|
| 137 |
+
distro==1.9.0
|
| 138 |
+
fonttools==4.56.0
|
| 139 |
+
typing-inspection==0.4.0
|
| 140 |
+
braceexpand==0.1.7
|
| 141 |
+
decorator==5.2.1
|
| 142 |
+
diskcache==5.6.3
|
| 143 |
+
yt-dlp==2025.3.27
|
| 144 |
+
shtab==1.7.1
|
| 145 |
+
gguf==0.10.0
|
| 146 |
+
interegular==0.3.3
|
| 147 |
+
compressed-tensors==0.9.2
|
| 148 |
+
pandas==2.2.3
|
| 149 |
+
huggingface-hub==0.29.3
|
| 150 |
+
pyarrow==19.0.1
|
| 151 |
+
lm-format-enforcer==0.10.11
|
| 152 |
+
GitPython==3.1.44
|
| 153 |
+
xxhash==3.5.0
|
| 154 |
+
packaging==24.2
|
| 155 |
+
setproctitle==1.3.5
|
| 156 |
+
llvmlite==0.43.0
|
| 157 |
+
tiktoken==0.9.0
|
| 158 |
+
mpmath==1.3.0
|
| 159 |
+
email_validator==2.2.0
|
| 160 |
+
nvidia-ml-py==12.570.86
|
| 161 |
+
pydantic==2.11.0
|
| 162 |
+
xformers==0.0.29.post2
|
| 163 |
+
httptools==0.6.4
|
| 164 |
+
librosa==0.11.0
|
| 165 |
+
pytorch-fid==0.3.0
|
| 166 |
+
hpsv2==1.2.0
|
| 167 |
+
nvidia-cudnn-cu12==9.1.0.70
|
| 168 |
+
iniconfig==2.1.0
|
| 169 |
+
sympy==1.13.1
|
| 170 |
+
safehttpx==0.1.6
|
| 171 |
+
jsonschema-specifications==2024.10.1
|
| 172 |
+
Jinja2==3.1.6
|
| 173 |
+
tyro==0.8.14
|
| 174 |
+
h11==0.14.0
|
| 175 |
+
aiohttp==3.11.14
|
| 176 |
+
diffusers==0.32.2
|
| 177 |
+
tqdm==4.67.1
|
| 178 |
+
blake3==1.0.4
|
| 179 |
+
vllm==0.8.2
|
| 180 |
+
scipy==1.10.1
|
| 181 |
+
audioread==3.0.1
|
| 182 |
+
proglog==0.1.10
|
| 183 |
+
fire==0.7.0
|
| 184 |
+
sentencepiece==0.2.0
|
| 185 |
+
pytz==2025.2
|
| 186 |
+
nvidia-nvjitlink-cu12==12.4.127
|
| 187 |
+
nvidia-curand-cu12==10.3.5.147
|
| 188 |
+
numpy==1.26.4
|
| 189 |
+
tzdata==2025.2
|
| 190 |
+
python-multipart==0.0.20
|
| 191 |
+
urllib3==2.3.0
|
| 192 |
+
pycryptodomex==3.22.0
|
| 193 |
+
yarl==1.18.3
|
| 194 |
+
outlines==0.1.11
|
| 195 |
+
nvidia-cusolver-cu12==11.6.1.9
|
| 196 |
+
pydub==0.25.1
|
| 197 |
+
mistral_common==1.5.4
|
| 198 |
+
pycparser==2.22
|
| 199 |
+
pytest-split==0.8.0
|
| 200 |
+
datasets==3.4.1
|
| 201 |
+
soundfile==0.13.1
|
| 202 |
+
transformers==4.50.0
|
| 203 |
+
image-reward==1.5
|
| 204 |
+
wcwidth==0.2.13
|
| 205 |
+
nvidia-cuda-nvrtc-cu12==12.4.127
|
| 206 |
+
groovy==0.1.2
|
| 207 |
+
nvidia-cuda-runtime-cu12==12.4.127
|
| 208 |
+
astor==0.8.1
|
| 209 |
+
anyio==4.9.0
|
| 210 |
+
wandb==0.19.8
|
| 211 |
+
joblib==1.4.2
|
| 212 |
+
fsspec==2024.12.0
|
| 213 |
+
accelerate==1.5.2
|
| 214 |
+
py-cpuinfo==9.0.0
|
| 215 |
+
docstring_parser==0.16
|
| 216 |
+
partial-json-parser==0.2.1.1.post5
|
| 217 |
+
nvidia-cusparse-cu12==12.3.1.170
|
| 218 |
+
protobuf==3.20.3
|
| 219 |
+
outlines_core==0.1.26
|
| 220 |
+
nvidia-cufft-cu12==11.2.1.3
|
| 221 |
+
cycler==0.12.1
|
| 222 |
+
uvicorn==0.34.0
|
| 223 |
+
orjson==3.10.16
|
| 224 |
+
av==14.2.0
|
| 225 |
+
Brotli==1.0.9
|
| 226 |
+
cupy-cuda12x==13.4.1
|
| 227 |
+
openai==1.69.0
|
| 228 |
+
rich==13.9.4
|
| 229 |
+
importlib_metadata==8.6.1
|
| 230 |
+
ninja==1.11.1.4
|
| 231 |
+
wheel==0.45.1
|
| 232 |
+
pillow==10.4.0
|
| 233 |
+
prometheus_client==0.21.1
|
| 234 |
+
threadpoolctl==3.6.0
|
| 235 |
+
gitdb==4.0.12
|
| 236 |
+
watchfiles==1.0.4
|
| 237 |
+
nvidia-cuda-cupti-cu12==12.4.127
|
| 238 |
+
opencv-python-headless==4.11.0.86
|
| 239 |
+
mkl-service==2.4.0
|
| 240 |
+
rich-toolkit==0.14.0
|
| 241 |
+
fastapi-cli==0.0.7
|
| 242 |
+
llamafactory==0.9.3.dev0
|
| 243 |
+
typing_extensions==4.12.2
|
| 244 |
+
tomli==2.0.1
|
| 245 |
+
zipp==3.19.2
|
| 246 |
+
wheel==0.43.0
|
| 247 |
+
jaraco.text==3.12.1
|
| 248 |
+
packaging==24.2
|
| 249 |
+
autocommand==2.2.2
|
| 250 |
+
jaraco.functools==4.0.1
|
| 251 |
+
jaraco.collections==5.1.0
|
| 252 |
+
platformdirs==4.2.2
|
| 253 |
+
more-itertools==10.3.0
|
| 254 |
+
inflect==7.3.1
|
| 255 |
+
jaraco.context==5.3.0
|
| 256 |
+
typeguard==4.3.0
|
| 257 |
+
backports.tarfile==1.2.0
|
| 258 |
+
importlib_metadata==8.0.0
|
wandb/run-20250331_173510-7rsczbti/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-5.15.0-1040-nvidia-x86_64-with-glibc2.35",
|
| 3 |
+
"python": "CPython 3.11.11",
|
| 4 |
+
"startedAt": "2025-03-31T09:35:10.840464Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"--local_rank=0",
|
| 7 |
+
"--model_name_or_path",
|
| 8 |
+
"/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf",
|
| 9 |
+
"--train_datasets",
|
| 10 |
+
"/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10",
|
| 11 |
+
"--train_template",
|
| 12 |
+
"AA_TI2T_LLAVA",
|
| 13 |
+
"--train_split",
|
| 14 |
+
"train",
|
| 15 |
+
"--train_name",
|
| 16 |
+
"text-image-to-text",
|
| 17 |
+
"--output_dir",
|
| 18 |
+
"../outputs/llava_1.6_mistral_7b_val/top1-10",
|
| 19 |
+
"--save_total_limit",
|
| 20 |
+
"12",
|
| 21 |
+
"--train_batch_size",
|
| 22 |
+
"8",
|
| 23 |
+
"--epochs",
|
| 24 |
+
"3"
|
| 25 |
+
],
|
| 26 |
+
"program": "-m align_anything.trainers.text_image_to_text.dpo",
|
| 27 |
+
"git": {
|
| 28 |
+
"remote": "git@github.com-hantao:PKU-Alignment/align-anything.git",
|
| 29 |
+
"commit": "106588f9802757a3283c1aff1f33ea9afd737f31"
|
| 30 |
+
},
|
| 31 |
+
"email": "2200017789@stu.pku.edu.cn",
|
| 32 |
+
"root": "../outputs/llava_1.6_mistral_7b_val/top1-10",
|
| 33 |
+
"host": "dgx-092",
|
| 34 |
+
"executable": "/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/python",
|
| 35 |
+
"cpu_count": 112,
|
| 36 |
+
"cpu_count_logical": 224,
|
| 37 |
+
"gpu": "NVIDIA H800",
|
| 38 |
+
"gpu_count": 8,
|
| 39 |
+
"disk": {
|
| 40 |
+
"/": {
|
| 41 |
+
"total": "1888556142592",
|
| 42 |
+
"used": "149000343552"
|
| 43 |
+
}
|
| 44 |
+
},
|
| 45 |
+
"memory": {
|
| 46 |
+
"total": "2164195454976"
|
| 47 |
+
},
|
| 48 |
+
"cpu": {
|
| 49 |
+
"count": 112,
|
| 50 |
+
"countLogical": 224
|
| 51 |
+
},
|
| 52 |
+
"gpu_nvidia": [
|
| 53 |
+
{
|
| 54 |
+
"name": "NVIDIA H800",
|
| 55 |
+
"memoryTotal": "85520809984",
|
| 56 |
+
"cudaCores": 16896,
|
| 57 |
+
"architecture": "Hopper"
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"name": "NVIDIA H800",
|
| 61 |
+
"memoryTotal": "85520809984",
|
| 62 |
+
"cudaCores": 16896,
|
| 63 |
+
"architecture": "Hopper"
|
| 64 |
+
},
|
| 65 |
+
{
|
| 66 |
+
"name": "NVIDIA H800",
|
| 67 |
+
"memoryTotal": "85520809984",
|
| 68 |
+
"cudaCores": 16896,
|
| 69 |
+
"architecture": "Hopper"
|
| 70 |
+
},
|
| 71 |
+
{
|
| 72 |
+
"name": "NVIDIA H800",
|
| 73 |
+
"memoryTotal": "85520809984",
|
| 74 |
+
"cudaCores": 16896,
|
| 75 |
+
"architecture": "Hopper"
|
| 76 |
+
},
|
| 77 |
+
{
|
| 78 |
+
"name": "NVIDIA H800",
|
| 79 |
+
"memoryTotal": "85520809984",
|
| 80 |
+
"cudaCores": 16896,
|
| 81 |
+
"architecture": "Hopper"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "NVIDIA H800",
|
| 85 |
+
"memoryTotal": "85520809984",
|
| 86 |
+
"cudaCores": 16896,
|
| 87 |
+
"architecture": "Hopper"
|
| 88 |
+
},
|
| 89 |
+
{
|
| 90 |
+
"name": "NVIDIA H800",
|
| 91 |
+
"memoryTotal": "85520809984",
|
| 92 |
+
"cudaCores": 16896,
|
| 93 |
+
"architecture": "Hopper"
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"name": "NVIDIA H800",
|
| 97 |
+
"memoryTotal": "85520809984",
|
| 98 |
+
"cudaCores": 16896,
|
| 99 |
+
"architecture": "Hopper"
|
| 100 |
+
}
|
| 101 |
+
],
|
| 102 |
+
"slurm": {
|
| 103 |
+
"conf": "/cm/shared/apps/slurm/var/etc/slurm/slurm.conf"
|
| 104 |
+
},
|
| 105 |
+
"cudaVersion": "12.2"
|
| 106 |
+
}
|
wandb/run-20250331_173510-7rsczbti/logs/debug-core.log
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:35:10.259294371+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpy9i5crdw/port-1800486.txt","pid":1800486,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false}
|
| 2 |
+
{"time":"2025-03-31T17:35:10.260539464+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1800486}
|
| 3 |
+
{"time":"2025-03-31T17:35:10.260534373+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":34021,"Zone":""}}
|
| 4 |
+
{"time":"2025-03-31T17:35:10.409005578+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:42092"}
|
| 5 |
+
{"time":"2025-03-31T17:35:10.842218444+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"7rsczbti","id":"127.0.0.1:42092"}
|
| 6 |
+
{"time":"2025-03-31T17:35:11.059457668+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"7rsczbti","id":"127.0.0.1:42092"}
|
| 7 |
+
{"time":"2025-03-31T17:36:11.289689563+08:00","level":"INFO","msg":"received shutdown signal","signal":15}
|
wandb/run-20250331_173510-7rsczbti/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:35:10.84379093+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.8","symlink path":"../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173510-7rsczbti/logs/debug-core.log"}
|
| 2 |
+
{"time":"2025-03-31T17:35:11.059368118+08:00","level":"INFO","msg":"created new stream","id":"7rsczbti"}
|
| 3 |
+
{"time":"2025-03-31T17:35:11.05945062+08:00","level":"INFO","msg":"stream: started","id":"7rsczbti"}
|
| 4 |
+
{"time":"2025-03-31T17:35:11.059508775+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"7rsczbti"}
|
| 5 |
+
{"time":"2025-03-31T17:35:11.059511735+08:00","level":"INFO","msg":"handler: started","stream_id":"7rsczbti"}
|
| 6 |
+
{"time":"2025-03-31T17:35:11.059514206+08:00","level":"INFO","msg":"sender: started","stream_id":"7rsczbti"}
|
| 7 |
+
{"time":"2025-03-31T17:35:11.777198217+08:00","level":"INFO","msg":"Starting system monitor"}
|
wandb/run-20250331_173510-7rsczbti/logs/debug.log
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_setup.py:_flush():67] Current SDK version is 0.19.8
|
| 2 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_setup.py:_flush():67] Configure stats pid to 1800486
|
| 3 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_setup.py:_flush():67] Loading settings from /home/yangyaodong/.config/wandb/settings
|
| 4 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_setup.py:_flush():67] Loading settings from /aifs4su/yaodong/hantao/align-anything/scripts/wandb/settings
|
| 5 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_setup.py:_flush():67] Loading settings from environment variables
|
| 6 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_init.py:setup_run_log_directory():647] Logging user logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173510-7rsczbti/logs/debug.log
|
| 7 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_init.py:setup_run_log_directory():648] Logging internal logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173510-7rsczbti/logs/debug-internal.log
|
| 8 |
+
2025-03-31 17:35:10,830 INFO MainThread:1800486 [wandb_init.py:init():761] calling init triggers
|
| 9 |
+
2025-03-31 17:35:10,831 INFO MainThread:1800486 [wandb_init.py:init():766] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'train_cfgs': {'save_checkpoint': False, 'load_checkpoint': False, 'ds_cfgs': 'ds_z3_config.json', 'epochs': 3, 'seed': 42, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 1, 'gradient_checkpointing': True, 'learning_rate': 1e-06, 'lr_scheduler_type': 'cosine', 'lr_warmup_ratio': 0.03, 'weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'bf16': True, 'fp16': False, 'eval_strategy': 'epoch', 'eval_interval': 10, 'regularization': 0.001, 'scale_coeff': 0.1, 'freeze_mm_proj': False, 'freeze_vision_tower': True, 'freeze_language_model': False}, 'data_cfgs': {'train_datasets': '/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10', 'train_template': 'AA_TI2T_LLAVA', 'train_size': {}, 'train_split': 'train', 'train_name': 'text-image-to-text', 'train_data_files': {}, 'train_optional_args': [], 'eval_datasets': {}, 'eval_template': {}, 'eval_size': {}, 'eval_split': {}, 'eval_subset': {}, 'eval_data_files': {}, 'eval_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'dpo', 'output_dir': '../outputs/llava_1.6_mistral_7b_val/top1-10', 'cache_dir': {}, 'save_total_limit': 12}, 'model_cfgs': {'model_name_or_path': '/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf', 'trust_remote_code': True, 'model_max_length': 4096}, 'special_tokens': {}, '_wandb': {}}
|
| 11 |
+
2025-03-31 17:35:10,831 INFO MainThread:1800486 [wandb_init.py:init():784] starting backend
|
| 12 |
+
2025-03-31 17:35:10,831 INFO MainThread:1800486 [wandb_init.py:init():788] sending inform_init request
|
| 13 |
+
2025-03-31 17:35:10,840 INFO MainThread:1800486 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
| 14 |
+
2025-03-31 17:35:10,840 INFO MainThread:1800486 [wandb_init.py:init():798] backend started and connected
|
| 15 |
+
2025-03-31 17:35:10,841 INFO MainThread:1800486 [wandb_init.py:init():891] updated telemetry
|
| 16 |
+
2025-03-31 17:35:10,853 INFO MainThread:1800486 [wandb_init.py:init():915] communicating run to backend with 90.0 second timeout
|
| 17 |
+
2025-03-31 17:35:11,770 INFO MainThread:1800486 [wandb_init.py:init():990] starting run threads in backend
|
| 18 |
+
2025-03-31 17:35:12,157 INFO MainThread:1800486 [wandb_run.py:_console_start():2375] atexit reg
|
| 19 |
+
2025-03-31 17:35:12,157 INFO MainThread:1800486 [wandb_run.py:_redirect():2227] redirect: wrap_raw
|
| 20 |
+
2025-03-31 17:35:12,157 INFO MainThread:1800486 [wandb_run.py:_redirect():2292] Wrapping output streams.
|
| 21 |
+
2025-03-31 17:35:12,157 INFO MainThread:1800486 [wandb_run.py:_redirect():2315] Redirects installed.
|
| 22 |
+
2025-03-31 17:35:12,162 INFO MainThread:1800486 [wandb_init.py:init():1032] run started, returning control to user process
|
wandb/run-20250331_173510-7rsczbti/run-7rsczbti.wandb
ADDED
|
File without changes
|
wandb/run-20250331_173726-4v7gx39g/files/output.log
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
***** Running training *****
|
| 2 |
+
Resuming from checkpoint 1/3 epoch : 0%| | 0/72 [00:00<?, ?it/s]/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/torch/utils/checkpoint.py:87: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
|
| 3 |
+
warnings.warn(
|
| 4 |
+
`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`.
|
| 5 |
+
Training 1/3 epoch (loss 0.6289): 7%|████████▌ | 5/72 [00:17<03:10, 2.84s/it]Token indices sequence length is longer than the specified maximum sequence length for this model (4205 > 4096). Running this sequence through the model will result in indexing errors
|
| 6 |
+
Training 1/3 epoch (loss 0.8125): 10%|████████████ | 7/72 [00:49<12:21, 11.41s/it]
|
| 7 |
+
Saving checkpoint at step 6 ...
|
| 8 |
+
Saving model to "../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6" ...
|
| 9 |
+
Saving 16-bit model...
|
| 10 |
+
[2025-03-31 17:37:56,524] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step6 is about to be saved!
|
| 11 |
+
[2025-03-31 17:37:56,525] [INFO] [engine.py:3831:save_16bit_model] Saving model weights to ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6/pytorch_model.bin, tag: global_step6
|
| 12 |
+
[2025-03-31 17:37:56,526] [INFO] [torch_checkpoint_engine.py:21:save] [Torch] Saving ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6/pytorch_model.bin...
|
| 13 |
+
[2025-03-31 17:38:14,981] [INFO] [torch_checkpoint_engine.py:23:save] [Torch] Saved ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_6/pytorch_model.bin.
|
| 14 |
+
[2025-03-31 17:38:14,981] [INFO] [torch_checkpoint_engine.py:33:commit] [Torch] Checkpoint global_step6 is ready now!
|
| 15 |
+
Model saved!
|
| 16 |
+
Checkpoint saved.
|
wandb/run-20250331_173726-4v7gx39g/files/requirements.txt
ADDED
|
@@ -0,0 +1,258 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
maskrcnn_benchmark==0.0.0
|
| 2 |
+
webdataset==0.2.111
|
| 3 |
+
websockets==15.0.1
|
| 4 |
+
typer==0.15.2
|
| 5 |
+
blobfile==3.0.0
|
| 6 |
+
pooch==1.8.2
|
| 7 |
+
filelock==3.18.0
|
| 8 |
+
referencing==0.36.2
|
| 9 |
+
matplotlib==3.10.1
|
| 10 |
+
cachetools==5.5.2
|
| 11 |
+
python-dateutil==2.9.0.post0
|
| 12 |
+
gmpy2==2.2.1
|
| 13 |
+
uvloop==0.21.0
|
| 14 |
+
nvidia-cusparselt-cu12==0.6.2
|
| 15 |
+
clip==0.2.0
|
| 16 |
+
httpcore==1.0.7
|
| 17 |
+
charset-normalizer==3.3.2
|
| 18 |
+
torchlibrosa==0.1.0
|
| 19 |
+
contourpy==1.3.1
|
| 20 |
+
multiprocess==0.70.16
|
| 21 |
+
nest-asyncio==1.6.0
|
| 22 |
+
Werkzeug==3.1.3
|
| 23 |
+
aiofiles==23.2.1
|
| 24 |
+
six==1.17.0
|
| 25 |
+
torch==2.6.0
|
| 26 |
+
sse-starlette==2.2.1
|
| 27 |
+
typing_extensions==4.12.2
|
| 28 |
+
xgrammar==0.1.16
|
| 29 |
+
psutil==7.0.0
|
| 30 |
+
kiwisolver==1.4.8
|
| 31 |
+
moviepy==2.1.2
|
| 32 |
+
frozenlist==1.5.0
|
| 33 |
+
jiter==0.9.0
|
| 34 |
+
einops==0.8.1
|
| 35 |
+
flash_attn==2.7.4.post1
|
| 36 |
+
PySocks==1.7.1
|
| 37 |
+
regex==2024.11.6
|
| 38 |
+
markdown-it-py==3.0.0
|
| 39 |
+
ruff==0.11.2
|
| 40 |
+
docker-pycreds==0.4.0
|
| 41 |
+
nvidia-nvtx-cu12==12.4.127
|
| 42 |
+
pyparsing==3.2.3
|
| 43 |
+
resampy==0.4.3
|
| 44 |
+
tokenizers==0.21.0
|
| 45 |
+
frechet-audio-distance==0.1.2
|
| 46 |
+
aiohappyeyeballs==2.6.1
|
| 47 |
+
llamafactory==0.9.3.dev0
|
| 48 |
+
msgspec==0.19.0
|
| 49 |
+
httpx==0.28.1
|
| 50 |
+
encodec==0.1.1
|
| 51 |
+
ffmpy==0.5.0
|
| 52 |
+
jsonschema==4.23.0
|
| 53 |
+
imageio-ffmpeg==0.6.0
|
| 54 |
+
mkl_random==1.2.8
|
| 55 |
+
fairscale==0.4.13
|
| 56 |
+
soxr==0.5.0.post1
|
| 57 |
+
lark==1.2.2
|
| 58 |
+
gradio==5.21.0
|
| 59 |
+
absl-py==2.2.1
|
| 60 |
+
dnspython==2.7.0
|
| 61 |
+
networkx==3.4.2
|
| 62 |
+
h5py==3.13.0
|
| 63 |
+
hjson==3.1.0
|
| 64 |
+
tensorboard==2.19.0
|
| 65 |
+
aiosignal==1.3.2
|
| 66 |
+
pip==25.0
|
| 67 |
+
nvidia-cublas-cu12==12.4.5.8
|
| 68 |
+
llguidance==0.7.11
|
| 69 |
+
zipp==3.21.0
|
| 70 |
+
ftfy==6.3.1
|
| 71 |
+
peft==0.15.0
|
| 72 |
+
attrs==25.3.0
|
| 73 |
+
trl==0.9.6
|
| 74 |
+
requests==2.32.3
|
| 75 |
+
progressbar==2.5
|
| 76 |
+
sniffio==1.3.1
|
| 77 |
+
pycountry==24.6.1
|
| 78 |
+
lxml==5.3.1
|
| 79 |
+
starlette==0.46.1
|
| 80 |
+
pytest==7.2.0
|
| 81 |
+
Markdown==3.7
|
| 82 |
+
mdurl==0.1.2
|
| 83 |
+
pyzmq==26.3.0
|
| 84 |
+
safetensors==0.5.3
|
| 85 |
+
opencv-python==4.6.0.66
|
| 86 |
+
prometheus-fastapi-instrumentator==7.1.0
|
| 87 |
+
shellingham==1.5.4
|
| 88 |
+
torchvision==0.21.0
|
| 89 |
+
pluggy==1.5.0
|
| 90 |
+
timm==1.0.15
|
| 91 |
+
multidict==6.2.0
|
| 92 |
+
semantic-version==2.10.0
|
| 93 |
+
airportsdata==20250224
|
| 94 |
+
numba==0.60.0
|
| 95 |
+
MarkupSafe==2.1.5
|
| 96 |
+
pydantic_core==2.33.0
|
| 97 |
+
imageio==2.37.0
|
| 98 |
+
nvidia-nccl-cu12==2.21.5
|
| 99 |
+
dill==0.3.8
|
| 100 |
+
msgpack==1.1.0
|
| 101 |
+
sentry-sdk==2.24.1
|
| 102 |
+
rpds-py==0.24.0
|
| 103 |
+
grpcio==1.71.0
|
| 104 |
+
fastrlock==0.8.3
|
| 105 |
+
python-json-logger==3.3.0
|
| 106 |
+
cffi==1.17.1
|
| 107 |
+
gradio_client==1.7.2
|
| 108 |
+
PyYAML==6.0.2
|
| 109 |
+
tensorboard-data-server==0.7.2
|
| 110 |
+
termcolor==2.5.0
|
| 111 |
+
torchaudio==2.6.0
|
| 112 |
+
triton==3.2.0
|
| 113 |
+
fastapi==0.115.12
|
| 114 |
+
clint==0.5.1
|
| 115 |
+
lazy_loader==0.4
|
| 116 |
+
depyf==0.18.0
|
| 117 |
+
mkl_fft==1.3.11
|
| 118 |
+
annotated-types==0.7.0
|
| 119 |
+
scikit-learn==1.6.1
|
| 120 |
+
wget==3.2
|
| 121 |
+
setuptools==75.8.0
|
| 122 |
+
args==0.1.0
|
| 123 |
+
certifi==2025.1.31
|
| 124 |
+
click==8.1.8
|
| 125 |
+
python-dotenv==1.1.0
|
| 126 |
+
laion_clap==1.1.5
|
| 127 |
+
Pygments==2.19.1
|
| 128 |
+
tomlkit==0.13.2
|
| 129 |
+
idna==3.7
|
| 130 |
+
propcache==0.3.1
|
| 131 |
+
platformdirs==4.3.7
|
| 132 |
+
align-anything==0.0.1.dev0
|
| 133 |
+
ray==2.44.1
|
| 134 |
+
cloudpickle==3.1.1
|
| 135 |
+
deepspeed==0.16.5
|
| 136 |
+
smmap==5.0.2
|
| 137 |
+
distro==1.9.0
|
| 138 |
+
fonttools==4.56.0
|
| 139 |
+
typing-inspection==0.4.0
|
| 140 |
+
braceexpand==0.1.7
|
| 141 |
+
decorator==5.2.1
|
| 142 |
+
diskcache==5.6.3
|
| 143 |
+
yt-dlp==2025.3.27
|
| 144 |
+
shtab==1.7.1
|
| 145 |
+
gguf==0.10.0
|
| 146 |
+
interegular==0.3.3
|
| 147 |
+
compressed-tensors==0.9.2
|
| 148 |
+
pandas==2.2.3
|
| 149 |
+
huggingface-hub==0.29.3
|
| 150 |
+
pyarrow==19.0.1
|
| 151 |
+
lm-format-enforcer==0.10.11
|
| 152 |
+
GitPython==3.1.44
|
| 153 |
+
xxhash==3.5.0
|
| 154 |
+
packaging==24.2
|
| 155 |
+
setproctitle==1.3.5
|
| 156 |
+
llvmlite==0.43.0
|
| 157 |
+
tiktoken==0.9.0
|
| 158 |
+
mpmath==1.3.0
|
| 159 |
+
email_validator==2.2.0
|
| 160 |
+
nvidia-ml-py==12.570.86
|
| 161 |
+
pydantic==2.11.0
|
| 162 |
+
xformers==0.0.29.post2
|
| 163 |
+
httptools==0.6.4
|
| 164 |
+
librosa==0.11.0
|
| 165 |
+
pytorch-fid==0.3.0
|
| 166 |
+
hpsv2==1.2.0
|
| 167 |
+
nvidia-cudnn-cu12==9.1.0.70
|
| 168 |
+
iniconfig==2.1.0
|
| 169 |
+
sympy==1.13.1
|
| 170 |
+
safehttpx==0.1.6
|
| 171 |
+
jsonschema-specifications==2024.10.1
|
| 172 |
+
Jinja2==3.1.6
|
| 173 |
+
tyro==0.8.14
|
| 174 |
+
h11==0.14.0
|
| 175 |
+
aiohttp==3.11.14
|
| 176 |
+
diffusers==0.32.2
|
| 177 |
+
tqdm==4.67.1
|
| 178 |
+
blake3==1.0.4
|
| 179 |
+
vllm==0.8.2
|
| 180 |
+
scipy==1.10.1
|
| 181 |
+
audioread==3.0.1
|
| 182 |
+
proglog==0.1.10
|
| 183 |
+
fire==0.7.0
|
| 184 |
+
sentencepiece==0.2.0
|
| 185 |
+
pytz==2025.2
|
| 186 |
+
nvidia-nvjitlink-cu12==12.4.127
|
| 187 |
+
nvidia-curand-cu12==10.3.5.147
|
| 188 |
+
numpy==1.26.4
|
| 189 |
+
tzdata==2025.2
|
| 190 |
+
python-multipart==0.0.20
|
| 191 |
+
urllib3==2.3.0
|
| 192 |
+
pycryptodomex==3.22.0
|
| 193 |
+
yarl==1.18.3
|
| 194 |
+
outlines==0.1.11
|
| 195 |
+
nvidia-cusolver-cu12==11.6.1.9
|
| 196 |
+
pydub==0.25.1
|
| 197 |
+
mistral_common==1.5.4
|
| 198 |
+
pycparser==2.22
|
| 199 |
+
pytest-split==0.8.0
|
| 200 |
+
datasets==3.4.1
|
| 201 |
+
soundfile==0.13.1
|
| 202 |
+
transformers==4.50.0
|
| 203 |
+
image-reward==1.5
|
| 204 |
+
wcwidth==0.2.13
|
| 205 |
+
nvidia-cuda-nvrtc-cu12==12.4.127
|
| 206 |
+
groovy==0.1.2
|
| 207 |
+
nvidia-cuda-runtime-cu12==12.4.127
|
| 208 |
+
astor==0.8.1
|
| 209 |
+
anyio==4.9.0
|
| 210 |
+
wandb==0.19.8
|
| 211 |
+
joblib==1.4.2
|
| 212 |
+
fsspec==2024.12.0
|
| 213 |
+
accelerate==1.5.2
|
| 214 |
+
py-cpuinfo==9.0.0
|
| 215 |
+
docstring_parser==0.16
|
| 216 |
+
partial-json-parser==0.2.1.1.post5
|
| 217 |
+
nvidia-cusparse-cu12==12.3.1.170
|
| 218 |
+
protobuf==3.20.3
|
| 219 |
+
outlines_core==0.1.26
|
| 220 |
+
nvidia-cufft-cu12==11.2.1.3
|
| 221 |
+
cycler==0.12.1
|
| 222 |
+
uvicorn==0.34.0
|
| 223 |
+
orjson==3.10.16
|
| 224 |
+
av==14.2.0
|
| 225 |
+
Brotli==1.0.9
|
| 226 |
+
cupy-cuda12x==13.4.1
|
| 227 |
+
openai==1.69.0
|
| 228 |
+
rich==13.9.4
|
| 229 |
+
importlib_metadata==8.6.1
|
| 230 |
+
ninja==1.11.1.4
|
| 231 |
+
wheel==0.45.1
|
| 232 |
+
pillow==10.4.0
|
| 233 |
+
prometheus_client==0.21.1
|
| 234 |
+
threadpoolctl==3.6.0
|
| 235 |
+
gitdb==4.0.12
|
| 236 |
+
watchfiles==1.0.4
|
| 237 |
+
nvidia-cuda-cupti-cu12==12.4.127
|
| 238 |
+
opencv-python-headless==4.11.0.86
|
| 239 |
+
mkl-service==2.4.0
|
| 240 |
+
rich-toolkit==0.14.0
|
| 241 |
+
fastapi-cli==0.0.7
|
| 242 |
+
llamafactory==0.9.3.dev0
|
| 243 |
+
typing_extensions==4.12.2
|
| 244 |
+
tomli==2.0.1
|
| 245 |
+
zipp==3.19.2
|
| 246 |
+
wheel==0.43.0
|
| 247 |
+
jaraco.text==3.12.1
|
| 248 |
+
packaging==24.2
|
| 249 |
+
autocommand==2.2.2
|
| 250 |
+
jaraco.functools==4.0.1
|
| 251 |
+
jaraco.collections==5.1.0
|
| 252 |
+
platformdirs==4.2.2
|
| 253 |
+
more-itertools==10.3.0
|
| 254 |
+
inflect==7.3.1
|
| 255 |
+
jaraco.context==5.3.0
|
| 256 |
+
typeguard==4.3.0
|
| 257 |
+
backports.tarfile==1.2.0
|
| 258 |
+
importlib_metadata==8.0.0
|
wandb/run-20250331_173726-4v7gx39g/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-5.15.0-1040-nvidia-x86_64-with-glibc2.35",
|
| 3 |
+
"python": "CPython 3.11.11",
|
| 4 |
+
"startedAt": "2025-03-31T09:37:27.001749Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"--local_rank=0",
|
| 7 |
+
"--model_name_or_path",
|
| 8 |
+
"/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf",
|
| 9 |
+
"--train_datasets",
|
| 10 |
+
"/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10",
|
| 11 |
+
"--train_template",
|
| 12 |
+
"AA_TI2T_LLAVA",
|
| 13 |
+
"--train_split",
|
| 14 |
+
"train",
|
| 15 |
+
"--train_name",
|
| 16 |
+
"text-image-to-text",
|
| 17 |
+
"--output_dir",
|
| 18 |
+
"../outputs/llava_1.6_mistral_7b_val/top1-10",
|
| 19 |
+
"--save_total_limit",
|
| 20 |
+
"12",
|
| 21 |
+
"--train_batch_size",
|
| 22 |
+
"8",
|
| 23 |
+
"--epochs",
|
| 24 |
+
"3"
|
| 25 |
+
],
|
| 26 |
+
"program": "-m align_anything.trainers.text_image_to_text.dpo",
|
| 27 |
+
"git": {
|
| 28 |
+
"remote": "git@github.com-hantao:PKU-Alignment/align-anything.git",
|
| 29 |
+
"commit": "106588f9802757a3283c1aff1f33ea9afd737f31"
|
| 30 |
+
},
|
| 31 |
+
"email": "2200017789@stu.pku.edu.cn",
|
| 32 |
+
"root": "../outputs/llava_1.6_mistral_7b_val/top1-10",
|
| 33 |
+
"host": "dgx-092",
|
| 34 |
+
"executable": "/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/python",
|
| 35 |
+
"cpu_count": 112,
|
| 36 |
+
"cpu_count_logical": 224,
|
| 37 |
+
"gpu": "NVIDIA H800",
|
| 38 |
+
"gpu_count": 8,
|
| 39 |
+
"disk": {
|
| 40 |
+
"/": {
|
| 41 |
+
"total": "1888556142592",
|
| 42 |
+
"used": "149000749056"
|
| 43 |
+
}
|
| 44 |
+
},
|
| 45 |
+
"memory": {
|
| 46 |
+
"total": "2164195454976"
|
| 47 |
+
},
|
| 48 |
+
"cpu": {
|
| 49 |
+
"count": 112,
|
| 50 |
+
"countLogical": 224
|
| 51 |
+
},
|
| 52 |
+
"gpu_nvidia": [
|
| 53 |
+
{
|
| 54 |
+
"name": "NVIDIA H800",
|
| 55 |
+
"memoryTotal": "85520809984",
|
| 56 |
+
"cudaCores": 16896,
|
| 57 |
+
"architecture": "Hopper"
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"name": "NVIDIA H800",
|
| 61 |
+
"memoryTotal": "85520809984",
|
| 62 |
+
"cudaCores": 16896,
|
| 63 |
+
"architecture": "Hopper"
|
| 64 |
+
},
|
| 65 |
+
{
|
| 66 |
+
"name": "NVIDIA H800",
|
| 67 |
+
"memoryTotal": "85520809984",
|
| 68 |
+
"cudaCores": 16896,
|
| 69 |
+
"architecture": "Hopper"
|
| 70 |
+
},
|
| 71 |
+
{
|
| 72 |
+
"name": "NVIDIA H800",
|
| 73 |
+
"memoryTotal": "85520809984",
|
| 74 |
+
"cudaCores": 16896,
|
| 75 |
+
"architecture": "Hopper"
|
| 76 |
+
},
|
| 77 |
+
{
|
| 78 |
+
"name": "NVIDIA H800",
|
| 79 |
+
"memoryTotal": "85520809984",
|
| 80 |
+
"cudaCores": 16896,
|
| 81 |
+
"architecture": "Hopper"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "NVIDIA H800",
|
| 85 |
+
"memoryTotal": "85520809984",
|
| 86 |
+
"cudaCores": 16896,
|
| 87 |
+
"architecture": "Hopper"
|
| 88 |
+
},
|
| 89 |
+
{
|
| 90 |
+
"name": "NVIDIA H800",
|
| 91 |
+
"memoryTotal": "85520809984",
|
| 92 |
+
"cudaCores": 16896,
|
| 93 |
+
"architecture": "Hopper"
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"name": "NVIDIA H800",
|
| 97 |
+
"memoryTotal": "85520809984",
|
| 98 |
+
"cudaCores": 16896,
|
| 99 |
+
"architecture": "Hopper"
|
| 100 |
+
}
|
| 101 |
+
],
|
| 102 |
+
"slurm": {
|
| 103 |
+
"conf": "/cm/shared/apps/slurm/var/etc/slurm/slurm.conf"
|
| 104 |
+
},
|
| 105 |
+
"cudaVersion": "12.2"
|
| 106 |
+
}
|
wandb/run-20250331_173726-4v7gx39g/logs/debug-core.log
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:37:26.427202666+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpp74thbzm/port-1807771.txt","pid":1807771,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false}
|
| 2 |
+
{"time":"2025-03-31T17:37:26.428083467+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1807771}
|
| 3 |
+
{"time":"2025-03-31T17:37:26.428083605+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":44573,"Zone":""}}
|
| 4 |
+
{"time":"2025-03-31T17:37:26.607453967+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:58146"}
|
| 5 |
+
{"time":"2025-03-31T17:37:27.003133886+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"4v7gx39g","id":"127.0.0.1:58146"}
|
| 6 |
+
{"time":"2025-03-31T17:37:27.321009034+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"4v7gx39g","id":"127.0.0.1:58146"}
|
| 7 |
+
{"time":"2025-03-31T17:38:18.696367064+08:00","level":"INFO","msg":"received shutdown signal","signal":15}
|
wandb/run-20250331_173726-4v7gx39g/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:37:27.005000621+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.8","symlink path":"../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173726-4v7gx39g/logs/debug-core.log"}
|
| 2 |
+
{"time":"2025-03-31T17:37:27.320946444+08:00","level":"INFO","msg":"created new stream","id":"4v7gx39g"}
|
| 3 |
+
{"time":"2025-03-31T17:37:27.321000682+08:00","level":"INFO","msg":"stream: started","id":"4v7gx39g"}
|
| 4 |
+
{"time":"2025-03-31T17:37:27.321021439+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"4v7gx39g"}
|
| 5 |
+
{"time":"2025-03-31T17:37:27.32102533+08:00","level":"INFO","msg":"handler: started","stream_id":"4v7gx39g"}
|
| 6 |
+
{"time":"2025-03-31T17:37:27.321061059+08:00","level":"INFO","msg":"sender: started","stream_id":"4v7gx39g"}
|
| 7 |
+
{"time":"2025-03-31T17:37:27.630894994+08:00","level":"INFO","msg":"Starting system monitor"}
|
wandb/run-20250331_173726-4v7gx39g/logs/debug.log
ADDED
|
@@ -0,0 +1,22 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-03-31 17:37:26,996 INFO MainThread:1807771 [wandb_setup.py:_flush():67] Current SDK version is 0.19.8
|
| 2 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_setup.py:_flush():67] Configure stats pid to 1807771
|
| 3 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_setup.py:_flush():67] Loading settings from /home/yangyaodong/.config/wandb/settings
|
| 4 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_setup.py:_flush():67] Loading settings from /aifs4su/yaodong/hantao/align-anything/scripts/wandb/settings
|
| 5 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_setup.py:_flush():67] Loading settings from environment variables
|
| 6 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_init.py:setup_run_log_directory():647] Logging user logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173726-4v7gx39g/logs/debug.log
|
| 7 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_init.py:setup_run_log_directory():648] Logging internal logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173726-4v7gx39g/logs/debug-internal.log
|
| 8 |
+
2025-03-31 17:37:26,997 INFO MainThread:1807771 [wandb_init.py:init():761] calling init triggers
|
| 9 |
+
2025-03-31 17:37:26,998 INFO MainThread:1807771 [wandb_init.py:init():766] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'train_cfgs': {'save_checkpoint': False, 'load_checkpoint': False, 'ds_cfgs': 'ds_z3_config.json', 'epochs': 3, 'seed': 42, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 1, 'gradient_checkpointing': True, 'learning_rate': 1e-06, 'lr_scheduler_type': 'cosine', 'lr_warmup_ratio': 0.03, 'weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'bf16': True, 'fp16': False, 'eval_strategy': 'epoch', 'eval_interval': 10, 'regularization': 0.001, 'scale_coeff': 0.1, 'freeze_mm_proj': False, 'freeze_vision_tower': True, 'freeze_language_model': False}, 'data_cfgs': {'train_datasets': '/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10', 'train_template': 'AA_TI2T_LLAVA', 'train_size': {}, 'train_split': 'train', 'train_name': 'text-image-to-text', 'train_data_files': {}, 'train_optional_args': [], 'eval_datasets': {}, 'eval_template': {}, 'eval_size': {}, 'eval_split': {}, 'eval_subset': {}, 'eval_data_files': {}, 'eval_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'dpo', 'output_dir': '../outputs/llava_1.6_mistral_7b_val/top1-10', 'cache_dir': {}, 'save_total_limit': 12}, 'model_cfgs': {'model_name_or_path': '/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf', 'trust_remote_code': True, 'model_max_length': 4096}, 'special_tokens': {}, '_wandb': {}}
|
| 11 |
+
2025-03-31 17:37:26,998 INFO MainThread:1807771 [wandb_init.py:init():784] starting backend
|
| 12 |
+
2025-03-31 17:37:26,998 INFO MainThread:1807771 [wandb_init.py:init():788] sending inform_init request
|
| 13 |
+
2025-03-31 17:37:27,001 INFO MainThread:1807771 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
| 14 |
+
2025-03-31 17:37:27,001 INFO MainThread:1807771 [wandb_init.py:init():798] backend started and connected
|
| 15 |
+
2025-03-31 17:37:27,002 INFO MainThread:1807771 [wandb_init.py:init():891] updated telemetry
|
| 16 |
+
2025-03-31 17:37:27,015 INFO MainThread:1807771 [wandb_init.py:init():915] communicating run to backend with 90.0 second timeout
|
| 17 |
+
2025-03-31 17:37:27,625 INFO MainThread:1807771 [wandb_init.py:init():990] starting run threads in backend
|
| 18 |
+
2025-03-31 17:37:27,862 INFO MainThread:1807771 [wandb_run.py:_console_start():2375] atexit reg
|
| 19 |
+
2025-03-31 17:37:27,862 INFO MainThread:1807771 [wandb_run.py:_redirect():2227] redirect: wrap_raw
|
| 20 |
+
2025-03-31 17:37:27,862 INFO MainThread:1807771 [wandb_run.py:_redirect():2292] Wrapping output streams.
|
| 21 |
+
2025-03-31 17:37:27,862 INFO MainThread:1807771 [wandb_run.py:_redirect():2315] Redirects installed.
|
| 22 |
+
2025-03-31 17:37:27,865 INFO MainThread:1807771 [wandb_init.py:init():1032] run started, returning control to user process
|
wandb/run-20250331_173726-4v7gx39g/run-4v7gx39g.wandb
ADDED
|
File without changes
|
wandb/run-20250331_173929-oflhglry/files/config.yaml
ADDED
|
@@ -0,0 +1,95 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
_wandb:
|
| 2 |
+
value:
|
| 3 |
+
cli_version: 0.19.8
|
| 4 |
+
m: []
|
| 5 |
+
python_version: 3.11.11
|
| 6 |
+
t:
|
| 7 |
+
"1":
|
| 8 |
+
- 1
|
| 9 |
+
- 5
|
| 10 |
+
- 11
|
| 11 |
+
- 41
|
| 12 |
+
- 49
|
| 13 |
+
- 51
|
| 14 |
+
- 53
|
| 15 |
+
- 55
|
| 16 |
+
- 63
|
| 17 |
+
- 71
|
| 18 |
+
- 83
|
| 19 |
+
- 98
|
| 20 |
+
- 105
|
| 21 |
+
"2":
|
| 22 |
+
- 1
|
| 23 |
+
- 5
|
| 24 |
+
- 11
|
| 25 |
+
- 41
|
| 26 |
+
- 49
|
| 27 |
+
- 51
|
| 28 |
+
- 53
|
| 29 |
+
- 55
|
| 30 |
+
- 63
|
| 31 |
+
- 71
|
| 32 |
+
- 83
|
| 33 |
+
- 98
|
| 34 |
+
- 105
|
| 35 |
+
"3":
|
| 36 |
+
- 2
|
| 37 |
+
- 13
|
| 38 |
+
- 16
|
| 39 |
+
- 23
|
| 40 |
+
- 55
|
| 41 |
+
- 61
|
| 42 |
+
"4": 3.11.11
|
| 43 |
+
"5": 0.19.8
|
| 44 |
+
"6": 4.50.0
|
| 45 |
+
"8":
|
| 46 |
+
- 5
|
| 47 |
+
"12": 0.19.8
|
| 48 |
+
"13": linux-x86_64
|
| 49 |
+
data_cfgs:
|
| 50 |
+
value:
|
| 51 |
+
eval_optional_args: []
|
| 52 |
+
train_datasets: /aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10
|
| 53 |
+
train_name: text-image-to-text
|
| 54 |
+
train_optional_args: []
|
| 55 |
+
train_split: train
|
| 56 |
+
train_template: AA_TI2T_LLAVA
|
| 57 |
+
logger_cfgs:
|
| 58 |
+
value:
|
| 59 |
+
log_project: align-anything
|
| 60 |
+
log_run_name: dpo
|
| 61 |
+
log_type: wandb
|
| 62 |
+
output_dir: ../outputs/llava_1.6_mistral_7b_val/top1-10
|
| 63 |
+
save_total_limit: 3
|
| 64 |
+
model_cfgs:
|
| 65 |
+
value:
|
| 66 |
+
model_max_length: 4096
|
| 67 |
+
model_name_or_path: /aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf
|
| 68 |
+
trust_remote_code: true
|
| 69 |
+
train_cfgs:
|
| 70 |
+
value:
|
| 71 |
+
adam_betas:
|
| 72 |
+
- 0.9
|
| 73 |
+
- 0.95
|
| 74 |
+
bf16: true
|
| 75 |
+
ds_cfgs: ds_z3_config.json
|
| 76 |
+
epochs: 3
|
| 77 |
+
eval_interval: 10
|
| 78 |
+
eval_strategy: epoch
|
| 79 |
+
fp16: false
|
| 80 |
+
freeze_language_model: false
|
| 81 |
+
freeze_mm_proj: false
|
| 82 |
+
freeze_vision_tower: true
|
| 83 |
+
gradient_accumulation_steps: 1
|
| 84 |
+
gradient_checkpointing: true
|
| 85 |
+
learning_rate: 1e-06
|
| 86 |
+
load_checkpoint: false
|
| 87 |
+
lr_scheduler_type: cosine
|
| 88 |
+
lr_warmup_ratio: 0.03
|
| 89 |
+
per_device_eval_batch_size: 1
|
| 90 |
+
per_device_train_batch_size: 1
|
| 91 |
+
regularization: 0.001
|
| 92 |
+
save_checkpoint: false
|
| 93 |
+
scale_coeff: 0.1
|
| 94 |
+
seed: 42
|
| 95 |
+
weight_decay: 0
|
wandb/run-20250331_173929-oflhglry/files/output.log
ADDED
|
@@ -0,0 +1,58 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
***** Running training *****
|
| 2 |
+
Resuming from checkpoint 1/3 epoch : 0%| | 0/72 [00:00<?, ?it/s]/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/torch/utils/checkpoint.py:87: UserWarning: None of the inputs have requires_grad=True. Gradients will be None
|
| 3 |
+
warnings.warn(
|
| 4 |
+
`use_cache=True` is incompatible with gradient checkpointing. Setting `use_cache=False`.
|
| 5 |
+
Training 1/3 epoch (loss 0.6211): 7%|████████▌ | 5/72 [00:18<03:14, 2.91s/it]Token indices sequence length is longer than the specified maximum sequence length for this model (4205 > 4096). Running this sequence through the model will result in indexing errors
|
| 6 |
+
Training 3/3 epoch (loss 0.0005): 82%|████████████████████████████████████████████████████████████████████████████████████████████████████▊ | 59/72 [03:15<00:35, 2.70s/it]
|
| 7 |
+
[2025-03-31 17:40:01,717] [INFO] [logging.py:107:log_dist] [Rank 0] step=10, skipped=0, lr=[9.681174353198686e-07, 9.681174353198686e-07], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 8 |
+
[2025-03-31 17:40:01,718] [INFO] [timer.py:264:stop] epoch=0/micro_step=10/global_step=10, RunningAvgSamplesPerSec=3.722313847981008, CurrSamplesPerSec=2.5449022257494556, MemAllocated=16.38GB, MaxMemAllocated=21.84GB
|
| 9 |
+
[2025-03-31 17:40:23,917] [INFO] [logging.py:107:log_dist] [Rank 0] step=20, skipped=0, lr=[8.455313244934324e-07, 8.455313244934324e-07], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 10 |
+
[2025-03-31 17:40:23,918] [INFO] [timer.py:264:stop] epoch=0/micro_step=20/global_step=20, RunningAvgSamplesPerSec=4.122870965859232, CurrSamplesPerSec=4.743181933673598, MemAllocated=16.38GB, MaxMemAllocated=21.84GB
|
| 11 |
+
Saving checkpoint at step 24 ...
|
| 12 |
+
Saving model to "../outputs/llava_1.6_mistral_7b_val/top1-10/slice_24" ...
|
| 13 |
+
Saving 16-bit model...
|
| 14 |
+
[2025-03-31 17:40:42,727] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step24 is about to be saved!
|
| 15 |
+
[2025-03-31 17:40:42,727] [INFO] [engine.py:3831:save_16bit_model] Saving model weights to ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_24/pytorch_model.bin, tag: global_step24
|
| 16 |
+
[2025-03-31 17:40:42,727] [INFO] [torch_checkpoint_engine.py:21:save] [Torch] Saving ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_24/pytorch_model.bin...
|
| 17 |
+
[2025-03-31 17:40:57,986] [INFO] [torch_checkpoint_engine.py:23:save] [Torch] Saved ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_24/pytorch_model.bin.
|
| 18 |
+
[2025-03-31 17:40:57,987] [INFO] [torch_checkpoint_engine.py:33:commit] [Torch] Checkpoint global_step24 is ready now!
|
| 19 |
+
Model saved!
|
| 20 |
+
Checkpoint saved.
|
| 21 |
+
[2025-03-31 17:41:12,404] [INFO] [logging.py:107:log_dist] [Rank 0] step=30, skipped=0, lr=[6.545084971874736e-07, 6.545084971874736e-07], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 22 |
+
[2025-03-31 17:41:12,405] [INFO] [timer.py:264:stop] epoch=1/micro_step=6/global_step=30, RunningAvgSamplesPerSec=4.02456931626677, CurrSamplesPerSec=3.240653361782424, MemAllocated=16.38GB, MaxMemAllocated=21.84GB
|
| 23 |
+
[2025-03-31 17:41:36,676] [INFO] [logging.py:107:log_dist] [Rank 0] step=40, skipped=0, lr=[4.328833670911724e-07, 4.328833670911724e-07], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 24 |
+
[2025-03-31 17:41:36,677] [INFO] [timer.py:264:stop] epoch=1/micro_step=16/global_step=40, RunningAvgSamplesPerSec=4.025850941473662, CurrSamplesPerSec=4.080984649502892, MemAllocated=16.38GB, MaxMemAllocated=21.84GB
|
| 25 |
+
Saving checkpoint at step 48 ...
|
| 26 |
+
Saving model to "../outputs/llava_1.6_mistral_7b_val/top1-10/slice_48" ...
|
| 27 |
+
Saving 16-bit model...
|
| 28 |
+
[2025-03-31 17:42:04,512] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step48 is about to be saved!
|
| 29 |
+
[2025-03-31 17:42:04,513] [INFO] [engine.py:3831:save_16bit_model] Saving model weights to ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_48/pytorch_model.bin, tag: global_step48
|
| 30 |
+
[2025-03-31 17:42:04,514] [INFO] [torch_checkpoint_engine.py:21:save] [Torch] Saving ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_48/pytorch_model.bin...
|
| 31 |
+
[2025-03-31 17:42:18,415] [INFO] [torch_checkpoint_engine.py:23:save] [Torch] Saved ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_48/pytorch_model.bin.
|
| 32 |
+
[2025-03-31 17:42:18,416] [INFO] [torch_checkpoint_engine.py:33:commit] [Torch] Checkpoint global_step48 is ready now!
|
| 33 |
+
Model saved!
|
| 34 |
+
Checkpoint saved.
|
| 35 |
+
[2025-03-31 17:42:22,734] [INFO] [logging.py:107:log_dist] [Rank 0] step=50, skipped=0, lr=[2.2455150927394878e-07, 2.2455150927394878e-07], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 36 |
+
[2025-03-31 17:42:22,735] [INFO] [timer.py:264:stop] epoch=2/micro_step=2/global_step=50, RunningAvgSamplesPerSec=4.051799326557334, CurrSamplesPerSec=5.4580888670132115, MemAllocated=16.37GB, MaxMemAllocated=21.84GB
|
| 37 |
+
[2025-03-31 17:42:47,922] [INFO] [logging.py:107:log_dist] [Rank 0] step=60, skipped=0, lr=[7.077560319906694e-08, 7.077560319906694e-08], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 38 |
+
[2025-03-31 17:42:47,923] [INFO] [timer.py:264:stop] epoch=2/micro_step=12/global_step=60, RunningAvgSamplesPerSec=4.025901311666018, CurrSamplesPerSec=3.85640209809824, MemAllocated=16.38GB, MaxMemAllocated=21.84GB
|
| 39 |
+
[2025-03-31 17:43:11,317] [INFO] [logging.py:107:log_dist] [Rank 0] step=70, skipped=0, lr=[2.0128530023804656e-09, 2.0128530023804656e-09], mom=[[0.9, 0.95], [0.9, 0.95]]
|
| 40 |
+
[2025-03-31 17:43:11,317] [INFO] [timer.py:264:stop] epoch=2/micro_step=22/global_step=70, RunningAvgSamplesPerSec=4.050512963952101, CurrSamplesPerSec=4.751966789154519, MemAllocated=16.38GB, MaxMemAllocated=21.84GB
|
| 41 |
+
Saving checkpoint at step 72 ...
|
| 42 |
+
Saving model to "../outputs/llava_1.6_mistral_7b_val/top1-10/slice_72" ...
|
| 43 |
+
Saving 16-bit model...
|
| 44 |
+
[2025-03-31 17:43:25,067] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step72 is about to be saved!
|
| 45 |
+
[2025-03-31 17:43:25,068] [INFO] [engine.py:3831:save_16bit_model] Saving model weights to ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_72/pytorch_model.bin, tag: global_step72
|
| 46 |
+
[2025-03-31 17:43:25,068] [INFO] [torch_checkpoint_engine.py:21:save] [Torch] Saving ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_72/pytorch_model.bin...
|
| 47 |
+
[2025-03-31 17:43:39,343] [INFO] [torch_checkpoint_engine.py:23:save] [Torch] Saved ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_72/pytorch_model.bin.
|
| 48 |
+
[2025-03-31 17:43:39,344] [INFO] [torch_checkpoint_engine.py:33:commit] [Torch] Checkpoint global_step72 is ready now!
|
| 49 |
+
Model saved!
|
| 50 |
+
Checkpoint saved.
|
| 51 |
+
Saving model to "../outputs/llava_1.6_mistral_7b_val/top1-10/slice_end" ...
|
| 52 |
+
Saving 16-bit model...
|
| 53 |
+
[2025-03-31 17:43:47,490] [INFO] [logging.py:107:log_dist] [Rank 0] [Torch] Checkpoint global_step72 is about to be saved!
|
| 54 |
+
[2025-03-31 17:43:47,491] [INFO] [engine.py:3831:save_16bit_model] Saving model weights to ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_end/pytorch_model.bin, tag: global_step72
|
| 55 |
+
[2025-03-31 17:43:47,491] [INFO] [torch_checkpoint_engine.py:21:save] [Torch] Saving ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_end/pytorch_model.bin...
|
| 56 |
+
[2025-03-31 17:44:01,346] [INFO] [torch_checkpoint_engine.py:23:save] [Torch] Saved ../outputs/llava_1.6_mistral_7b_val/top1-10/slice_end/pytorch_model.bin.
|
| 57 |
+
[2025-03-31 17:44:01,347] [INFO] [torch_checkpoint_engine.py:33:commit] [Torch] Checkpoint global_step72 is ready now!
|
| 58 |
+
Model saved!
|
wandb/run-20250331_173929-oflhglry/files/requirements.txt
ADDED
|
@@ -0,0 +1,258 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
maskrcnn_benchmark==0.0.0
|
| 2 |
+
webdataset==0.2.111
|
| 3 |
+
websockets==15.0.1
|
| 4 |
+
typer==0.15.2
|
| 5 |
+
blobfile==3.0.0
|
| 6 |
+
pooch==1.8.2
|
| 7 |
+
filelock==3.18.0
|
| 8 |
+
referencing==0.36.2
|
| 9 |
+
matplotlib==3.10.1
|
| 10 |
+
cachetools==5.5.2
|
| 11 |
+
python-dateutil==2.9.0.post0
|
| 12 |
+
gmpy2==2.2.1
|
| 13 |
+
uvloop==0.21.0
|
| 14 |
+
nvidia-cusparselt-cu12==0.6.2
|
| 15 |
+
clip==0.2.0
|
| 16 |
+
httpcore==1.0.7
|
| 17 |
+
charset-normalizer==3.3.2
|
| 18 |
+
torchlibrosa==0.1.0
|
| 19 |
+
contourpy==1.3.1
|
| 20 |
+
multiprocess==0.70.16
|
| 21 |
+
nest-asyncio==1.6.0
|
| 22 |
+
Werkzeug==3.1.3
|
| 23 |
+
aiofiles==23.2.1
|
| 24 |
+
six==1.17.0
|
| 25 |
+
torch==2.6.0
|
| 26 |
+
sse-starlette==2.2.1
|
| 27 |
+
typing_extensions==4.12.2
|
| 28 |
+
xgrammar==0.1.16
|
| 29 |
+
psutil==7.0.0
|
| 30 |
+
kiwisolver==1.4.8
|
| 31 |
+
moviepy==2.1.2
|
| 32 |
+
frozenlist==1.5.0
|
| 33 |
+
jiter==0.9.0
|
| 34 |
+
einops==0.8.1
|
| 35 |
+
flash_attn==2.7.4.post1
|
| 36 |
+
PySocks==1.7.1
|
| 37 |
+
regex==2024.11.6
|
| 38 |
+
markdown-it-py==3.0.0
|
| 39 |
+
ruff==0.11.2
|
| 40 |
+
docker-pycreds==0.4.0
|
| 41 |
+
nvidia-nvtx-cu12==12.4.127
|
| 42 |
+
pyparsing==3.2.3
|
| 43 |
+
resampy==0.4.3
|
| 44 |
+
tokenizers==0.21.0
|
| 45 |
+
frechet-audio-distance==0.1.2
|
| 46 |
+
aiohappyeyeballs==2.6.1
|
| 47 |
+
llamafactory==0.9.3.dev0
|
| 48 |
+
msgspec==0.19.0
|
| 49 |
+
httpx==0.28.1
|
| 50 |
+
encodec==0.1.1
|
| 51 |
+
ffmpy==0.5.0
|
| 52 |
+
jsonschema==4.23.0
|
| 53 |
+
imageio-ffmpeg==0.6.0
|
| 54 |
+
mkl_random==1.2.8
|
| 55 |
+
fairscale==0.4.13
|
| 56 |
+
soxr==0.5.0.post1
|
| 57 |
+
lark==1.2.2
|
| 58 |
+
gradio==5.21.0
|
| 59 |
+
absl-py==2.2.1
|
| 60 |
+
dnspython==2.7.0
|
| 61 |
+
networkx==3.4.2
|
| 62 |
+
h5py==3.13.0
|
| 63 |
+
hjson==3.1.0
|
| 64 |
+
tensorboard==2.19.0
|
| 65 |
+
aiosignal==1.3.2
|
| 66 |
+
pip==25.0
|
| 67 |
+
nvidia-cublas-cu12==12.4.5.8
|
| 68 |
+
llguidance==0.7.11
|
| 69 |
+
zipp==3.21.0
|
| 70 |
+
ftfy==6.3.1
|
| 71 |
+
peft==0.15.0
|
| 72 |
+
attrs==25.3.0
|
| 73 |
+
trl==0.9.6
|
| 74 |
+
requests==2.32.3
|
| 75 |
+
progressbar==2.5
|
| 76 |
+
sniffio==1.3.1
|
| 77 |
+
pycountry==24.6.1
|
| 78 |
+
lxml==5.3.1
|
| 79 |
+
starlette==0.46.1
|
| 80 |
+
pytest==7.2.0
|
| 81 |
+
Markdown==3.7
|
| 82 |
+
mdurl==0.1.2
|
| 83 |
+
pyzmq==26.3.0
|
| 84 |
+
safetensors==0.5.3
|
| 85 |
+
opencv-python==4.6.0.66
|
| 86 |
+
prometheus-fastapi-instrumentator==7.1.0
|
| 87 |
+
shellingham==1.5.4
|
| 88 |
+
torchvision==0.21.0
|
| 89 |
+
pluggy==1.5.0
|
| 90 |
+
timm==1.0.15
|
| 91 |
+
multidict==6.2.0
|
| 92 |
+
semantic-version==2.10.0
|
| 93 |
+
airportsdata==20250224
|
| 94 |
+
numba==0.60.0
|
| 95 |
+
MarkupSafe==2.1.5
|
| 96 |
+
pydantic_core==2.33.0
|
| 97 |
+
imageio==2.37.0
|
| 98 |
+
nvidia-nccl-cu12==2.21.5
|
| 99 |
+
dill==0.3.8
|
| 100 |
+
msgpack==1.1.0
|
| 101 |
+
sentry-sdk==2.24.1
|
| 102 |
+
rpds-py==0.24.0
|
| 103 |
+
grpcio==1.71.0
|
| 104 |
+
fastrlock==0.8.3
|
| 105 |
+
python-json-logger==3.3.0
|
| 106 |
+
cffi==1.17.1
|
| 107 |
+
gradio_client==1.7.2
|
| 108 |
+
PyYAML==6.0.2
|
| 109 |
+
tensorboard-data-server==0.7.2
|
| 110 |
+
termcolor==2.5.0
|
| 111 |
+
torchaudio==2.6.0
|
| 112 |
+
triton==3.2.0
|
| 113 |
+
fastapi==0.115.12
|
| 114 |
+
clint==0.5.1
|
| 115 |
+
lazy_loader==0.4
|
| 116 |
+
depyf==0.18.0
|
| 117 |
+
mkl_fft==1.3.11
|
| 118 |
+
annotated-types==0.7.0
|
| 119 |
+
scikit-learn==1.6.1
|
| 120 |
+
wget==3.2
|
| 121 |
+
setuptools==75.8.0
|
| 122 |
+
args==0.1.0
|
| 123 |
+
certifi==2025.1.31
|
| 124 |
+
click==8.1.8
|
| 125 |
+
python-dotenv==1.1.0
|
| 126 |
+
laion_clap==1.1.5
|
| 127 |
+
Pygments==2.19.1
|
| 128 |
+
tomlkit==0.13.2
|
| 129 |
+
idna==3.7
|
| 130 |
+
propcache==0.3.1
|
| 131 |
+
platformdirs==4.3.7
|
| 132 |
+
align-anything==0.0.1.dev0
|
| 133 |
+
ray==2.44.1
|
| 134 |
+
cloudpickle==3.1.1
|
| 135 |
+
deepspeed==0.16.5
|
| 136 |
+
smmap==5.0.2
|
| 137 |
+
distro==1.9.0
|
| 138 |
+
fonttools==4.56.0
|
| 139 |
+
typing-inspection==0.4.0
|
| 140 |
+
braceexpand==0.1.7
|
| 141 |
+
decorator==5.2.1
|
| 142 |
+
diskcache==5.6.3
|
| 143 |
+
yt-dlp==2025.3.27
|
| 144 |
+
shtab==1.7.1
|
| 145 |
+
gguf==0.10.0
|
| 146 |
+
interegular==0.3.3
|
| 147 |
+
compressed-tensors==0.9.2
|
| 148 |
+
pandas==2.2.3
|
| 149 |
+
huggingface-hub==0.29.3
|
| 150 |
+
pyarrow==19.0.1
|
| 151 |
+
lm-format-enforcer==0.10.11
|
| 152 |
+
GitPython==3.1.44
|
| 153 |
+
xxhash==3.5.0
|
| 154 |
+
packaging==24.2
|
| 155 |
+
setproctitle==1.3.5
|
| 156 |
+
llvmlite==0.43.0
|
| 157 |
+
tiktoken==0.9.0
|
| 158 |
+
mpmath==1.3.0
|
| 159 |
+
email_validator==2.2.0
|
| 160 |
+
nvidia-ml-py==12.570.86
|
| 161 |
+
pydantic==2.11.0
|
| 162 |
+
xformers==0.0.29.post2
|
| 163 |
+
httptools==0.6.4
|
| 164 |
+
librosa==0.11.0
|
| 165 |
+
pytorch-fid==0.3.0
|
| 166 |
+
hpsv2==1.2.0
|
| 167 |
+
nvidia-cudnn-cu12==9.1.0.70
|
| 168 |
+
iniconfig==2.1.0
|
| 169 |
+
sympy==1.13.1
|
| 170 |
+
safehttpx==0.1.6
|
| 171 |
+
jsonschema-specifications==2024.10.1
|
| 172 |
+
Jinja2==3.1.6
|
| 173 |
+
tyro==0.8.14
|
| 174 |
+
h11==0.14.0
|
| 175 |
+
aiohttp==3.11.14
|
| 176 |
+
diffusers==0.32.2
|
| 177 |
+
tqdm==4.67.1
|
| 178 |
+
blake3==1.0.4
|
| 179 |
+
vllm==0.8.2
|
| 180 |
+
scipy==1.10.1
|
| 181 |
+
audioread==3.0.1
|
| 182 |
+
proglog==0.1.10
|
| 183 |
+
fire==0.7.0
|
| 184 |
+
sentencepiece==0.2.0
|
| 185 |
+
pytz==2025.2
|
| 186 |
+
nvidia-nvjitlink-cu12==12.4.127
|
| 187 |
+
nvidia-curand-cu12==10.3.5.147
|
| 188 |
+
numpy==1.26.4
|
| 189 |
+
tzdata==2025.2
|
| 190 |
+
python-multipart==0.0.20
|
| 191 |
+
urllib3==2.3.0
|
| 192 |
+
pycryptodomex==3.22.0
|
| 193 |
+
yarl==1.18.3
|
| 194 |
+
outlines==0.1.11
|
| 195 |
+
nvidia-cusolver-cu12==11.6.1.9
|
| 196 |
+
pydub==0.25.1
|
| 197 |
+
mistral_common==1.5.4
|
| 198 |
+
pycparser==2.22
|
| 199 |
+
pytest-split==0.8.0
|
| 200 |
+
datasets==3.4.1
|
| 201 |
+
soundfile==0.13.1
|
| 202 |
+
transformers==4.50.0
|
| 203 |
+
image-reward==1.5
|
| 204 |
+
wcwidth==0.2.13
|
| 205 |
+
nvidia-cuda-nvrtc-cu12==12.4.127
|
| 206 |
+
groovy==0.1.2
|
| 207 |
+
nvidia-cuda-runtime-cu12==12.4.127
|
| 208 |
+
astor==0.8.1
|
| 209 |
+
anyio==4.9.0
|
| 210 |
+
wandb==0.19.8
|
| 211 |
+
joblib==1.4.2
|
| 212 |
+
fsspec==2024.12.0
|
| 213 |
+
accelerate==1.5.2
|
| 214 |
+
py-cpuinfo==9.0.0
|
| 215 |
+
docstring_parser==0.16
|
| 216 |
+
partial-json-parser==0.2.1.1.post5
|
| 217 |
+
nvidia-cusparse-cu12==12.3.1.170
|
| 218 |
+
protobuf==3.20.3
|
| 219 |
+
outlines_core==0.1.26
|
| 220 |
+
nvidia-cufft-cu12==11.2.1.3
|
| 221 |
+
cycler==0.12.1
|
| 222 |
+
uvicorn==0.34.0
|
| 223 |
+
orjson==3.10.16
|
| 224 |
+
av==14.2.0
|
| 225 |
+
Brotli==1.0.9
|
| 226 |
+
cupy-cuda12x==13.4.1
|
| 227 |
+
openai==1.69.0
|
| 228 |
+
rich==13.9.4
|
| 229 |
+
importlib_metadata==8.6.1
|
| 230 |
+
ninja==1.11.1.4
|
| 231 |
+
wheel==0.45.1
|
| 232 |
+
pillow==10.4.0
|
| 233 |
+
prometheus_client==0.21.1
|
| 234 |
+
threadpoolctl==3.6.0
|
| 235 |
+
gitdb==4.0.12
|
| 236 |
+
watchfiles==1.0.4
|
| 237 |
+
nvidia-cuda-cupti-cu12==12.4.127
|
| 238 |
+
opencv-python-headless==4.11.0.86
|
| 239 |
+
mkl-service==2.4.0
|
| 240 |
+
rich-toolkit==0.14.0
|
| 241 |
+
fastapi-cli==0.0.7
|
| 242 |
+
llamafactory==0.9.3.dev0
|
| 243 |
+
typing_extensions==4.12.2
|
| 244 |
+
tomli==2.0.1
|
| 245 |
+
zipp==3.19.2
|
| 246 |
+
wheel==0.43.0
|
| 247 |
+
jaraco.text==3.12.1
|
| 248 |
+
packaging==24.2
|
| 249 |
+
autocommand==2.2.2
|
| 250 |
+
jaraco.functools==4.0.1
|
| 251 |
+
jaraco.collections==5.1.0
|
| 252 |
+
platformdirs==4.2.2
|
| 253 |
+
more-itertools==10.3.0
|
| 254 |
+
inflect==7.3.1
|
| 255 |
+
jaraco.context==5.3.0
|
| 256 |
+
typeguard==4.3.0
|
| 257 |
+
backports.tarfile==1.2.0
|
| 258 |
+
importlib_metadata==8.0.0
|
wandb/run-20250331_173929-oflhglry/files/wandb-metadata.json
ADDED
|
@@ -0,0 +1,106 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"os": "Linux-5.15.0-1040-nvidia-x86_64-with-glibc2.35",
|
| 3 |
+
"python": "CPython 3.11.11",
|
| 4 |
+
"startedAt": "2025-03-31T09:39:29.355313Z",
|
| 5 |
+
"args": [
|
| 6 |
+
"--local_rank=0",
|
| 7 |
+
"--model_name_or_path",
|
| 8 |
+
"/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf",
|
| 9 |
+
"--train_datasets",
|
| 10 |
+
"/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10",
|
| 11 |
+
"--train_template",
|
| 12 |
+
"AA_TI2T_LLAVA",
|
| 13 |
+
"--train_split",
|
| 14 |
+
"train",
|
| 15 |
+
"--train_name",
|
| 16 |
+
"text-image-to-text",
|
| 17 |
+
"--output_dir",
|
| 18 |
+
"../outputs/llava_1.6_mistral_7b_val/top1-10",
|
| 19 |
+
"--save_total_limit",
|
| 20 |
+
"3",
|
| 21 |
+
"--train_batch_size",
|
| 22 |
+
"8",
|
| 23 |
+
"--epochs",
|
| 24 |
+
"3"
|
| 25 |
+
],
|
| 26 |
+
"program": "-m align_anything.trainers.text_image_to_text.dpo",
|
| 27 |
+
"git": {
|
| 28 |
+
"remote": "git@github.com-hantao:PKU-Alignment/align-anything.git",
|
| 29 |
+
"commit": "106588f9802757a3283c1aff1f33ea9afd737f31"
|
| 30 |
+
},
|
| 31 |
+
"email": "2200017789@stu.pku.edu.cn",
|
| 32 |
+
"root": "../outputs/llava_1.6_mistral_7b_val/top1-10",
|
| 33 |
+
"host": "dgx-092",
|
| 34 |
+
"executable": "/aifs4su/yaodong/miniconda3/envs/hantao_llama/bin/python",
|
| 35 |
+
"cpu_count": 112,
|
| 36 |
+
"cpu_count_logical": 224,
|
| 37 |
+
"gpu": "NVIDIA H800",
|
| 38 |
+
"gpu_count": 8,
|
| 39 |
+
"disk": {
|
| 40 |
+
"/": {
|
| 41 |
+
"total": "1888556142592",
|
| 42 |
+
"used": "149001129984"
|
| 43 |
+
}
|
| 44 |
+
},
|
| 45 |
+
"memory": {
|
| 46 |
+
"total": "2164195454976"
|
| 47 |
+
},
|
| 48 |
+
"cpu": {
|
| 49 |
+
"count": 112,
|
| 50 |
+
"countLogical": 224
|
| 51 |
+
},
|
| 52 |
+
"gpu_nvidia": [
|
| 53 |
+
{
|
| 54 |
+
"name": "NVIDIA H800",
|
| 55 |
+
"memoryTotal": "85520809984",
|
| 56 |
+
"cudaCores": 16896,
|
| 57 |
+
"architecture": "Hopper"
|
| 58 |
+
},
|
| 59 |
+
{
|
| 60 |
+
"name": "NVIDIA H800",
|
| 61 |
+
"memoryTotal": "85520809984",
|
| 62 |
+
"cudaCores": 16896,
|
| 63 |
+
"architecture": "Hopper"
|
| 64 |
+
},
|
| 65 |
+
{
|
| 66 |
+
"name": "NVIDIA H800",
|
| 67 |
+
"memoryTotal": "85520809984",
|
| 68 |
+
"cudaCores": 16896,
|
| 69 |
+
"architecture": "Hopper"
|
| 70 |
+
},
|
| 71 |
+
{
|
| 72 |
+
"name": "NVIDIA H800",
|
| 73 |
+
"memoryTotal": "85520809984",
|
| 74 |
+
"cudaCores": 16896,
|
| 75 |
+
"architecture": "Hopper"
|
| 76 |
+
},
|
| 77 |
+
{
|
| 78 |
+
"name": "NVIDIA H800",
|
| 79 |
+
"memoryTotal": "85520809984",
|
| 80 |
+
"cudaCores": 16896,
|
| 81 |
+
"architecture": "Hopper"
|
| 82 |
+
},
|
| 83 |
+
{
|
| 84 |
+
"name": "NVIDIA H800",
|
| 85 |
+
"memoryTotal": "85520809984",
|
| 86 |
+
"cudaCores": 16896,
|
| 87 |
+
"architecture": "Hopper"
|
| 88 |
+
},
|
| 89 |
+
{
|
| 90 |
+
"name": "NVIDIA H800",
|
| 91 |
+
"memoryTotal": "85520809984",
|
| 92 |
+
"cudaCores": 16896,
|
| 93 |
+
"architecture": "Hopper"
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"name": "NVIDIA H800",
|
| 97 |
+
"memoryTotal": "85520809984",
|
| 98 |
+
"cudaCores": 16896,
|
| 99 |
+
"architecture": "Hopper"
|
| 100 |
+
}
|
| 101 |
+
],
|
| 102 |
+
"slurm": {
|
| 103 |
+
"conf": "/cm/shared/apps/slurm/var/etc/slurm/slurm.conf"
|
| 104 |
+
},
|
| 105 |
+
"cudaVersion": "12.2"
|
| 106 |
+
}
|
wandb/run-20250331_173929-oflhglry/files/wandb-summary.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"train/reward_accuracy":1,"train/better_sample_reward":12.8125,"train/lr":0,"train/loss":2.372264862060547e-05,"_timestamp":1.7434141965928717e+09,"train/reward_margin":23.25,"train/reward":2.53125,"_wandb":{"runtime":272},"train/epoch":3,"train/step":72,"_runtime":272.034614709,"train/worse_sample_reward":-10.25,"_step":72}
|
wandb/run-20250331_173929-oflhglry/logs/debug-core.log
ADDED
|
@@ -0,0 +1,14 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:39:28.778210084+08:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpv0i9nviv/port-1814399.txt","pid":1814399,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false}
|
| 2 |
+
{"time":"2025-03-31T17:39:28.779087954+08:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1814399}
|
| 3 |
+
{"time":"2025-03-31T17:39:28.77907394+08:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":39655,"Zone":""}}
|
| 4 |
+
{"time":"2025-03-31T17:39:28.95893708+08:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:59496"}
|
| 5 |
+
{"time":"2025-03-31T17:39:29.356525972+08:00","level":"INFO","msg":"handleInformInit: received","streamId":"oflhglry","id":"127.0.0.1:59496"}
|
| 6 |
+
{"time":"2025-03-31T17:39:29.573174417+08:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"oflhglry","id":"127.0.0.1:59496"}
|
| 7 |
+
{"time":"2025-03-31T17:44:02.902545963+08:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:59496"}
|
| 8 |
+
{"time":"2025-03-31T17:44:02.902715372+08:00","level":"INFO","msg":"server is shutting down"}
|
| 9 |
+
{"time":"2025-03-31T17:44:02.902714048+08:00","level":"INFO","msg":"connection: closing","id":"127.0.0.1:59496"}
|
| 10 |
+
{"time":"2025-03-31T17:44:02.902792273+08:00","level":"INFO","msg":"connection: closed successfully","id":"127.0.0.1:59496"}
|
| 11 |
+
{"time":"2025-03-31T17:44:03.099880037+08:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write tcp 127.0.0.1:39655->127.0.0.1:59496: use of closed network connection","id":"127.0.0.1:59496"}
|
| 12 |
+
{"time":"2025-03-31T17:44:03.100151996+08:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:59496"}
|
| 13 |
+
{"time":"2025-03-31T17:44:03.100161974+08:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:59496"}
|
| 14 |
+
{"time":"2025-03-31T17:44:03.100166112+08:00","level":"INFO","msg":"server is closed"}
|
wandb/run-20250331_173929-oflhglry/logs/debug-internal.log
ADDED
|
@@ -0,0 +1,16 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"time":"2025-03-31T17:39:29.357857467+08:00","level":"INFO","msg":"stream: starting","core version":"0.19.8","symlink path":"../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173929-oflhglry/logs/debug-core.log"}
|
| 2 |
+
{"time":"2025-03-31T17:39:29.573098876+08:00","level":"INFO","msg":"created new stream","id":"oflhglry"}
|
| 3 |
+
{"time":"2025-03-31T17:39:29.573167483+08:00","level":"INFO","msg":"stream: started","id":"oflhglry"}
|
| 4 |
+
{"time":"2025-03-31T17:39:29.573190711+08:00","level":"INFO","msg":"handler: started","stream_id":"oflhglry"}
|
| 5 |
+
{"time":"2025-03-31T17:39:29.573196411+08:00","level":"INFO","msg":"writer: Do: started","stream_id":"oflhglry"}
|
| 6 |
+
{"time":"2025-03-31T17:39:29.573195051+08:00","level":"INFO","msg":"sender: started","stream_id":"oflhglry"}
|
| 7 |
+
{"time":"2025-03-31T17:39:29.880478815+08:00","level":"INFO","msg":"Starting system monitor"}
|
| 8 |
+
{"time":"2025-03-31T17:44:01.38994239+08:00","level":"INFO","msg":"Stopping system monitor"}
|
| 9 |
+
{"time":"2025-03-31T17:44:01.39093428+08:00","level":"INFO","msg":"Stopped system monitor"}
|
| 10 |
+
{"time":"2025-03-31T17:44:02.852694686+08:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"}
|
| 11 |
+
{"time":"2025-03-31T17:44:02.902712341+08:00","level":"INFO","msg":"stream: closing","id":"oflhglry"}
|
| 12 |
+
{"time":"2025-03-31T17:44:02.902739889+08:00","level":"WARN","msg":"sender: received Exit record more than once, ignoring"}
|
| 13 |
+
{"time":"2025-03-31T17:44:03.099767177+08:00","level":"INFO","msg":"handler: closed","stream_id":"oflhglry"}
|
| 14 |
+
{"time":"2025-03-31T17:44:03.099816798+08:00","level":"INFO","msg":"writer: Close: closed","stream_id":"oflhglry"}
|
| 15 |
+
{"time":"2025-03-31T17:44:03.099920876+08:00","level":"INFO","msg":"sender: closed","stream_id":"oflhglry"}
|
| 16 |
+
{"time":"2025-03-31T17:44:03.100069001+08:00","level":"INFO","msg":"stream: closed","id":"oflhglry"}
|
wandb/run-20250331_173929-oflhglry/logs/debug.log
ADDED
|
@@ -0,0 +1,47 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Current SDK version is 0.19.8
|
| 2 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Configure stats pid to 1814399
|
| 3 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Loading settings from /home/yangyaodong/.config/wandb/settings
|
| 4 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Loading settings from /aifs4su/yaodong/hantao/align-anything/scripts/wandb/settings
|
| 5 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_setup.py:_flush():67] Loading settings from environment variables
|
| 6 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_init.py:setup_run_log_directory():647] Logging user logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173929-oflhglry/logs/debug.log
|
| 7 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_init.py:setup_run_log_directory():648] Logging internal logs to ../outputs/llava_1.6_mistral_7b_val/top1-10/wandb/run-20250331_173929-oflhglry/logs/debug-internal.log
|
| 8 |
+
2025-03-31 17:39:29,349 INFO MainThread:1814399 [wandb_init.py:init():761] calling init triggers
|
| 9 |
+
2025-03-31 17:39:29,350 INFO MainThread:1814399 [wandb_init.py:init():766] wandb.init called with sweep_config: {}
|
| 10 |
+
config: {'train_cfgs': {'save_checkpoint': False, 'load_checkpoint': False, 'ds_cfgs': 'ds_z3_config.json', 'epochs': 3, 'seed': 42, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 1, 'gradient_checkpointing': True, 'learning_rate': 1e-06, 'lr_scheduler_type': 'cosine', 'lr_warmup_ratio': 0.03, 'weight_decay': 0.0, 'adam_betas': [0.9, 0.95], 'bf16': True, 'fp16': False, 'eval_strategy': 'epoch', 'eval_interval': 10, 'regularization': 0.001, 'scale_coeff': 0.1, 'freeze_mm_proj': False, 'freeze_vision_tower': True, 'freeze_language_model': False}, 'data_cfgs': {'train_datasets': '/aifs4su/yaodong/hantao/datasets/AA_preference_cosi/val/merged/top1-10', 'train_template': 'AA_TI2T_LLAVA', 'train_size': {}, 'train_split': 'train', 'train_name': 'text-image-to-text', 'train_data_files': {}, 'train_optional_args': [], 'eval_datasets': {}, 'eval_template': {}, 'eval_size': {}, 'eval_split': {}, 'eval_subset': {}, 'eval_data_files': {}, 'eval_optional_args': []}, 'logger_cfgs': {'log_type': 'wandb', 'log_project': 'align-anything', 'log_run_name': 'dpo', 'output_dir': '../outputs/llava_1.6_mistral_7b_val/top1-10', 'cache_dir': {}, 'save_total_limit': 3}, 'model_cfgs': {'model_name_or_path': '/aifs4su/yaodong/hantao/models/llava-v1.6-mistral-7b-hf', 'trust_remote_code': True, 'model_max_length': 4096}, 'special_tokens': {}, '_wandb': {}}
|
| 11 |
+
2025-03-31 17:39:29,350 INFO MainThread:1814399 [wandb_init.py:init():784] starting backend
|
| 12 |
+
2025-03-31 17:39:29,350 INFO MainThread:1814399 [wandb_init.py:init():788] sending inform_init request
|
| 13 |
+
2025-03-31 17:39:29,355 INFO MainThread:1814399 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn
|
| 14 |
+
2025-03-31 17:39:29,355 INFO MainThread:1814399 [wandb_init.py:init():798] backend started and connected
|
| 15 |
+
2025-03-31 17:39:29,357 INFO MainThread:1814399 [wandb_init.py:init():891] updated telemetry
|
| 16 |
+
2025-03-31 17:39:29,371 INFO MainThread:1814399 [wandb_init.py:init():915] communicating run to backend with 90.0 second timeout
|
| 17 |
+
2025-03-31 17:39:29,874 INFO MainThread:1814399 [wandb_init.py:init():990] starting run threads in backend
|
| 18 |
+
2025-03-31 17:39:30,115 INFO MainThread:1814399 [wandb_run.py:_console_start():2375] atexit reg
|
| 19 |
+
2025-03-31 17:39:30,115 INFO MainThread:1814399 [wandb_run.py:_redirect():2227] redirect: wrap_raw
|
| 20 |
+
2025-03-31 17:39:30,115 INFO MainThread:1814399 [wandb_run.py:_redirect():2292] Wrapping output streams.
|
| 21 |
+
2025-03-31 17:39:30,116 INFO MainThread:1814399 [wandb_run.py:_redirect():2315] Redirects installed.
|
| 22 |
+
2025-03-31 17:39:30,118 INFO MainThread:1814399 [wandb_init.py:init():1032] run started, returning control to user process
|
| 23 |
+
2025-03-31 17:44:01,349 INFO MainThread:1814399 [wandb_run.py:_finish():2112] finishing run htlou/align-anything/oflhglry
|
| 24 |
+
2025-03-31 17:44:01,350 INFO MainThread:1814399 [wandb_run.py:_atexit_cleanup():2340] got exitcode: 0
|
| 25 |
+
2025-03-31 17:44:01,350 INFO MainThread:1814399 [wandb_run.py:_restore():2322] restore
|
| 26 |
+
2025-03-31 17:44:01,351 INFO MainThread:1814399 [wandb_run.py:_restore():2328] restore done
|
| 27 |
+
2025-03-31 17:44:02,351 INFO MainThread:1814399 [wandb_run.py:_restore():2322] restore
|
| 28 |
+
2025-03-31 17:44:02,352 INFO MainThread:1814399 [wandb_run.py:_restore():2328] restore done
|
| 29 |
+
2025-03-31 17:44:02,352 ERROR MainThread:1814399 [wandb_run.py:_atexit_cleanup():2361] Problem finishing run
|
| 30 |
+
Traceback (most recent call last):
|
| 31 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2352, in _atexit_cleanup
|
| 32 |
+
self._on_finish()
|
| 33 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2609, in _on_finish
|
| 34 |
+
wait_with_progress(
|
| 35 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/mailbox/wait_with_progress.py", line 24, in wait_with_progress
|
| 36 |
+
return wait_all_with_progress(
|
| 37 |
+
^^^^^^^^^^^^^^^^^^^^^^^
|
| 38 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/mailbox/wait_with_progress.py", line 87, in wait_all_with_progress
|
| 39 |
+
return asyncio_compat.run(progress_loop_with_timeout)
|
| 40 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 41 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/site-packages/wandb/sdk/lib/asyncio_compat.py", line 27, in run
|
| 42 |
+
future = executor.submit(runner.run, fn)
|
| 43 |
+
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
| 44 |
+
File "/aifs4su/yaodong/miniconda3/envs/hantao_llama/lib/python3.11/concurrent/futures/thread.py", line 169, in submit
|
| 45 |
+
raise RuntimeError('cannot schedule new futures after '
|
| 46 |
+
RuntimeError: cannot schedule new futures after interpreter shutdown
|
| 47 |
+
2025-03-31 17:44:02,902 INFO MsgRouterThr:1814399 [mailbox.py:close():129] Closing mailbox, abandoning 2 handles.
|
wandb/run-20250331_173929-oflhglry/run-oflhglry.wandb
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a26e6c684540f750d23ff07cee6a7862cb1eec1d4e30a0eb65cf9720a19af204
|
| 3 |
+
size 204949
|