diff --git a/.gitattributes b/.gitattributes index a6344aac8c09253b3b630fb776ae94478aa0275b..5baa917cb79db4bb60739232316a2b2fe232802f 100644 --- a/.gitattributes +++ b/.gitattributes @@ -33,3 +33,9 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text +tokenizer.json filter=lfs diff=lfs merge=lfs -text +wandb/run-20250508_192410-bvjuhkx9/run-bvjuhkx9.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250508_195305-eg5dmumm/run-eg5dmumm.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250508_200345-11tfsh2x/run-11tfsh2x.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250508_200812-r9ll1pfp/run-r9ll1pfp.wandb filter=lfs diff=lfs merge=lfs -text +wandb/run-20250508_212154-is8630ik/run-is8630ik.wandb filter=lfs diff=lfs merge=lfs -text diff --git a/README.md b/README.md index 7b95401dc46245ac339fc25059d4a56d90b4cde5..0d31128190920e45b61115944d16e773c2ec94c3 100644 --- a/README.md +++ b/README.md @@ -1,3 +1,202 @@ ---- -license: apache-2.0 ---- +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.15.2 \ No newline at end of file diff --git a/adapter_config.json b/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..7ff702fc023bc6dea1f65204bbc26da513ff19c1 --- /dev/null +++ b/adapter_config.json @@ -0,0 +1,43 @@ +{ + "alpha_pattern": {}, + "auto_mapping": { + "base_model_class": "LlamaForCausalLM", + "parent_library": "transformers.models.llama.modeling_llama" + }, + "base_model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "bias": "none", + "corda_config": null, + "eva_config": null, + "exclude_modules": null, + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 32, + "lora_bias": false, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 16, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "gate_proj", + "up_proj", + "k_proj", + "v_proj", + "o_proj", + "lm_head", + "q_proj", + "down_proj" + ], + "task_type": null, + "trainable_token_indices": null, + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/adapter_model.safetensors b/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..047087c3d7feba1b654b4fa8ece48929310e9742 --- /dev/null +++ b/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f80f97ba737f3c031e76d69dccad49a09d735bfb2e9091172f15f50637eb165 +size 1138873304 diff --git a/arguments.json b/arguments.json new file mode 100644 index 0000000000000000000000000000000000000000..82e333b41396d3d8dbe6bec5ceb81299cf69f1e2 --- /dev/null +++ b/arguments.json @@ -0,0 +1,60 @@ +{ + "model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "recompute_baseline": false, + "cache_dir": "/home/panda/pda-llm/cache/sft-tools", + "max_length": 2048, + "trust_remote_code": true, + "train_datasets": [ + [ + "tools", + { + "proportion": 1.0 + } + ] + ], + "eval_datasets": null, + "safety_ratio_tol": 10.0, + "important_sft": false, + "resilient_coeff": 1.0, + "epochs": 3, + "per_device_train_batch_size": 1, + "per_device_eval_batch_size": 1, + "gradient_accumulation_steps": 48, + "gradient_checkpointing": true, + "lr": 0.0001, + "lr_scheduler_type": "cosine", + "lr_warmup_ratio": 0.1, + "weight_decay": 0.0, + "seed": 42, + "fp16": false, + "bf16": true, + "tf32": false, + "lora_r": 16, + "lora_alpha": 32, + "lora_dropout": 0.05, + "eval_strategy": "epoch", + "eval_interval": 1000000, + "need_eval": true, + "eval_split_ratio": null, + "output_dir": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "log_type": "wandb", + "log_dir": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "log_project": "TOOLS-SFT", + "log_run_name": "tools-sft-2025-05-08-21-21-54", + "save_16bit": false, + "save_interval": 1000000, + "local_rank": 0, + "zero_stage": 0, + "offload": "none", + "deepspeed": false, + "deepspeed_config": null, + "deepscale": false, + "deepscale_config": null, + "global_rank": 0, + "device": { + "type": "torch.device", + "repr": "device(type='cuda', index=0)" + }, + "num_update_steps_per_epoch": 118, + "total_training_steps": 354 +} \ No newline at end of file diff --git a/arguments.pkl b/arguments.pkl new file mode 100644 index 0000000000000000000000000000000000000000..9b82c20ec24f03fb7ec807d239646ecdd7759dbc --- /dev/null +++ b/arguments.pkl @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fafb0bdf1ddeef98c143db9a23a20856b243c468bf8542e9bda3541f03d46598 +size 1225 diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..3c5ae56b84061071c386d25c1a237bba9d9d1ac4 --- /dev/null +++ b/config.json @@ -0,0 +1,37 @@ +{ + "_attn_implementation_autoset": true, + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128009, + "head_dim": 128, + "hidden_act": "silu", + "hidden_size": 4096, + "initializer_range": 0.02, + "intermediate_size": 14336, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 32, + "num_hidden_layers": 32, + "num_key_value_heads": 8, + "pad_token_id": 128256, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "bfloat16", + "transformers_version": "4.51.3", + "use_cache": true, + "vocab_size": 128258 +} diff --git a/environ.txt b/environ.txt new file mode 100644 index 0000000000000000000000000000000000000000..c56f60a26290141251c926217d847567f0508db6 --- /dev/null +++ b/environ.txt @@ -0,0 +1,58 @@ +BROWSER=/home/panda/.vscode-server/cli/servers/Stable-4949701c880d4bdb949e3c0e6b400288da7f474b/server/bin/helpers/browser.sh +COLORTERM=truecolor +CONDA_DEFAULT_ENV=pda +CONDA_EXE=/home/panda/miniconda3/bin/conda +CONDA_PREFIX=/home/panda/miniconda3/envs/pda +CONDA_PREFIX_1=/home/panda/miniconda3 +CONDA_PROMPT_MODIFIER=(pda) +CONDA_PYTHON_EXE=/home/panda/miniconda3/bin/python +CONDA_SHLVL=2 +CROSS_RANK=0 +CROSS_SIZE=1 +CUDA_MODULE_LOADING=LAZY +CUDA_VISIBLE_DEVICES=0,1 +DBUS_SESSION_BUS_ADDRESS=unix:path=/run/user/1008/bus +GIT_ASKPASS=/home/panda/.vscode-server/cli/servers/Stable-4949701c880d4bdb949e3c0e6b400288da7f474b/server/extensions/git/dist/askpass.sh +HOME=/home/panda +LANG=en_US.UTF-8 +LOCAL_RANK=0 +LOCAL_SIZE=2 +LOGLEVEL=WARNING +LOGNAME=panda +MASTER_ADDR=127.0.0.1 +MASTER_PORT=35777 +MOTD_SHOWN=pam +OLDPWD=/home/panda +PATH=/home/panda/miniconda3/envs/pda/bin:/home/panda/miniconda3/condabin:/home/panda/.vscode-server/cli/servers/Stable-4949701c880d4bdb949e3c0e6b400288da7f474b/server/bin/remote-cli:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/games:/usr/local/games:/snap/bin:/home/panda/.vscode-server/data/User/globalStorage/github.copilot-chat/debugCommand +PWD=/home/panda/pda-llm/scripts +PYTHONHASHSEED=42 +PYTHONPATH=/home/panda/pda-llm +RANK=0 +SHELL=/bin/sh +SHLVL=4 +SSH_CLIENT=10.100.129.163 58321 22 +SSH_CONNECTION=10.100.129.163 58321 158.130.110.127 22 +SSL_CERT_DIR=/usr/lib/ssl/certs +SSL_CERT_FILE=/usr/lib/ssl/certs/ca-certificates.crt +TERM=screen +TERM_PROGRAM=vscode +TERM_PROGRAM_VERSION=1.99.2 +TMUX=/tmp//tmux-1008/default,750869,0 +TMUX_PANE=%0 +USER=panda +VSCODE_GIT_ASKPASS_EXTRA_ARGS= +VSCODE_GIT_ASKPASS_MAIN=/home/panda/.vscode-server/cli/servers/Stable-4949701c880d4bdb949e3c0e6b400288da7f474b/server/extensions/git/dist/askpass-main.js +VSCODE_GIT_ASKPASS_NODE=/home/panda/.vscode-server/cli/servers/Stable-4949701c880d4bdb949e3c0e6b400288da7f474b/server/node +VSCODE_GIT_IPC_HANDLE=/run/user/1008/vscode-git-3d8733097b.sock +VSCODE_IPC_HOOK_CLI=/run/user/1008/vscode-ipc-6dd97596-0fde-4c4b-bbcb-61896b21f983.sock +WANDB_API_KEY=6a71e7fad84fe1aa8f6ccaa01e4e02fcf4c7ffb4 +WANDB_ENTITY=alelab +WANDB_MODE=online +WANDB_SERVICE=2-1548317-tcp-localhost-45581 +WORLD_SIZE=2 +XDG_DATA_DIRS=/usr/local/share:/usr/share:/var/lib/snapd/desktop +XDG_RUNTIME_DIR=/run/user/1008 +XDG_SESSION_CLASS=user +XDG_SESSION_ID=1945 +XDG_SESSION_TYPE=tty +_=/home/panda/miniconda3/envs/pda/bin/deepspeed \ No newline at end of file diff --git a/script.sh b/script.sh new file mode 100644 index 0000000000000000000000000000000000000000..bb746e5446f0fc89a64740a039be8bcaafa1b6c0 --- /dev/null +++ b/script.sh @@ -0,0 +1,197 @@ +#!/usr/bin/env bash +# +# Copyright 2023-2024 PKU-Alignment Team. All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. +# ============================================================================== + +if [ -z "${BASH_VERSION}" ]; then + echo "Please use bash to run this script." >&2 + exit 1 +fi + +set -x + +SCRIPT_DIR="$(cd "$(dirname "$0")" &>/dev/null && pwd)" +ROOT_DIR="$(dirname "${SCRIPT_DIR}")" +export PYTHONPATH="${ROOT_DIR}${PYTHONPATH:+:${PYTHONPATH}}" +export LOGLEVEL="${LOGLEVEL:-WARNING}" +export WANDB_MODE=online +export WANDB_ENTITY="alelab" + +MODEL_NAME_OR_PATH="meta-llama/Llama-3.1-8B-Instruct" +timestamp="$(date +%Y%m%d-%H%M%S)" +unset HOSTFILE +ZERO_STAGE=0 +OFFLOAD="none" +IMPORTANT_SFT=False +# GRIDSEARCH PARAMS +SAFETY_RATIO_TOL=10 +RESILIENT_COEFF=1 +LEARNING_RATE=1e-4 +EPOCHS=3 +while [[ "$#" -gt 0 ]]; do + arg="$1" + shift + case "${arg}" in + --model_name_or_path) + MODEL_NAME_OR_PATH="$1" + shift + ;; + --model_name_or_path=*) + MODEL_NAME_OR_PATH="${arg#*=}" + ;; + --output_dir) + OUTPUT_DIR="$1" + shift + ;; + --output_dir=*) + OUTPUT_DIR="${arg#*=}" + ;; + --hostfile) + HOSTFILE="$1" + shift + ;; + --hostfile=*) + HOSTFILE="${arg#*=}" + ;; + --zero_stage) + ZERO_STAGE="$1" + shift + ;; + --zero_stage=*) + ZERO_STAGE="${arg#*=}" + ;; + --offload) + OFFLOAD="$1" + shift + ;; + --offload=*) + OFFLOAD="${arg#*=}" + ;; + --safety_ratio_tol) + SAFETY_RATIO_TOL="$1" + shift + ;; + --safety_ratio_tol=*) + SAFETY_RATIO_TOL="${arg#*=}" + ;; + --resilient_coeff) + RESILIENT_COEFF="$1" + shift + ;; + --resilient_coeff=*) + RESILIENT_COEFF="${arg#*=}" + ;; + --learning_rate) + LEARNING_RATE="$1" + shift + ;; + --learning_rate=*) + LEARNING_RATE="${arg#*=}" + ;; + --epochs) + EPOCHS="$1" + shift + ;; + --epochs=*) + EPOCHS="${arg#*=}" + ;; + --important_sft) + IMPORTANT_SFT="$1" + shift + ;; + --important_sft=*) + IMPORTANT_SFT="${arg#*=}" + ;; + *) + echo "Unknown parameter passed: '${arg}'" >&2 + exit 1 + ;; + esac +done + +OUTPUT_DIR="${ROOT_DIR}/output/sft-tools/run-${IMPORTANT_SFT}-${RESILIENT_COEFF}-${SAFETY_RATIO_TOL}" +mkdir -p "${OUTPUT_DIR}" +OUTPUT_DIR="$(cd "${OUTPUT_DIR}" &>/dev/null && pwd)" +if [[ ! -f "${OUTPUT_DIR}/.gitignore" ]]; then + echo '*' >"${OUTPUT_DIR}/.gitignore" +fi + +cp -f "$0" "${OUTPUT_DIR}/script.sh" + +export WANDB_API_KEY="6a71e7fad84fe1aa8f6ccaa01e4e02fcf4c7ffb4" +if [[ -z "${WANDB_API_KEY}" ]]; then + export WANDB_MODE="offline" +fi + +MASTER_PORT_START=10000 +MASTER_PORT_END=65535 +MASTER_PORT="$( + comm -23 \ + <(seq "${MASTER_PORT_START}" "${MASTER_PORT_END}" | sort) \ + <(ss -Htan | awk '{ print $4 }' | awk -F ':' '{ print $NF }' | sort -u) | + shuf | head -n 1 +)" + +DEEPSPEED_ARGS=() +if [[ -n "${HOSTFILE+x}" ]]; then + DEEPSPEED_ARGS+=("--hostfile" "${HOSTFILE}") +fi +DEEPSPEED_ARGS+=("--master_port" "${MASTER_PORT}") + +exec 1> >(tee "${OUTPUT_DIR}/stdout.log" >&1) 2> >(tee "${OUTPUT_DIR}/stderr.log" >&2) +echo "--------- Environment sanity check ---------" +echo "shell: $0 running under bash $BASH_VERSION" +echo "conda env: ${CONDA_DEFAULT_ENV:-}" +echo "python: $(which python)" +python - <<'PY' +import sys, os +print("sys.executable :", sys.executable) +print("python version :", sys.version.split()[0]) +print("CONDA_PREFIX :", os.environ.get("CONDA_PREFIX")) +PY +echo "deepspeed: $(command -v deepspeed)" +echo "--------------------------------------------" + +CUDA_VISIBLE_DEVICES=0,1 deepspeed "${DEEPSPEED_ARGS[@]}" \ + --module safe_rlhf.algorithms.tools_ft \ + --train_datasets "tools" \ + --model_name_or_path "${MODEL_NAME_OR_PATH}" \ + --cache_dir "${ROOT_DIR}/cache/sft-tools" \ + --important_sft "${IMPORTANT_SFT}" \ + --max_length 2048 \ + --trust_remote_code True \ + --epochs "${EPOCHS}" \ + --per_device_train_batch_size 1 \ + --per_device_eval_batch_size 1 \ + --gradient_accumulation_steps 48 \ + --gradient_checkpointing \ + --learning_rate "${LEARNING_RATE}" \ + --lr_scheduler_type cosine \ + --lr_warmup_ratio 0.1 \ + --weight_decay 0.0 \ + --seed 42 \ + --output_dir "${OUTPUT_DIR}" \ + --log_type wandb \ + --log_project TOOLS-SFT \ + --zero_stage "${ZERO_STAGE}" \ + --offload "${OFFLOAD}" \ + --safety_ratio_tol "${SAFETY_RATIO_TOL}" \ + --resilient_coeff "${RESILIENT_COEFF}" \ + --lora_r "16" \ + --lora_alpha "32" \ + --lora_dropout "0.05" \ + --bf16 True \ + --fp16 False \ + --tf32 False diff --git a/special_tokens_map.json b/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..6a2f43c3870b2266021ea201101ffffa6f244695 --- /dev/null +++ b/special_tokens_map.json @@ -0,0 +1,30 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "unk_token": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + } +} diff --git a/stderr.log b/stderr.log new file mode 100644 index 0000000000000000000000000000000000000000..29e67fe04f57eb052b3715b4aa4d5d2be3be2ebc --- /dev/null +++ b/stderr.log @@ -0,0 +1,76 @@ ++ echo '--------- Environment sanity check ---------' ++ echo 'shell: ./sft-tools.sh running under bash 5.0.17(1)-release' ++ echo 'conda env: pda' +++ which python ++ echo 'python: /home/panda/miniconda3/envs/pda/bin/python' ++ python - +++ command -v deepspeed ++ echo 'deepspeed: /home/panda/miniconda3/envs/pda/bin/deepspeed' ++ echo -------------------------------------------- ++ CUDA_VISIBLE_DEVICES=0,1 ++ deepspeed --master_port 35777 --module safe_rlhf.algorithms.tools_ft --train_datasets tools --model_name_or_path meta-llama/Llama-3.1-8B-Instruct --cache_dir /home/panda/pda-llm/cache/sft-tools --important_sft false --max_length 2048 --trust_remote_code True --epochs 3 --per_device_train_batch_size 1 --per_device_eval_batch_size 1 --gradient_accumulation_steps 48 --gradient_checkpointing --learning_rate 1e-4 --lr_scheduler_type cosine --lr_warmup_ratio 0.1 --weight_decay 0.0 --seed 42 --output_dir /home/panda/pda-llm/output/sft-tools/run-false-1-10 --log_type wandb --log_project TOOLS-SFT --zero_stage 0 --offload none --safety_ratio_tol 10 --resilient_coeff 1 --lora_r 16 --lora_alpha 32 --lora_dropout 0.05 --bf16 True --fp16 False --tf32 False +[rank1]:[W508 20:43:10.558049870 ProcessGroupNCCL.cpp:4115] [PG ID 0 PG GUID 0 Rank 1] using GPU 1 to perform barrier as devices used by this process are currently unknown. This can potentially cause a hang if this rank to GPU mapping is incorrect.Specify device_ids in barrier() to force use of a particular device,or call init_process_group() with a device_id. +[rank0]:[W508 20:43:10.580350715 ProcessGroupNCCL.cpp:4115] [PG ID 0 PG GUID 0 Rank 0] using GPU 0 to perform barrier as devices used by this process are currently unknown. This can potentially cause a hang if this rank to GPU mapping is incorrect.Specify device_ids in barrier() to force use of a particular device,or call init_process_group() with a device_id. + Loading checkpoint shards: 0%| | 0/4 [00:00, {'localhost': [0, 1]}) +[2025-05-08 20:43:05,396] [INFO] [launch.py:164:main] dist_world_size=2 +[2025-05-08 20:43:05,396] [INFO] [launch.py:168:main] Setting CUDA_VISIBLE_DEVICES=0,1 +[2025-05-08 20:43:05,396] [INFO] [launch.py:256:main] process 1548317 spawned with command: ['/home/panda/miniconda3/envs/pda/bin/python', '-u', '-m', 'safe_rlhf.algorithms.tools_ft', '--local_rank=0', '--train_datasets', 'tools', '--model_name_or_path', 'meta-llama/Llama-3.1-8B-Instruct', '--cache_dir', '/home/panda/pda-llm/cache/sft-tools', '--important_sft', 'false', '--max_length', '2048', '--trust_remote_code', 'True', '--epochs', '3', '--per_device_train_batch_size', '1', '--per_device_eval_batch_size', '1', '--gradient_accumulation_steps', '48', '--gradient_checkpointing', '--learning_rate', '1e-4', '--lr_scheduler_type', 'cosine', '--lr_warmup_ratio', '0.1', '--weight_decay', '0.0', '--seed', '42', '--output_dir', '/home/panda/pda-llm/output/sft-tools/run-false-1-10', '--log_type', 'wandb', '--log_project', 'TOOLS-SFT', '--zero_stage', '0', '--offload', 'none', '--safety_ratio_tol', '10', '--resilient_coeff', '1', '--lora_r', '16', '--lora_alpha', '32', '--lora_dropout', '0.05', '--bf16', 'True', '--fp16', 'False', '--tf32', 'False'] +[2025-05-08 20:43:05,397] [INFO] [launch.py:256:main] process 1548318 spawned with command: ['/home/panda/miniconda3/envs/pda/bin/python', '-u', '-m', 'safe_rlhf.algorithms.tools_ft', '--local_rank=1', '--train_datasets', 'tools', '--model_name_or_path', 'meta-llama/Llama-3.1-8B-Instruct', '--cache_dir', '/home/panda/pda-llm/cache/sft-tools', '--important_sft', 'false', '--max_length', '2048', '--trust_remote_code', 'True', '--epochs', '3', '--per_device_train_batch_size', '1', '--per_device_eval_batch_size', '1', '--gradient_accumulation_steps', '48', '--gradient_checkpointing', '--learning_rate', '1e-4', '--lr_scheduler_type', 'cosine', '--lr_warmup_ratio', '0.1', '--weight_decay', '0.0', '--seed', '42', '--output_dir', '/home/panda/pda-llm/output/sft-tools/run-false-1-10', '--log_type', 'wandb', '--log_project', 'TOOLS-SFT', '--zero_stage', '0', '--offload', 'none', '--safety_ratio_tol', '10', '--resilient_coeff', '1', '--lora_r', '16', '--lora_alpha', '32', '--lora_dropout', '0.05', '--bf16', 'True', '--fp16', 'False', '--tf32', 'False'] +[2025-05-08 20:43:06,560] [INFO] [real_accelerator.py:239:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-05-08 20:43:06,564] [INFO] [real_accelerator.py:239:get_accelerator] Setting ds_accelerator to cuda (auto detect) +[2025-05-08 20:43:09,694] [INFO] [comm.py:669:init_distributed] cdb=None +[2025-05-08 20:43:09,706] [INFO] [comm.py:669:init_distributed] cdb=None +[2025-05-08 20:43:09,707] [INFO] [comm.py:700:init_distributed] Initializing TorchBackend in DeepSpeed with backend nccl +Set logger level to WARNING. +calculating baseline ... +calculating baseline ... +Computing baseline logprobs... +ninja: no work to do. +Time to load fused_adam op: 0.032095909118652344 seconds +Saving computed baseline logprobs to /home/panda/pda-llm/cache/sft-tools/cached_baseline_logprobs.pt +Saved baseline logprobs successfully +ninja: no work to do. +Time to load fused_adam op: 0.034185171127319336 seconds +***** Running training ***** + +***** Evaluating at the beginning ***** + +***** Evaluating at epoch 1/3 ***** + +***** Evaluating at epoch 2/3 ***** + +***** Evaluating at epoch 3/3 ***** +Saving model to "/home/panda/pda-llm/output/sft-tools/run-false-1-10" ... +Saving Hugging Face Checkpoints... +[2025-05-09 00:28:41,058] [INFO] [launch.py:351:main] Process 1548318 exits successfully. +Model saved! +[2025-05-09 00:28:45,059] [INFO] [launch.py:351:main] Process 1548317 exits successfully. diff --git a/tokenizer.json b/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6c121de2e8c636f5025887b6f8041dc7ba379ea7 --- /dev/null +++ b/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:65b66ccdde0ce45c83b06f31e9f11272cade6aff26ec3cc9d6c49ede82b3ee2d +size 17210383 diff --git a/tokenizer_config.json b/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..d1613fb66d55f8779ee8427b9db3d2a2d9910017 --- /dev/null +++ b/tokenizer_config.json @@ -0,0 +1,2082 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128256": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128257": { + "content": "", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 2048, + "pad_token": "", + "padding_side": "right", + "tokenizer_class": "PreTrainedTokenizer", + "unk_token": "" +} diff --git a/wandb/debug-internal.log b/wandb/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d37033d7f8e5ba7af95d0fef4bc3dd9187388fa4 --- /dev/null +++ b/wandb/debug-internal.log @@ -0,0 +1,18 @@ +{"time":"2025-05-08T21:21:54.561866061-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_212154-is8630ik/logs/debug-core.log"} +{"time":"2025-05-08T21:21:54.740453474-04:00","level":"INFO","msg":"created new stream","id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740487984-04:00","level":"INFO","msg":"stream: started","id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740504345-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740556706-04:00","level":"INFO","msg":"handler: started","stream_id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740592257-04:00","level":"INFO","msg":"sender: started","stream_id":"is8630ik"} +{"time":"2025-05-08T21:21:54.884261607-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T21:21:54.884293377-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T21:21:54.884348759-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} +{"time":"2025-05-09T00:28:41.16342178-04:00","level":"INFO","msg":"Stopping system monitor"} +{"time":"2025-05-09T00:28:41.163499301-04:00","level":"INFO","msg":"Stopped system monitor"} +{"time":"2025-05-09T00:28:41.635621981-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-05-09T00:28:41.760403458-04:00","level":"INFO","msg":"handler: operation stats","stats":{}} +{"time":"2025-05-09T00:28:41.773696754-04:00","level":"INFO","msg":"stream: closing","id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773713434-04:00","level":"INFO","msg":"handler: closed","stream_id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773723895-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773731765-04:00","level":"INFO","msg":"sender: closed","stream_id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773786876-04:00","level":"INFO","msg":"stream: closed","id":"is8630ik"} diff --git a/wandb/debug.log b/wandb/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c0d34bc4185ae49d7b75ac0313612476f5540c1c --- /dev/null +++ b/wandb/debug.log @@ -0,0 +1,29 @@ +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Configure stats pid to 1548317 +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_212154-is8630ik/logs/debug.log +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_212154-is8630ik/logs/debug-internal.log +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():852] calling init triggers +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 2048, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 48, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-21-21-54', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():893] starting backend +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():897] sending inform_init request +2025-05-08 21:21:54,558 INFO MainThread:1548317 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 21:21:54,558 INFO MainThread:1548317 [wandb_init.py:init():907] backend started and connected +2025-05-08 21:21:54,559 INFO MainThread:1548317 [wandb_init.py:init():1002] updated telemetry +2025-05-08 21:21:54,564 INFO MainThread:1548317 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 21:21:54,882 INFO MainThread:1548317 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 21:21:54,948 INFO MainThread:1548317 [wandb_init.py:init():1147] run started, returning control to user process +2025-05-09 00:28:41,160 INFO MainThread:1548317 [wandb_run.py:_finish():2314] finishing run alelab/TOOLS-SFT/is8630ik +2025-05-09 00:28:41,162 INFO MainThread:1548317 [wandb_run.py:_atexit_cleanup():2531] got exitcode: 0 +2025-05-09 00:28:41,162 INFO MainThread:1548317 [wandb_run.py:_restore():2513] restore +2025-05-09 00:28:41,162 INFO MainThread:1548317 [wandb_run.py:_restore():2519] restore done +2025-05-09 00:28:41,762 INFO MainThread:1548317 [wandb_run.py:_footer_history_summary_info():4160] rendering history +2025-05-09 00:28:41,765 INFO MainThread:1548317 [wandb_run.py:_footer_history_summary_info():4192] rendering summary +2025-05-09 00:28:41,773 INFO MainThread:1548317 [wandb_run.py:_footer_sync_info():4121] logging synced files diff --git a/wandb/run-20250508_005928-1jhxkbtl/files/config.yaml b/wandb/run-20250508_005928-1jhxkbtl/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..2a16b8ac5486c1bd5159f1ac719e24e9946a5800 --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/files/config.yaml @@ -0,0 +1,133 @@ +_wandb: + value: + cli_version: 0.19.10 + m: [] + python_version: 3.11.11 + t: + "1": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "2": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "3": + - 2 + - 13 + - 16 + - 23 + - 55 + "4": 3.11.11 + "5": 0.19.10 + "6": 4.49.0 + "8": + - 5 + "12": 0.19.10 + "13": linux-x86_64 +bf16: + value: false +cache_dir: + value: /home/panda/pda-llm/cache/sft-tools +deepscale: + value: false +deepscale_config: + value: null +deepspeed: + value: false +deepspeed_config: + value: null +device: + value: cuda:0 +epochs: + value: 3 +eval_datasets: + value: null +eval_interval: + value: 1000000 +eval_split_ratio: + value: null +eval_strategy: + value: epoch +fp16: + value: true +global_rank: + value: 0 +gradient_accumulation_steps: + value: 12 +gradient_checkpointing: + value: true +important_sft: + value: false +local_rank: + value: 0 +log_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +log_project: + value: TOOLS-SFT +log_run_name: + value: tools-sft-2025-05-08-00-59-27 +log_type: + value: wandb +lora_alpha: + value: 32 +lora_dropout: + value: 0.05 +lora_r: + value: 16 +lr: + value: 0.0001 +lr_scheduler_type: + value: COSINE +lr_warmup_ratio: + value: 0.1 +max_length: + value: 4096 +model_name_or_path: + value: meta-llama/Llama-3.1-8B-Instruct +need_eval: + value: true +num_update_steps_per_epoch: + value: 236 +offload: + value: none +output_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +per_device_eval_batch_size: + value: 4 +per_device_train_batch_size: + value: 4 +recompute_baseline: + value: false +resilient_coeff: + value: 1 +safety_ratio_tol: + value: 10 +save_16bit: + value: false +save_interval: + value: 1000000 +seed: + value: 42 +tf32: + value: true +total_training_steps: + value: 708 +train_datasets: + value: + - - tools + - proportion: 1 +trust_remote_code: + value: true +weight_decay: + value: 0 +zero_stage: + value: 0 diff --git a/wandb/run-20250508_005928-1jhxkbtl/files/output.log b/wandb/run-20250508_005928-1jhxkbtl/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..5e368e38e767dd86d259aebbf6fb94d2ec03adfa --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/files/output.log @@ -0,0 +1,33 @@ +***** Running training ***** +Training 1/3 epoch: 0%| | 0/8475 [00:00", line 198, in _run_module_as_main + File "", line 88, in _run_code + File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/__main__.py", line 23, in + sys.exit(main()) + ^^^^^^ + File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/main.py", line 366, in main + trainer.train() + File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 292, in train + self.logger.log(self.eval(), step=0) + ^^^^^^^^^^^ + File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 344, in eval + is_important = batch['is_important'] + ~~~~~^^^^^^^^^^^^^^^^ +KeyError: 'is_important' +[rank0]: Traceback (most recent call last): +[rank0]: File "", line 198, in _run_module_as_main +[rank0]: File "", line 88, in _run_code +[rank0]: File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/__main__.py", line 23, in +[rank0]: sys.exit(main()) +[rank0]: ^^^^^^ +[rank0]: File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/main.py", line 366, in main +[rank0]: trainer.train() +[rank0]: File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 292, in train +[rank0]: self.logger.log(self.eval(), step=0) +[rank0]: ^^^^^^^^^^^ +[rank0]: File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 344, in eval +[rank0]: is_important = batch['is_important'] +[rank0]: ~~~~~^^^^^^^^^^^^^^^^ +[rank0]: KeyError: 'is_important' diff --git a/wandb/run-20250508_005928-1jhxkbtl/files/requirements.txt b/wandb/run-20250508_005928-1jhxkbtl/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..e2b7f5395d84a9b80d09934d0e417c5884a7edca --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/files/requirements.txt @@ -0,0 +1,108 @@ +PySocks==1.7.1 +certifi==2025.4.26 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +smmap==4.0.0 +xxhash==3.5.0 +tzdata==2025.2 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +Markdown==3.8 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +sentry-sdk==2.18.0 +tokenizers==0.21.0 +wandb==0.19.10 +einops==0.8.1 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +wheel==0.45.1 +networkx==3.4.2 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +GitPython==3.1.43 +ninja==1.11.1.4 +fonttools==4.55.3 +annotated-types==0.6.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +transformers==4.49.0 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +setuptools==78.1.1 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +dill==0.3.8 +psutil==5.9.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +packaging==24.2 +pyarrow==19.0.0 +PyYAML==6.0.2 +pandas==2.2.3 +tensorboard==2.19.0 +regex==2024.11.6 +pip==25.1 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +triton==3.1.0 +multiprocess==0.70.15 +platformdirs==4.3.7 +numpy==2.0.1 +protobuf==5.29.3 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +setproctitle==1.2.2 +huggingface_hub==0.29.2 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_005928-1jhxkbtl/files/wandb-metadata.json b/wandb/run-20250508_005928-1jhxkbtl/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..9e4acc7c957740a7a3a1f37aafbe5fc4e7574c20 --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T04:59:28.129836Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "4", + "--per_device_eval_batch_size", + "4", + "--gradient_accumulation_steps", + "12", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "False", + "--fp16", + "True", + "--tf32", + "True" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "2f331e2abf3b7b76829bc02cfb20143a948f9298" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "886362767360" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_005928-1jhxkbtl/files/wandb-summary.json b/wandb/run-20250508_005928-1jhxkbtl/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..6c37fe1cbbb8aed86fd461a79642cb991e4d35cf --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":0}} \ No newline at end of file diff --git a/wandb/run-20250508_005928-1jhxkbtl/logs/debug-core.log b/wandb/run-20250508_005928-1jhxkbtl/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..5694109a088c37b30b19a6f020582068e5d0dd85 --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-05-08T00:59:27.792254317-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpgz00y30n/port-770208.txt","pid":770208,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T00:59:27.798415148-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":770208} +{"time":"2025-05-08T00:59:27.798422588-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":34881,"Zone":""}} +{"time":"2025-05-08T00:59:27.976670485-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:28.131725052-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"1jhxkbtl","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:28.326665772-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"1jhxkbtl","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:29.41791717-04:00","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"1jhxkbtl","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:29.418040741-04:00","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"1jhxkbtl","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:30.418475132-04:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:30.418514823-04:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:30.418524663-04:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-05-08T00:59:30.418559823-04:00","level":"INFO","msg":"connection: closing","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:30.418645274-04:00","level":"INFO","msg":"connection: closed successfully","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:30.418654894-04:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:41134"} +{"time":"2025-05-08T00:59:30.418665734-04:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250508_005928-1jhxkbtl/logs/debug-internal.log b/wandb/run-20250508_005928-1jhxkbtl/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..2280fd6b4d1686d30bac45e668ffb0115eeccda0 --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/logs/debug-internal.log @@ -0,0 +1,18 @@ +{"time":"2025-05-08T00:59:28.132017565-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_005928-1jhxkbtl/logs/debug-core.log"} +{"time":"2025-05-08T00:59:28.326622421-04:00","level":"INFO","msg":"created new stream","id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:28.326657422-04:00","level":"INFO","msg":"stream: started","id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:28.326676812-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:28.326745383-04:00","level":"INFO","msg":"sender: started","stream_id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:28.326827564-04:00","level":"INFO","msg":"handler: started","stream_id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:28.55136741-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T00:59:28.551407061-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T00:59:28.551479361-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} +{"time":"2025-05-08T00:59:28.645082906-04:00","level":"INFO","msg":"Stopping system monitor"} +{"time":"2025-05-08T00:59:28.645130137-04:00","level":"INFO","msg":"Stopped system monitor"} +{"time":"2025-05-08T00:59:29.314315271-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-05-08T00:59:29.414164881-04:00","level":"INFO","msg":"handler: operation stats","stats":{}} +{"time":"2025-05-08T00:59:29.417948-04:00","level":"INFO","msg":"stream: closing","id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:29.41796369-04:00","level":"INFO","msg":"handler: closed","stream_id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:29.41797832-04:00","level":"INFO","msg":"sender: closed","stream_id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:29.41797466-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":"1jhxkbtl"} +{"time":"2025-05-08T00:59:29.418031771-04:00","level":"INFO","msg":"stream: closed","id":"1jhxkbtl"} diff --git a/wandb/run-20250508_005928-1jhxkbtl/logs/debug.log b/wandb/run-20250508_005928-1jhxkbtl/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..21d6a654d03c0f5bfc8f702f24e6a75d51b23f03 --- /dev/null +++ b/wandb/run-20250508_005928-1jhxkbtl/logs/debug.log @@ -0,0 +1,29 @@ +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_setup.py:_flush():68] Configure stats pid to 770208 +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_005928-1jhxkbtl/logs/debug.log +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_005928-1jhxkbtl/logs/debug-internal.log +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_init.py:init():852] calling init triggers +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 4, 'per_device_eval_batch_size': 4, 'gradient_accumulation_steps': 12, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': True, 'bf16': False, 'tf32': True, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-00-59-27', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 236, 'total_training_steps': 708, '_wandb': {}} +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_init.py:init():893] starting backend +2025-05-08 00:59:28,127 INFO MainThread:770208 [wandb_init.py:init():897] sending inform_init request +2025-05-08 00:59:28,129 INFO MainThread:770208 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 00:59:28,129 INFO MainThread:770208 [wandb_init.py:init():907] backend started and connected +2025-05-08 00:59:28,131 INFO MainThread:770208 [wandb_init.py:init():1002] updated telemetry +2025-05-08 00:59:28,136 INFO MainThread:770208 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 00:59:28,548 INFO MainThread:770208 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 00:59:28,603 INFO MainThread:770208 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 00:59:28,604 INFO MainThread:770208 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 00:59:28,604 INFO MainThread:770208 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 00:59:28,604 INFO MainThread:770208 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 00:59:28,605 INFO MainThread:770208 [wandb_init.py:init():1147] run started, returning control to user process +2025-05-08 00:59:28,613 INFO MainThread:770208 [wandb_run.py:_finish():2314] finishing run alelab/TOOLS-SFT/1jhxkbtl +2025-05-08 00:59:28,613 INFO MainThread:770208 [wandb_run.py:_atexit_cleanup():2531] got exitcode: 0 +2025-05-08 00:59:28,613 INFO MainThread:770208 [wandb_run.py:_restore():2513] restore +2025-05-08 00:59:28,613 INFO MainThread:770208 [wandb_run.py:_restore():2519] restore done +2025-05-08 00:59:29,417 INFO MainThread:770208 [wandb_run.py:_footer_history_summary_info():4160] rendering history +2025-05-08 00:59:29,417 INFO MainThread:770208 [wandb_run.py:_footer_history_summary_info():4192] rendering summary +2025-05-08 00:59:29,417 INFO MainThread:770208 [wandb_run.py:_footer_sync_info():4121] logging synced files diff --git a/wandb/run-20250508_005928-1jhxkbtl/run-1jhxkbtl.wandb b/wandb/run-20250508_005928-1jhxkbtl/run-1jhxkbtl.wandb new file mode 100644 index 0000000000000000000000000000000000000000..8c90a2891fc969130f56b83118fb8a0f9f468db1 Binary files /dev/null and b/wandb/run-20250508_005928-1jhxkbtl/run-1jhxkbtl.wandb differ diff --git a/wandb/run-20250508_160817-0lem0nck/files/output.log b/wandb/run-20250508_160817-0lem0nck/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..1b5e843da6aaa40067b4b60ea36e77ebabe44d36 --- /dev/null +++ b/wandb/run-20250508_160817-0lem0nck/files/output.log @@ -0,0 +1,2 @@ +***** Running training ***** +Training 1/3 epoch (loss 0.1149): 0%| | 10/16950 [00:05<2:23:52, 1.96it/s] diff --git a/wandb/run-20250508_160817-0lem0nck/files/requirements.txt b/wandb/run-20250508_160817-0lem0nck/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_160817-0lem0nck/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_160817-0lem0nck/files/wandb-metadata.json b/wandb/run-20250508_160817-0lem0nck/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..f5805aaa293f9aa234d58abca154486d1b77419f --- /dev/null +++ b/wandb/run-20250508_160817-0lem0nck/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T20:08:17.504166Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "48", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "False", + "--fp16", + "True", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888964751360" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_160817-0lem0nck/logs/debug-core.log b/wandb/run-20250508_160817-0lem0nck/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..9a11e63163170153266b8797d097e6479471a783 --- /dev/null +++ b/wandb/run-20250508_160817-0lem0nck/logs/debug-core.log @@ -0,0 +1,7 @@ +{"time":"2025-05-08T16:08:17.190213216-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpmnui6ecd/port-1356767.txt","pid":1356767,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T16:08:17.196180277-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1356767} +{"time":"2025-05-08T16:08:17.196169706-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":33129,"Zone":""}} +{"time":"2025-05-08T16:08:17.376078695-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:43214"} +{"time":"2025-05-08T16:08:17.506096175-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"0lem0nck","id":"127.0.0.1:43214"} +{"time":"2025-05-08T16:08:17.699304314-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"0lem0nck","id":"127.0.0.1:43214"} +{"time":"2025-05-08T16:08:23.606079392-04:00","level":"INFO","msg":"received shutdown signal","signal":15} diff --git a/wandb/run-20250508_160817-0lem0nck/logs/debug-internal.log b/wandb/run-20250508_160817-0lem0nck/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..bc3d7d49cc5601b7d5dac9970ccdb580b9c3d781 --- /dev/null +++ b/wandb/run-20250508_160817-0lem0nck/logs/debug-internal.log @@ -0,0 +1,9 @@ +{"time":"2025-05-08T16:08:17.506377451-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_160817-0lem0nck/logs/debug-core.log"} +{"time":"2025-05-08T16:08:17.699250973-04:00","level":"INFO","msg":"created new stream","id":"0lem0nck"} +{"time":"2025-05-08T16:08:17.699295584-04:00","level":"INFO","msg":"stream: started","id":"0lem0nck"} +{"time":"2025-05-08T16:08:17.699316065-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"0lem0nck"} +{"time":"2025-05-08T16:08:17.699320815-04:00","level":"INFO","msg":"handler: started","stream_id":"0lem0nck"} +{"time":"2025-05-08T16:08:17.699350985-04:00","level":"INFO","msg":"sender: started","stream_id":"0lem0nck"} +{"time":"2025-05-08T16:08:17.867900772-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T16:08:17.867932482-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T16:08:17.867991013-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} diff --git a/wandb/run-20250508_160817-0lem0nck/logs/debug.log b/wandb/run-20250508_160817-0lem0nck/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..86f6f2e318d6e864e09ef0b291f42bcd9c4282dc --- /dev/null +++ b/wandb/run-20250508_160817-0lem0nck/logs/debug.log @@ -0,0 +1,22 @@ +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_setup.py:_flush():68] Configure stats pid to 1356767 +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_160817-0lem0nck/logs/debug.log +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_160817-0lem0nck/logs/debug-internal.log +2025-05-08 16:08:17,501 INFO MainThread:1356767 [wandb_init.py:init():852] calling init triggers +2025-05-08 16:08:17,502 INFO MainThread:1356767 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 48, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': True, 'bf16': False, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': False, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-16-08-17', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 16:08:17,502 INFO MainThread:1356767 [wandb_init.py:init():893] starting backend +2025-05-08 16:08:17,502 INFO MainThread:1356767 [wandb_init.py:init():897] sending inform_init request +2025-05-08 16:08:17,503 INFO MainThread:1356767 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 16:08:17,504 INFO MainThread:1356767 [wandb_init.py:init():907] backend started and connected +2025-05-08 16:08:17,505 INFO MainThread:1356767 [wandb_init.py:init():1002] updated telemetry +2025-05-08 16:08:17,510 INFO MainThread:1356767 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 16:08:17,865 INFO MainThread:1356767 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 16:08:17,927 INFO MainThread:1356767 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 16:08:17,927 INFO MainThread:1356767 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 16:08:17,927 INFO MainThread:1356767 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 16:08:17,927 INFO MainThread:1356767 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 16:08:17,929 INFO MainThread:1356767 [wandb_init.py:init():1147] run started, returning control to user process diff --git a/wandb/run-20250508_160817-0lem0nck/run-0lem0nck.wandb b/wandb/run-20250508_160817-0lem0nck/run-0lem0nck.wandb new file mode 100644 index 0000000000000000000000000000000000000000..e69de29bb2d1d6434b8b29ae775ad8c2e48c5391 diff --git a/wandb/run-20250508_160902-sm1tkrdz/files/output.log b/wandb/run-20250508_160902-sm1tkrdz/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..97394f483c61ec0ae3fb5128f58d181845eecce9 --- /dev/null +++ b/wandb/run-20250508_160902-sm1tkrdz/files/output.log @@ -0,0 +1,2 @@ +***** Running training ***** +Training 1/3 epoch (loss 0.0966): 0%| | 48/16950 [00:31<3:03:57, 1.53it/s] diff --git a/wandb/run-20250508_160902-sm1tkrdz/files/requirements.txt b/wandb/run-20250508_160902-sm1tkrdz/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_160902-sm1tkrdz/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_160902-sm1tkrdz/files/wandb-metadata.json b/wandb/run-20250508_160902-sm1tkrdz/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..02feaccb1a10b585c3682dcc4fb2a93a548aa3e9 --- /dev/null +++ b/wandb/run-20250508_160902-sm1tkrdz/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T20:09:02.642208Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "48", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "False", + "--fp16", + "True", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888964878336" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_160902-sm1tkrdz/logs/debug-core.log b/wandb/run-20250508_160902-sm1tkrdz/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..122f46d7e09b4f222a0e652f810daa93e359a7bc --- /dev/null +++ b/wandb/run-20250508_160902-sm1tkrdz/logs/debug-core.log @@ -0,0 +1,7 @@ +{"time":"2025-05-08T16:09:02.351406897-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpab3smczz/port-1359956.txt","pid":1359956,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T16:09:02.357060512-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1359956} +{"time":"2025-05-08T16:09:02.357067772-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":34899,"Zone":""}} +{"time":"2025-05-08T16:09:02.536932502-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:52382"} +{"time":"2025-05-08T16:09:02.644160501-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"sm1tkrdz","id":"127.0.0.1:52382"} +{"time":"2025-05-08T16:09:02.838054407-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"sm1tkrdz","id":"127.0.0.1:52382"} +{"time":"2025-05-08T16:09:35.696088794-04:00","level":"INFO","msg":"received shutdown signal","signal":15} diff --git a/wandb/run-20250508_160902-sm1tkrdz/logs/debug-internal.log b/wandb/run-20250508_160902-sm1tkrdz/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..8b119000b623d220b62c9a110ad8497c802553ca --- /dev/null +++ b/wandb/run-20250508_160902-sm1tkrdz/logs/debug-internal.log @@ -0,0 +1,9 @@ +{"time":"2025-05-08T16:09:02.644440987-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_160902-sm1tkrdz/logs/debug-core.log"} +{"time":"2025-05-08T16:09:02.838002086-04:00","level":"INFO","msg":"created new stream","id":"sm1tkrdz"} +{"time":"2025-05-08T16:09:02.838045347-04:00","level":"INFO","msg":"stream: started","id":"sm1tkrdz"} +{"time":"2025-05-08T16:09:02.838062637-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"sm1tkrdz"} +{"time":"2025-05-08T16:09:02.838103338-04:00","level":"INFO","msg":"handler: started","stream_id":"sm1tkrdz"} +{"time":"2025-05-08T16:09:02.838129438-04:00","level":"INFO","msg":"sender: started","stream_id":"sm1tkrdz"} +{"time":"2025-05-08T16:09:02.986767932-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T16:09:02.986795302-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T16:09:02.986852694-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} diff --git a/wandb/run-20250508_160902-sm1tkrdz/logs/debug.log b/wandb/run-20250508_160902-sm1tkrdz/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..dcf027abd6d100016fccbb90f35906775d29ed06 --- /dev/null +++ b/wandb/run-20250508_160902-sm1tkrdz/logs/debug.log @@ -0,0 +1,22 @@ +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_setup.py:_flush():68] Configure stats pid to 1359956 +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_160902-sm1tkrdz/logs/debug.log +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_160902-sm1tkrdz/logs/debug-internal.log +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_init.py:init():852] calling init triggers +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 48, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': True, 'bf16': False, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': False, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-16-09-02', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_init.py:init():893] starting backend +2025-05-08 16:09:02,639 INFO MainThread:1359956 [wandb_init.py:init():897] sending inform_init request +2025-05-08 16:09:02,642 INFO MainThread:1359956 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 16:09:02,642 INFO MainThread:1359956 [wandb_init.py:init():907] backend started and connected +2025-05-08 16:09:02,643 INFO MainThread:1359956 [wandb_init.py:init():1002] updated telemetry +2025-05-08 16:09:02,648 INFO MainThread:1359956 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 16:09:02,984 INFO MainThread:1359956 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 16:09:03,046 INFO MainThread:1359956 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 16:09:03,046 INFO MainThread:1359956 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 16:09:03,046 INFO MainThread:1359956 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 16:09:03,046 INFO MainThread:1359956 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 16:09:03,048 INFO MainThread:1359956 [wandb_init.py:init():1147] run started, returning control to user process diff --git a/wandb/run-20250508_160902-sm1tkrdz/run-sm1tkrdz.wandb b/wandb/run-20250508_160902-sm1tkrdz/run-sm1tkrdz.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ac45b567630f5585072bc6324fa142b963febac9 Binary files /dev/null and b/wandb/run-20250508_160902-sm1tkrdz/run-sm1tkrdz.wandb differ diff --git a/wandb/run-20250508_183954-8zs3b4m6/files/config.yaml b/wandb/run-20250508_183954-8zs3b4m6/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..f53be718868bc2293426c148cf1383aefbe8cf2c --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/files/config.yaml @@ -0,0 +1,133 @@ +_wandb: + value: + cli_version: 0.19.10 + m: [] + python_version: 3.11.11 + t: + "1": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "2": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "3": + - 2 + - 13 + - 16 + - 23 + - 55 + "4": 3.11.11 + "5": 0.19.10 + "6": 4.51.3 + "8": + - 5 + "12": 0.19.10 + "13": linux-x86_64 +bf16: + value: true +cache_dir: + value: /home/panda/pda-llm/cache/sft-tools +deepscale: + value: false +deepscale_config: + value: null +deepspeed: + value: false +deepspeed_config: + value: null +device: + value: cuda:0 +epochs: + value: 3 +eval_datasets: + value: null +eval_interval: + value: 1000000 +eval_split_ratio: + value: null +eval_strategy: + value: epoch +fp16: + value: false +global_rank: + value: 0 +gradient_accumulation_steps: + value: 24 +gradient_checkpointing: + value: true +important_sft: + value: false +local_rank: + value: 0 +log_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +log_project: + value: TOOLS-SFT +log_run_name: + value: tools-sft-2025-05-08-18-39-54 +log_type: + value: wandb +lora_alpha: + value: 32 +lora_dropout: + value: 0.05 +lora_r: + value: 16 +lr: + value: 0.0001 +lr_scheduler_type: + value: COSINE +lr_warmup_ratio: + value: 0.1 +max_length: + value: 4096 +model_name_or_path: + value: meta-llama/Llama-3.1-8B-Instruct +need_eval: + value: true +num_update_steps_per_epoch: + value: 118 +offload: + value: none +output_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +per_device_eval_batch_size: + value: 2 +per_device_train_batch_size: + value: 2 +recompute_baseline: + value: false +resilient_coeff: + value: 1 +safety_ratio_tol: + value: 10 +save_16bit: + value: false +save_interval: + value: 1000000 +seed: + value: 42 +tf32: + value: false +total_training_steps: + value: 354 +train_datasets: + value: + - - tools + - proportion: 1 +trust_remote_code: + value: true +weight_decay: + value: 0 +zero_stage: + value: 0 diff --git a/wandb/run-20250508_183954-8zs3b4m6/files/output.log b/wandb/run-20250508_183954-8zs3b4m6/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..b36e3568eeebaf7f234e753453f4e45da253c298 --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/files/output.log @@ -0,0 +1,33 @@ +***** Running training ***** +Training 1/3 epoch: 0%| | 0/8475 [00:00", line 198, in _run_module_as_main + File "", line 88, in _run_code + File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/__main__.py", line 23, in + sys.exit(main()) + ^^^^^^ + File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/main.py", line 367, in main + trainer.train() + File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 292, in train + self.logger.log(self.eval(), step=0) + ^^^^^^^^^^^ + File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 361, in eval + safe_log_ratios = safe_log_ratios.cpu().numpy() + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +TypeError: Got unsupported ScalarType BFloat16 +[rank0]: Traceback (most recent call last): +[rank0]: File "", line 198, in _run_module_as_main +[rank0]: File "", line 88, in _run_code +[rank0]: File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/__main__.py", line 23, in +[rank0]: sys.exit(main()) +[rank0]: ^^^^^^ +[rank0]: File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/main.py", line 367, in main +[rank0]: trainer.train() +[rank0]: File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 292, in train +[rank0]: self.logger.log(self.eval(), step=0) +[rank0]: ^^^^^^^^^^^ +[rank0]: File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 361, in eval +[rank0]: safe_log_ratios = safe_log_ratios.cpu().numpy() +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: TypeError: Got unsupported ScalarType BFloat16 diff --git a/wandb/run-20250508_183954-8zs3b4m6/files/requirements.txt b/wandb/run-20250508_183954-8zs3b4m6/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_183954-8zs3b4m6/files/wandb-metadata.json b/wandb/run-20250508_183954-8zs3b4m6/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..95ccdd3c60a61a5a49dfecb93ecf6e2c7e9a27b5 --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T22:39:54.527119Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "2", + "--per_device_eval_batch_size", + "2", + "--gradient_accumulation_steps", + "24", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888969158656" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_183954-8zs3b4m6/files/wandb-summary.json b/wandb/run-20250508_183954-8zs3b4m6/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..1e721ee0e6a3739077251a63440b7756199b1b46 --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":234}} \ No newline at end of file diff --git a/wandb/run-20250508_183954-8zs3b4m6/logs/debug-core.log b/wandb/run-20250508_183954-8zs3b4m6/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..b613cb41fb229dcbe39a6a648c2b44791160f025 --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-05-08T18:39:54.176871407-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpm9rm19xp/port-1436037.txt","pid":1436037,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T18:39:54.182885566-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1436037} +{"time":"2025-05-08T18:39:54.182859406-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":46483,"Zone":""}} +{"time":"2025-05-08T18:39:54.363158642-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:39:54.529022063-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"8zs3b4m6","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:39:54.721870969-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"8zs3b4m6","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:49.728612807-04:00","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"8zs3b4m6","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:49.72878048-04:00","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"8zs3b4m6","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:50.729045178-04:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:50.729069669-04:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:50.729078859-04:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-05-08T18:43:50.729084919-04:00","level":"INFO","msg":"connection: closing","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:50.72916792-04:00","level":"INFO","msg":"connection: closed successfully","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:50.729198191-04:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:44756"} +{"time":"2025-05-08T18:43:50.729212691-04:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250508_183954-8zs3b4m6/logs/debug-internal.log b/wandb/run-20250508_183954-8zs3b4m6/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..e32549d45cc78fb97b55316d9e3bb3b0b380d377 --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/logs/debug-internal.log @@ -0,0 +1,18 @@ +{"time":"2025-05-08T18:39:54.529292648-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_183954-8zs3b4m6/logs/debug-core.log"} +{"time":"2025-05-08T18:39:54.721820928-04:00","level":"INFO","msg":"created new stream","id":"8zs3b4m6"} +{"time":"2025-05-08T18:39:54.721862709-04:00","level":"INFO","msg":"stream: started","id":"8zs3b4m6"} +{"time":"2025-05-08T18:39:54.72190933-04:00","level":"INFO","msg":"handler: started","stream_id":"8zs3b4m6"} +{"time":"2025-05-08T18:39:54.7218919-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"8zs3b4m6"} +{"time":"2025-05-08T18:39:54.721943141-04:00","level":"INFO","msg":"sender: started","stream_id":"8zs3b4m6"} +{"time":"2025-05-08T18:39:55.093280788-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T18:39:55.093320509-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T18:39:55.09338449-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} +{"time":"2025-05-08T18:43:49.04487936-04:00","level":"INFO","msg":"Stopping system monitor"} +{"time":"2025-05-08T18:43:49.045105404-04:00","level":"INFO","msg":"Stopped system monitor"} +{"time":"2025-05-08T18:43:49.452801239-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-05-08T18:43:49.726313582-04:00","level":"INFO","msg":"handler: operation stats","stats":{}} +{"time":"2025-05-08T18:43:49.728641487-04:00","level":"INFO","msg":"stream: closing","id":"8zs3b4m6"} +{"time":"2025-05-08T18:43:49.728657388-04:00","level":"INFO","msg":"handler: closed","stream_id":"8zs3b4m6"} +{"time":"2025-05-08T18:43:49.728667428-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":"8zs3b4m6"} +{"time":"2025-05-08T18:43:49.728696378-04:00","level":"INFO","msg":"sender: closed","stream_id":"8zs3b4m6"} +{"time":"2025-05-08T18:43:49.72877225-04:00","level":"INFO","msg":"stream: closed","id":"8zs3b4m6"} diff --git a/wandb/run-20250508_183954-8zs3b4m6/logs/debug.log b/wandb/run-20250508_183954-8zs3b4m6/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7642ad12837c9e959156f8965ebaaac42ab8eea0 --- /dev/null +++ b/wandb/run-20250508_183954-8zs3b4m6/logs/debug.log @@ -0,0 +1,29 @@ +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_setup.py:_flush():68] Configure stats pid to 1436037 +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_183954-8zs3b4m6/logs/debug.log +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_183954-8zs3b4m6/logs/debug-internal.log +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_init.py:init():852] calling init triggers +2025-05-08 18:39:54,524 INFO MainThread:1436037 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'gradient_accumulation_steps': 24, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-18-39-54', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 18:39:54,525 INFO MainThread:1436037 [wandb_init.py:init():893] starting backend +2025-05-08 18:39:54,525 INFO MainThread:1436037 [wandb_init.py:init():897] sending inform_init request +2025-05-08 18:39:54,526 INFO MainThread:1436037 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 18:39:54,526 INFO MainThread:1436037 [wandb_init.py:init():907] backend started and connected +2025-05-08 18:39:54,528 INFO MainThread:1436037 [wandb_init.py:init():1002] updated telemetry +2025-05-08 18:39:54,534 INFO MainThread:1436037 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 18:39:55,090 INFO MainThread:1436037 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 18:39:55,157 INFO MainThread:1436037 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 18:39:55,157 INFO MainThread:1436037 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 18:39:55,157 INFO MainThread:1436037 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 18:39:55,157 INFO MainThread:1436037 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 18:39:55,158 INFO MainThread:1436037 [wandb_init.py:init():1147] run started, returning control to user process +2025-05-08 18:43:49,004 INFO MainThread:1436037 [wandb_run.py:_finish():2314] finishing run alelab/TOOLS-SFT/8zs3b4m6 +2025-05-08 18:43:49,005 INFO MainThread:1436037 [wandb_run.py:_atexit_cleanup():2531] got exitcode: 0 +2025-05-08 18:43:49,005 INFO MainThread:1436037 [wandb_run.py:_restore():2513] restore +2025-05-08 18:43:49,005 INFO MainThread:1436037 [wandb_run.py:_restore():2519] restore done +2025-05-08 18:43:49,727 INFO MainThread:1436037 [wandb_run.py:_footer_history_summary_info():4160] rendering history +2025-05-08 18:43:49,728 INFO MainThread:1436037 [wandb_run.py:_footer_history_summary_info():4192] rendering summary +2025-05-08 18:43:49,728 INFO MainThread:1436037 [wandb_run.py:_footer_sync_info():4121] logging synced files diff --git a/wandb/run-20250508_183954-8zs3b4m6/run-8zs3b4m6.wandb b/wandb/run-20250508_183954-8zs3b4m6/run-8zs3b4m6.wandb new file mode 100644 index 0000000000000000000000000000000000000000..addad4f8e9365f5f95527300e1f81a8c54a20386 Binary files /dev/null and b/wandb/run-20250508_183954-8zs3b4m6/run-8zs3b4m6.wandb differ diff --git a/wandb/run-20250508_190248-2v7dhhsz/files/config.yaml b/wandb/run-20250508_190248-2v7dhhsz/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..e2cefd0bf196aa2f0beaf2a01a5ee67e163b68da --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/files/config.yaml @@ -0,0 +1,134 @@ +_wandb: + value: + cli_version: 0.19.10 + m: [] + python_version: 3.11.11 + t: + "1": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "2": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "3": + - 2 + - 13 + - 16 + - 23 + - 55 + - 61 + "4": 3.11.11 + "5": 0.19.10 + "6": 4.51.3 + "8": + - 5 + "12": 0.19.10 + "13": linux-x86_64 +bf16: + value: true +cache_dir: + value: /home/panda/pda-llm/cache/sft-tools +deepscale: + value: false +deepscale_config: + value: null +deepspeed: + value: false +deepspeed_config: + value: null +device: + value: cuda:0 +epochs: + value: 3 +eval_datasets: + value: null +eval_interval: + value: 1000000 +eval_split_ratio: + value: null +eval_strategy: + value: epoch +fp16: + value: false +global_rank: + value: 0 +gradient_accumulation_steps: + value: 24 +gradient_checkpointing: + value: true +important_sft: + value: false +local_rank: + value: 0 +log_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +log_project: + value: TOOLS-SFT +log_run_name: + value: tools-sft-2025-05-08-19-02-47 +log_type: + value: wandb +lora_alpha: + value: 32 +lora_dropout: + value: 0.05 +lora_r: + value: 16 +lr: + value: 0.0001 +lr_scheduler_type: + value: COSINE +lr_warmup_ratio: + value: 0.1 +max_length: + value: 4096 +model_name_or_path: + value: meta-llama/Llama-3.1-8B-Instruct +need_eval: + value: true +num_update_steps_per_epoch: + value: 118 +offload: + value: none +output_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +per_device_eval_batch_size: + value: 2 +per_device_train_batch_size: + value: 2 +recompute_baseline: + value: false +resilient_coeff: + value: 1 +safety_ratio_tol: + value: 10 +save_16bit: + value: false +save_interval: + value: 1000000 +seed: + value: 42 +tf32: + value: false +total_training_steps: + value: 354 +train_datasets: + value: + - - tools + - proportion: 1 +trust_remote_code: + value: true +weight_decay: + value: 0 +zero_stage: + value: 0 diff --git a/wandb/run-20250508_190248-2v7dhhsz/files/output.log b/wandb/run-20250508_190248-2v7dhhsz/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..74fa91b8c2727bdf87afbab2550c8e5acc7096b4 --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/files/output.log @@ -0,0 +1,141 @@ +***** Running training ***** +Training 1/3 epoch: 0%| | 0/8475 [00:00", line 198, in _run_module_as_main + File "", line 88, in _run_code + File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/__main__.py", line 23, in + sys.exit(main()) + ^^^^^^ + File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/main.py", line 367, in main + trainer.train() + File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 292, in train + self.logger.log(self.eval(), step=0) + File "/home/panda/pda-llm/safe_rlhf/utils.py", line 195, in wrapper + return func(*args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/pda-llm/safe_rlhf/logger.py", line 158, in log + self.wandb.log(metrics, step=step) + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 406, in wrapper + return func(self, *args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 503, in wrapper + return func(self, *args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 464, in wrapper_fn + return func(self, *args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 451, in wrapper + return func(self, *args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2064, in log + self._log(data=data, step=step, commit=commit) + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 1776, in _log + self._partial_history_callback(data, step, commit) + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 406, in wrapper + return func(self, *args, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 1603, in _partial_history_callback + self._backend.interface.publish_partial_history( + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/interface/interface.py", line 674, in publish_partial_history + data = history_dict_to_json(run, data, step=user_step, ignore_copy_err=True) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/data_types/utils.py", line 54, in history_dict_to_json + payload[key] = val_to_json( + ^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/data_types/utils.py", line 162, in val_to_json + val.bind_to_run(run, key, namespace) + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/data_types/table.py", line 510, in bind_to_run + util.json_dump_safer(data, fp) + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/util.py", line 812, in json_dump_safer + return dump(obj, fp, cls=WandBJSONEncoder, **kwargs) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/__init__.py", line 179, in dump + for chunk in iterable: + File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 432, in _iterencode + yield from _iterencode_dict(o, _current_indent_level) + File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 406, in _iterencode_dict + yield from chunks + File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 326, in _iterencode_list + yield from chunks + File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 326, in _iterencode_list + yield from chunks + File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 439, in _iterencode + o = _default(o) + ^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/util.py", line 763, in default + tmp_obj, converted = json_friendly(obj) + ^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/util.py", line 595, in json_friendly + obj = obj.cpu().detach().numpy() + ^^^^^^^^^^^^^^^^^^^^^^^^^^ +TypeError: Got unsupported ScalarType BFloat16 +[rank0]: Traceback (most recent call last): +[rank0]: File "", line 198, in _run_module_as_main +[rank0]: File "", line 88, in _run_code +[rank0]: File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/__main__.py", line 23, in +[rank0]: sys.exit(main()) +[rank0]: ^^^^^^ +[rank0]: File "/home/panda/pda-llm/safe_rlhf/algorithms/tools_ft/main.py", line 367, in main +[rank0]: trainer.train() +[rank0]: File "/home/panda/pda-llm/safe_rlhf/trainers/supervised_tools_trainer.py", line 292, in train +[rank0]: self.logger.log(self.eval(), step=0) +[rank0]: File "/home/panda/pda-llm/safe_rlhf/utils.py", line 195, in wrapper +[rank0]: return func(*args, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/pda-llm/safe_rlhf/logger.py", line 158, in log +[rank0]: self.wandb.log(metrics, step=step) +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 406, in wrapper +[rank0]: return func(self, *args, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 503, in wrapper +[rank0]: return func(self, *args, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 464, in wrapper_fn +[rank0]: return func(self, *args, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 451, in wrapper +[rank0]: return func(self, *args, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2064, in log +[rank0]: self._log(data=data, step=step, commit=commit) +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 1776, in _log +[rank0]: self._partial_history_callback(data, step, commit) +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 406, in wrapper +[rank0]: return func(self, *args, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 1603, in _partial_history_callback +[rank0]: self._backend.interface.publish_partial_history( +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/interface/interface.py", line 674, in publish_partial_history +[rank0]: data = history_dict_to_json(run, data, step=user_step, ignore_copy_err=True) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/data_types/utils.py", line 54, in history_dict_to_json +[rank0]: payload[key] = val_to_json( +[rank0]: ^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/data_types/utils.py", line 162, in val_to_json +[rank0]: val.bind_to_run(run, key, namespace) +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/data_types/table.py", line 510, in bind_to_run +[rank0]: util.json_dump_safer(data, fp) +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/util.py", line 812, in json_dump_safer +[rank0]: return dump(obj, fp, cls=WandBJSONEncoder, **kwargs) +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/__init__.py", line 179, in dump +[rank0]: for chunk in iterable: +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 432, in _iterencode +[rank0]: yield from _iterencode_dict(o, _current_indent_level) +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 406, in _iterencode_dict +[rank0]: yield from chunks +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 326, in _iterencode_list +[rank0]: yield from chunks +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 326, in _iterencode_list +[rank0]: yield from chunks +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/json/encoder.py", line 439, in _iterencode +[rank0]: o = _default(o) +[rank0]: ^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/util.py", line 763, in default +[rank0]: tmp_obj, converted = json_friendly(obj) +[rank0]: ^^^^^^^^^^^^^^^^^^ +[rank0]: File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/util.py", line 595, in json_friendly +[rank0]: obj = obj.cpu().detach().numpy() +[rank0]: ^^^^^^^^^^^^^^^^^^^^^^^^^^ +[rank0]: TypeError: Got unsupported ScalarType BFloat16 diff --git a/wandb/run-20250508_190248-2v7dhhsz/files/requirements.txt b/wandb/run-20250508_190248-2v7dhhsz/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_190248-2v7dhhsz/files/wandb-metadata.json b/wandb/run-20250508_190248-2v7dhhsz/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..5d0b2fef20add4052c96dfd07e9dcc9b46c4526b --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T23:02:48.306604Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "2", + "--per_device_eval_batch_size", + "2", + "--gradient_accumulation_steps", + "24", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888971706368" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_190248-2v7dhhsz/files/wandb-summary.json b/wandb/run-20250508_190248-2v7dhhsz/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..5998a97095cb161150819561b7d0ecdc63597dd3 --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/files/wandb-summary.json @@ -0,0 +1 @@ +{"_wandb":{"runtime":280}} \ No newline at end of file diff --git a/wandb/run-20250508_190248-2v7dhhsz/logs/debug-core.log b/wandb/run-20250508_190248-2v7dhhsz/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..689a2e8da92d37763a266f95e36418103861091c --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/logs/debug-core.log @@ -0,0 +1,14 @@ +{"time":"2025-05-08T19:02:47.960541849-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpihmishva/port-1477147.txt","pid":1477147,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T19:02:47.966289105-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1477147} +{"time":"2025-05-08T19:02:47.966293555-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":41139,"Zone":""}} +{"time":"2025-05-08T19:02:48.147653935-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:02:48.308576825-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"2v7dhhsz","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:02:48.512505858-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"2v7dhhsz","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:29.938545851-04:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:29.938626302-04:00","level":"INFO","msg":"connection: closing","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:29.938655533-04:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-05-08T19:07:29.938762645-04:00","level":"INFO","msg":"connection: closed successfully","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:30.213249369-04:00","level":"ERROR","msg":"processOutgoingData: flush error","error":"write tcp 127.0.0.1:41139->127.0.0.1:35080: use of closed network connection","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:30.213368232-04:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:30.213382952-04:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:35080"} +{"time":"2025-05-08T19:07:30.213392242-04:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250508_190248-2v7dhhsz/logs/debug-internal.log b/wandb/run-20250508_190248-2v7dhhsz/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..34f5d55f61c69aa4572d9d6190d4075134b9b161 --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/logs/debug-internal.log @@ -0,0 +1,18 @@ +{"time":"2025-05-08T19:02:48.3088284-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_190248-2v7dhhsz/logs/debug-core.log"} +{"time":"2025-05-08T19:02:48.512462137-04:00","level":"INFO","msg":"created new stream","id":"2v7dhhsz"} +{"time":"2025-05-08T19:02:48.512497528-04:00","level":"INFO","msg":"stream: started","id":"2v7dhhsz"} +{"time":"2025-05-08T19:02:48.512513028-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"2v7dhhsz"} +{"time":"2025-05-08T19:02:48.512533188-04:00","level":"INFO","msg":"handler: started","stream_id":"2v7dhhsz"} +{"time":"2025-05-08T19:02:48.51259588-04:00","level":"INFO","msg":"sender: started","stream_id":"2v7dhhsz"} +{"time":"2025-05-08T19:02:48.704760577-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T19:02:48.704823668-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T19:02:48.704894959-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} +{"time":"2025-05-08T19:07:28.85735877-04:00","level":"INFO","msg":"Stopping system monitor"} +{"time":"2025-05-08T19:07:28.857413541-04:00","level":"INFO","msg":"Stopped system monitor"} +{"time":"2025-05-08T19:07:29.938677033-04:00","level":"INFO","msg":"stream: closing","id":"2v7dhhsz"} +{"time":"2025-05-08T19:07:29.938711754-04:00","level":"WARN","msg":"sender: received Exit record more than once, ignoring"} +{"time":"2025-05-08T19:07:29.98872358-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-05-08T19:07:30.213208069-04:00","level":"INFO","msg":"handler: closed","stream_id":"2v7dhhsz"} +{"time":"2025-05-08T19:07:30.213244659-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":"2v7dhhsz"} +{"time":"2025-05-08T19:07:30.2132889-04:00","level":"INFO","msg":"sender: closed","stream_id":"2v7dhhsz"} +{"time":"2025-05-08T19:07:30.213314711-04:00","level":"INFO","msg":"stream: closed","id":"2v7dhhsz"} diff --git a/wandb/run-20250508_190248-2v7dhhsz/logs/debug.log b/wandb/run-20250508_190248-2v7dhhsz/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..700b6393998050709842cd64f16cc803a9ff316a --- /dev/null +++ b/wandb/run-20250508_190248-2v7dhhsz/logs/debug.log @@ -0,0 +1,47 @@ +2025-05-08 19:02:48,303 INFO MainThread:1477147 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_setup.py:_flush():68] Configure stats pid to 1477147 +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_190248-2v7dhhsz/logs/debug.log +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_190248-2v7dhhsz/logs/debug-internal.log +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_init.py:init():852] calling init triggers +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 2, 'per_device_eval_batch_size': 2, 'gradient_accumulation_steps': 24, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-19-02-47', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_init.py:init():893] starting backend +2025-05-08 19:02:48,304 INFO MainThread:1477147 [wandb_init.py:init():897] sending inform_init request +2025-05-08 19:02:48,306 INFO MainThread:1477147 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 19:02:48,306 INFO MainThread:1477147 [wandb_init.py:init():907] backend started and connected +2025-05-08 19:02:48,308 INFO MainThread:1477147 [wandb_init.py:init():1002] updated telemetry +2025-05-08 19:02:48,314 INFO MainThread:1477147 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 19:02:48,702 INFO MainThread:1477147 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 19:02:48,765 INFO MainThread:1477147 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 19:02:48,765 INFO MainThread:1477147 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 19:02:48,765 INFO MainThread:1477147 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 19:02:48,765 INFO MainThread:1477147 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 19:02:48,766 INFO MainThread:1477147 [wandb_init.py:init():1147] run started, returning control to user process +2025-05-08 19:07:28,824 INFO MainThread:1477147 [wandb_run.py:_finish():2314] finishing run alelab/TOOLS-SFT/2v7dhhsz +2025-05-08 19:07:28,824 INFO MainThread:1477147 [wandb_run.py:_atexit_cleanup():2531] got exitcode: 0 +2025-05-08 19:07:28,824 INFO MainThread:1477147 [wandb_run.py:_restore():2513] restore +2025-05-08 19:07:28,824 INFO MainThread:1477147 [wandb_run.py:_restore():2519] restore done +2025-05-08 19:07:29,825 INFO MainThread:1477147 [wandb_run.py:_restore():2513] restore +2025-05-08 19:07:29,825 INFO MainThread:1477147 [wandb_run.py:_restore():2519] restore done +2025-05-08 19:07:29,825 ERROR MainThread:1477147 [wandb_run.py:_atexit_cleanup():2552] Problem finishing run +Traceback (most recent call last): + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2543, in _atexit_cleanup + self._on_finish() + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/wandb_run.py", line 2799, in _on_finish + wait_with_progress( + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/mailbox/wait_with_progress.py", line 24, in wait_with_progress + return wait_all_with_progress( + ^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/mailbox/wait_with_progress.py", line 87, in wait_all_with_progress + return asyncio_compat.run(progress_loop_with_timeout) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/wandb/sdk/lib/asyncio_compat.py", line 27, in run + future = executor.submit(runner.run, fn) + ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ + File "/home/panda/miniconda3/envs/pda/lib/python3.11/concurrent/futures/thread.py", line 169, in submit + raise RuntimeError('cannot schedule new futures after ' +RuntimeError: cannot schedule new futures after interpreter shutdown +2025-05-08 19:07:29,938 INFO MsgRouterThr:1477147 [mailbox.py:close():129] [no run ID] Closing mailbox, abandoning 2 handles. diff --git a/wandb/run-20250508_190248-2v7dhhsz/run-2v7dhhsz.wandb b/wandb/run-20250508_190248-2v7dhhsz/run-2v7dhhsz.wandb new file mode 100644 index 0000000000000000000000000000000000000000..ce1bc2edb68de3562cae81ad92dacb50beaa7147 Binary files /dev/null and b/wandb/run-20250508_190248-2v7dhhsz/run-2v7dhhsz.wandb differ diff --git a/wandb/run-20250508_192410-bvjuhkx9/files/output.log b/wandb/run-20250508_192410-bvjuhkx9/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..d7a3c4bfe2ab4ebcc6f2729646cce46fcd5ca277 --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/files/output.log @@ -0,0 +1,4 @@ +***** Running training ***** +Training 1/3 epoch (loss 1.2245): 2%|▏ | 277/16950 [07:23<3:07:41, 1.48it/s] + +***** Evaluating at the beginning ***** diff --git a/wandb/run-20250508_192410-bvjuhkx9/files/requirements.txt b/wandb/run-20250508_192410-bvjuhkx9/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_192410-bvjuhkx9/files/wandb-metadata.json b/wandb/run-20250508_192410-bvjuhkx9/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..3a8dee75b4ccf56c453befb0a5fd9814e18b846a --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T23:24:10.153063Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "24", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888972967936" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_192410-bvjuhkx9/logs/debug-core.log b/wandb/run-20250508_192410-bvjuhkx9/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..6cc1d182b3754a78a9904972ba7941acf676820b --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/logs/debug-core.log @@ -0,0 +1,7 @@ +{"time":"2025-05-08T19:24:09.82237144-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpm7ixav89/port-1493943.txt","pid":1493943,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T19:24:09.828178485-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":42089,"Zone":""}} +{"time":"2025-05-08T19:24:09.828213636-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1493943} +{"time":"2025-05-08T19:24:10.006299553-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:33400"} +{"time":"2025-05-08T19:24:10.156512128-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"bvjuhkx9","id":"127.0.0.1:33400"} +{"time":"2025-05-08T19:24:10.359218782-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"bvjuhkx9","id":"127.0.0.1:33400"} +{"time":"2025-05-08T19:31:37.213002082-04:00","level":"INFO","msg":"received shutdown signal","signal":15} diff --git a/wandb/run-20250508_192410-bvjuhkx9/logs/debug-internal.log b/wandb/run-20250508_192410-bvjuhkx9/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..00ba703bc805d41ca409c07cf6cae5e19295e318 --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/logs/debug-internal.log @@ -0,0 +1,9 @@ +{"time":"2025-05-08T19:24:10.156798224-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_192410-bvjuhkx9/logs/debug-core.log"} +{"time":"2025-05-08T19:24:10.359168401-04:00","level":"INFO","msg":"created new stream","id":"bvjuhkx9"} +{"time":"2025-05-08T19:24:10.359210061-04:00","level":"INFO","msg":"stream: started","id":"bvjuhkx9"} +{"time":"2025-05-08T19:24:10.359221702-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"bvjuhkx9"} +{"time":"2025-05-08T19:24:10.359274803-04:00","level":"INFO","msg":"sender: started","stream_id":"bvjuhkx9"} +{"time":"2025-05-08T19:24:10.359247342-04:00","level":"INFO","msg":"handler: started","stream_id":"bvjuhkx9"} +{"time":"2025-05-08T19:24:10.503713274-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T19:24:10.503761015-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T19:24:10.503813646-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} diff --git a/wandb/run-20250508_192410-bvjuhkx9/logs/debug.log b/wandb/run-20250508_192410-bvjuhkx9/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..306193443d9ea3abeaf1050688a561ec6101bb2c --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/logs/debug.log @@ -0,0 +1,22 @@ +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_setup.py:_flush():68] Configure stats pid to 1493943 +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_192410-bvjuhkx9/logs/debug.log +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_192410-bvjuhkx9/logs/debug-internal.log +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_init.py:init():852] calling init triggers +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 24, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-19-24-09', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 236, 'total_training_steps': 708, '_wandb': {}} +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_init.py:init():893] starting backend +2025-05-08 19:24:10,150 INFO MainThread:1493943 [wandb_init.py:init():897] sending inform_init request +2025-05-08 19:24:10,152 INFO MainThread:1493943 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 19:24:10,152 INFO MainThread:1493943 [wandb_init.py:init():907] backend started and connected +2025-05-08 19:24:10,154 INFO MainThread:1493943 [wandb_init.py:init():1002] updated telemetry +2025-05-08 19:24:10,160 INFO MainThread:1493943 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 19:24:10,501 INFO MainThread:1493943 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 19:24:10,567 INFO MainThread:1493943 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 19:24:10,568 INFO MainThread:1493943 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 19:24:10,568 INFO MainThread:1493943 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 19:24:10,568 INFO MainThread:1493943 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 19:24:10,569 INFO MainThread:1493943 [wandb_init.py:init():1147] run started, returning control to user process diff --git a/wandb/run-20250508_192410-bvjuhkx9/run-bvjuhkx9.wandb b/wandb/run-20250508_192410-bvjuhkx9/run-bvjuhkx9.wandb new file mode 100644 index 0000000000000000000000000000000000000000..482226ea15d9aa4febffbb60e928e33bd152f61e --- /dev/null +++ b/wandb/run-20250508_192410-bvjuhkx9/run-bvjuhkx9.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5d93fab7ce997c8c7c3003593d553cca8a51958f86a1dc2741b6d2033a70d328 +size 229376 diff --git a/wandb/run-20250508_195305-eg5dmumm/files/output.log b/wandb/run-20250508_195305-eg5dmumm/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..6c859fbd815f2298f4775a87cfa97dfc135033b1 --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/files/output.log @@ -0,0 +1,2 @@ +***** Running training ***** +Training 1/3 epoch (loss 1.2252): 2%|▏ | 277/16950 [02:50<3:15:41, 1.42it/s] diff --git a/wandb/run-20250508_195305-eg5dmumm/files/requirements.txt b/wandb/run-20250508_195305-eg5dmumm/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_195305-eg5dmumm/files/wandb-metadata.json b/wandb/run-20250508_195305-eg5dmumm/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..4ff7d6d19bd64ca0cd605038f33232f3413125fa --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-08T23:53:05.519065Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "4096", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "24", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888974594048" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_195305-eg5dmumm/logs/debug-core.log b/wandb/run-20250508_195305-eg5dmumm/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..30f175db739be1452e13480015db12fa7e07255c --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/logs/debug-core.log @@ -0,0 +1,7 @@ +{"time":"2025-05-08T19:53:05.178546582-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpzz5ecly8/port-1513282.txt","pid":1513282,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T19:53:05.184358662-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1513282} +{"time":"2025-05-08T19:53:05.184343092-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":37623,"Zone":""}} +{"time":"2025-05-08T19:53:05.364749873-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:42854"} +{"time":"2025-05-08T19:53:05.521830994-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"eg5dmumm","id":"127.0.0.1:42854"} +{"time":"2025-05-08T19:53:05.715689443-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"eg5dmumm","id":"127.0.0.1:42854"} +{"time":"2025-05-08T19:56:00.128045328-04:00","level":"INFO","msg":"received shutdown signal","signal":15} diff --git a/wandb/run-20250508_195305-eg5dmumm/logs/debug-internal.log b/wandb/run-20250508_195305-eg5dmumm/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..6f1c10a90db27ade4f608e5916d1dfa4b5a7e0c1 --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/logs/debug-internal.log @@ -0,0 +1,9 @@ +{"time":"2025-05-08T19:53:05.522072549-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_195305-eg5dmumm/logs/debug-core.log"} +{"time":"2025-05-08T19:53:05.715632182-04:00","level":"INFO","msg":"created new stream","id":"eg5dmumm"} +{"time":"2025-05-08T19:53:05.715680643-04:00","level":"INFO","msg":"stream: started","id":"eg5dmumm"} +{"time":"2025-05-08T19:53:05.715707744-04:00","level":"INFO","msg":"handler: started","stream_id":"eg5dmumm"} +{"time":"2025-05-08T19:53:05.715714574-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"eg5dmumm"} +{"time":"2025-05-08T19:53:05.715744524-04:00","level":"INFO","msg":"sender: started","stream_id":"eg5dmumm"} +{"time":"2025-05-08T19:53:05.882615356-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T19:53:05.882643227-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T19:53:05.882703378-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} diff --git a/wandb/run-20250508_195305-eg5dmumm/logs/debug.log b/wandb/run-20250508_195305-eg5dmumm/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..4432340e5bf6fbad0bbaddfcb8b1097530d0b77f --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/logs/debug.log @@ -0,0 +1,22 @@ +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_setup.py:_flush():68] Configure stats pid to 1513282 +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_195305-eg5dmumm/logs/debug.log +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_195305-eg5dmumm/logs/debug-internal.log +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_init.py:init():852] calling init triggers +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 4096, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 24, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': False, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-19-53-05', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 236, 'total_training_steps': 708, '_wandb': {}} +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_init.py:init():893] starting backend +2025-05-08 19:53:05,516 INFO MainThread:1513282 [wandb_init.py:init():897] sending inform_init request +2025-05-08 19:53:05,518 INFO MainThread:1513282 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 19:53:05,518 INFO MainThread:1513282 [wandb_init.py:init():907] backend started and connected +2025-05-08 19:53:05,521 INFO MainThread:1513282 [wandb_init.py:init():1002] updated telemetry +2025-05-08 19:53:05,527 INFO MainThread:1513282 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 19:53:05,880 INFO MainThread:1513282 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 19:53:05,945 INFO MainThread:1513282 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 19:53:05,945 INFO MainThread:1513282 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 19:53:05,945 INFO MainThread:1513282 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 19:53:05,945 INFO MainThread:1513282 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 19:53:05,946 INFO MainThread:1513282 [wandb_init.py:init():1147] run started, returning control to user process diff --git a/wandb/run-20250508_195305-eg5dmumm/run-eg5dmumm.wandb b/wandb/run-20250508_195305-eg5dmumm/run-eg5dmumm.wandb new file mode 100644 index 0000000000000000000000000000000000000000..d5f1722a10063a776f1e635ea56b5ccae2297149 --- /dev/null +++ b/wandb/run-20250508_195305-eg5dmumm/run-eg5dmumm.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:77775e6c57a7d09999c20c0b9ecb06d6e6d8c179a239b81fee8fb2b0a589476b +size 229376 diff --git a/wandb/run-20250508_200345-11tfsh2x/files/output.log b/wandb/run-20250508_200345-11tfsh2x/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..96fc640765a28284b44ca00e7ed0cb0d98691b73 --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/files/output.log @@ -0,0 +1,2 @@ +***** Running training ***** +Training 1/3 epoch (loss 0.1382): 2%|▏ | 322/16950 [03:15<2:55:07, 1.58it/s] diff --git a/wandb/run-20250508_200345-11tfsh2x/files/requirements.txt b/wandb/run-20250508_200345-11tfsh2x/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_200345-11tfsh2x/files/wandb-metadata.json b/wandb/run-20250508_200345-11tfsh2x/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..fd821d3bb4b348ed673a1deb4c20c2e9729fc251 --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-09T00:03:45.425570Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "2048", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "48", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888978477056" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_200345-11tfsh2x/logs/debug-core.log b/wandb/run-20250508_200345-11tfsh2x/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..6d9b6f8f76ae39a7957085934680421fe32ec5fa --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/logs/debug-core.log @@ -0,0 +1,7 @@ +{"time":"2025-05-08T20:03:45.106333664-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpm1f2r9ji/port-1521451.txt","pid":1521451,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T20:03:45.112205483-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1521451} +{"time":"2025-05-08T20:03:45.112199702-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":39791,"Zone":""}} +{"time":"2025-05-08T20:03:45.291091092-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:38834"} +{"time":"2025-05-08T20:03:45.427445086-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"11tfsh2x","id":"127.0.0.1:38834"} +{"time":"2025-05-08T20:03:45.603023788-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"11tfsh2x","id":"127.0.0.1:38834"} +{"time":"2025-05-08T20:07:01.95135013-04:00","level":"INFO","msg":"received shutdown signal","signal":15} diff --git a/wandb/run-20250508_200345-11tfsh2x/logs/debug-internal.log b/wandb/run-20250508_200345-11tfsh2x/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..decc0e06bbad4bcdfb9602a53e5d5907597cdd55 --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/logs/debug-internal.log @@ -0,0 +1,9 @@ +{"time":"2025-05-08T20:03:45.427702031-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_200345-11tfsh2x/logs/debug-core.log"} +{"time":"2025-05-08T20:03:45.602972307-04:00","level":"INFO","msg":"created new stream","id":"11tfsh2x"} +{"time":"2025-05-08T20:03:45.603015138-04:00","level":"INFO","msg":"stream: started","id":"11tfsh2x"} +{"time":"2025-05-08T20:03:45.603036709-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"11tfsh2x"} +{"time":"2025-05-08T20:03:45.603052919-04:00","level":"INFO","msg":"handler: started","stream_id":"11tfsh2x"} +{"time":"2025-05-08T20:03:45.60309844-04:00","level":"INFO","msg":"sender: started","stream_id":"11tfsh2x"} +{"time":"2025-05-08T20:03:45.789794907-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T20:03:45.789827868-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T20:03:45.789885869-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} diff --git a/wandb/run-20250508_200345-11tfsh2x/logs/debug.log b/wandb/run-20250508_200345-11tfsh2x/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..7fa5c3b4d8e9935dcb0c321339b18d101b259dd5 --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/logs/debug.log @@ -0,0 +1,22 @@ +2025-05-08 20:03:45,422 INFO MainThread:1521451 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_setup.py:_flush():68] Configure stats pid to 1521451 +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_200345-11tfsh2x/logs/debug.log +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_200345-11tfsh2x/logs/debug-internal.log +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_init.py:init():852] calling init triggers +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 2048, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 48, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': False, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-20-03-45', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_init.py:init():893] starting backend +2025-05-08 20:03:45,423 INFO MainThread:1521451 [wandb_init.py:init():897] sending inform_init request +2025-05-08 20:03:45,425 INFO MainThread:1521451 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 20:03:45,425 INFO MainThread:1521451 [wandb_init.py:init():907] backend started and connected +2025-05-08 20:03:45,427 INFO MainThread:1521451 [wandb_init.py:init():1002] updated telemetry +2025-05-08 20:03:45,433 INFO MainThread:1521451 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 20:03:45,787 INFO MainThread:1521451 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 20:03:45,850 INFO MainThread:1521451 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 20:03:45,850 INFO MainThread:1521451 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 20:03:45,850 INFO MainThread:1521451 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 20:03:45,850 INFO MainThread:1521451 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 20:03:45,852 INFO MainThread:1521451 [wandb_init.py:init():1147] run started, returning control to user process diff --git a/wandb/run-20250508_200345-11tfsh2x/run-11tfsh2x.wandb b/wandb/run-20250508_200345-11tfsh2x/run-11tfsh2x.wandb new file mode 100644 index 0000000000000000000000000000000000000000..7d9ca2a7c0b99eed16abff8bbb2488ec056d2a6f --- /dev/null +++ b/wandb/run-20250508_200345-11tfsh2x/run-11tfsh2x.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:482e4443af9fff6620a3802caaa139af0aecc06f1ce4e56001d93c489d683362 +size 262144 diff --git a/wandb/run-20250508_200812-r9ll1pfp/files/output.log b/wandb/run-20250508_200812-r9ll1pfp/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..4bc3c057626a9080af6800cd6294aaaa5aeb2233 --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/files/output.log @@ -0,0 +1,4 @@ +***** Running training ***** +Training 1/3 epoch (loss 0.6629): 18%|█▊ | 2973/16950 [34:26<2:26:27, 1.59it/s] + +***** Evaluating at the beginning ***** diff --git a/wandb/run-20250508_200812-r9ll1pfp/files/requirements.txt b/wandb/run-20250508_200812-r9ll1pfp/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_200812-r9ll1pfp/files/wandb-metadata.json b/wandb/run-20250508_200812-r9ll1pfp/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..443e7fa5ab2b9cd5574cadfa3e01824c7528d6f4 --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-09T00:08:12.242292Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "2048", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "48", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "888978935808" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_200812-r9ll1pfp/logs/debug-core.log b/wandb/run-20250508_200812-r9ll1pfp/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..bf5109dd1c6ef62d0d52f7500d9d0a9f46b4ab82 --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/logs/debug-core.log @@ -0,0 +1,7 @@ +{"time":"2025-05-08T20:08:11.923316769-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmp4qh6mtww/port-1525237.txt","pid":1525237,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T20:08:11.929274359-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1525237} +{"time":"2025-05-08T20:08:11.929273519-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":39583,"Zone":""}} +{"time":"2025-05-08T20:08:12.107579258-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:54756"} +{"time":"2025-05-08T20:08:12.244322974-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"r9ll1pfp","id":"127.0.0.1:54756"} +{"time":"2025-05-08T20:08:12.42003295-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"r9ll1pfp","id":"127.0.0.1:54756"} +{"time":"2025-05-08T20:42:39.674287806-04:00","level":"INFO","msg":"received shutdown signal","signal":15} diff --git a/wandb/run-20250508_200812-r9ll1pfp/logs/debug-internal.log b/wandb/run-20250508_200812-r9ll1pfp/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..4ab7cc27fa4392be05b448ebe41148821c380364 --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/logs/debug-internal.log @@ -0,0 +1,11 @@ +{"time":"2025-05-08T20:08:12.24464186-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_200812-r9ll1pfp/logs/debug-core.log"} +{"time":"2025-05-08T20:08:12.419977979-04:00","level":"INFO","msg":"created new stream","id":"r9ll1pfp"} +{"time":"2025-05-08T20:08:12.42002456-04:00","level":"INFO","msg":"stream: started","id":"r9ll1pfp"} +{"time":"2025-05-08T20:08:12.420050511-04:00","level":"INFO","msg":"handler: started","stream_id":"r9ll1pfp"} +{"time":"2025-05-08T20:08:12.420061401-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"r9ll1pfp"} +{"time":"2025-05-08T20:08:12.420107742-04:00","level":"INFO","msg":"sender: started","stream_id":"r9ll1pfp"} +{"time":"2025-05-08T20:08:12.597153795-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T20:08:12.597212296-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T20:08:12.597274918-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} +{"time":"2025-05-08T20:42:39.655861396-04:00","level":"INFO","msg":"Stopping system monitor"} +{"time":"2025-05-08T20:42:39.655921007-04:00","level":"INFO","msg":"Stopped system monitor"} diff --git a/wandb/run-20250508_200812-r9ll1pfp/logs/debug.log b/wandb/run-20250508_200812-r9ll1pfp/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..61eb0d4e4a6e142653f9019d82970e2d9ad1a696 --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/logs/debug.log @@ -0,0 +1,26 @@ +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_setup.py:_flush():68] Configure stats pid to 1525237 +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_200812-r9ll1pfp/logs/debug.log +2025-05-08 20:08:12,239 INFO MainThread:1525237 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_200812-r9ll1pfp/logs/debug-internal.log +2025-05-08 20:08:12,240 INFO MainThread:1525237 [wandb_init.py:init():852] calling init triggers +2025-05-08 20:08:12,240 INFO MainThread:1525237 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 2048, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 48, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-20-08-11', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 20:08:12,240 INFO MainThread:1525237 [wandb_init.py:init():893] starting backend +2025-05-08 20:08:12,240 INFO MainThread:1525237 [wandb_init.py:init():897] sending inform_init request +2025-05-08 20:08:12,242 INFO MainThread:1525237 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 20:08:12,242 INFO MainThread:1525237 [wandb_init.py:init():907] backend started and connected +2025-05-08 20:08:12,243 INFO MainThread:1525237 [wandb_init.py:init():1002] updated telemetry +2025-05-08 20:08:12,250 INFO MainThread:1525237 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 20:08:12,594 INFO MainThread:1525237 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 20:08:12,659 INFO MainThread:1525237 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 20:08:12,659 INFO MainThread:1525237 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 20:08:12,659 INFO MainThread:1525237 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 20:08:12,659 INFO MainThread:1525237 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 20:08:12,661 INFO MainThread:1525237 [wandb_init.py:init():1147] run started, returning control to user process +2025-05-08 20:42:39,655 INFO MainThread:1525237 [wandb_run.py:_finish():2314] finishing run alelab/TOOLS-SFT/r9ll1pfp +2025-05-08 20:42:39,655 INFO MainThread:1525237 [wandb_run.py:_atexit_cleanup():2531] got exitcode: 0 +2025-05-08 20:42:39,655 INFO MainThread:1525237 [wandb_run.py:_restore():2513] restore +2025-05-08 20:42:39,655 INFO MainThread:1525237 [wandb_run.py:_restore():2519] restore done diff --git a/wandb/run-20250508_200812-r9ll1pfp/run-r9ll1pfp.wandb b/wandb/run-20250508_200812-r9ll1pfp/run-r9ll1pfp.wandb new file mode 100644 index 0000000000000000000000000000000000000000..2bf041431b732664adf2910c5be690461de5cae9 --- /dev/null +++ b/wandb/run-20250508_200812-r9ll1pfp/run-r9ll1pfp.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5905e280278c394658f5270c5a695a04c9b91303a4562df73d46ca263f57661 +size 2490368 diff --git a/wandb/run-20250508_212154-is8630ik/files/config.yaml b/wandb/run-20250508_212154-is8630ik/files/config.yaml new file mode 100644 index 0000000000000000000000000000000000000000..566a2bda79248236b1e455a226b22be58dc42b4d --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/files/config.yaml @@ -0,0 +1,134 @@ +_wandb: + value: + cli_version: 0.19.10 + m: [] + python_version: 3.11.11 + t: + "1": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "2": + - 1 + - 11 + - 49 + - 51 + - 55 + - 71 + - 98 + "3": + - 2 + - 13 + - 16 + - 23 + - 55 + - 61 + "4": 3.11.11 + "5": 0.19.10 + "6": 4.51.3 + "8": + - 5 + "12": 0.19.10 + "13": linux-x86_64 +bf16: + value: true +cache_dir: + value: /home/panda/pda-llm/cache/sft-tools +deepscale: + value: false +deepscale_config: + value: null +deepspeed: + value: false +deepspeed_config: + value: null +device: + value: cuda:0 +epochs: + value: 3 +eval_datasets: + value: null +eval_interval: + value: 1000000 +eval_split_ratio: + value: null +eval_strategy: + value: epoch +fp16: + value: false +global_rank: + value: 0 +gradient_accumulation_steps: + value: 48 +gradient_checkpointing: + value: true +important_sft: + value: false +local_rank: + value: 0 +log_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +log_project: + value: TOOLS-SFT +log_run_name: + value: tools-sft-2025-05-08-21-21-54 +log_type: + value: wandb +lora_alpha: + value: 32 +lora_dropout: + value: 0.05 +lora_r: + value: 16 +lr: + value: 0.0001 +lr_scheduler_type: + value: COSINE +lr_warmup_ratio: + value: 0.1 +max_length: + value: 2048 +model_name_or_path: + value: meta-llama/Llama-3.1-8B-Instruct +need_eval: + value: true +num_update_steps_per_epoch: + value: 118 +offload: + value: none +output_dir: + value: /home/panda/pda-llm/output/sft-tools/run-false-1-10 +per_device_eval_batch_size: + value: 1 +per_device_train_batch_size: + value: 1 +recompute_baseline: + value: false +resilient_coeff: + value: 1 +safety_ratio_tol: + value: 10 +save_16bit: + value: false +save_interval: + value: 1000000 +seed: + value: 42 +tf32: + value: false +total_training_steps: + value: 354 +train_datasets: + value: + - - tools + - proportion: 1 +trust_remote_code: + value: true +weight_decay: + value: 0 +zero_stage: + value: 0 diff --git a/wandb/run-20250508_212154-is8630ik/files/output.log b/wandb/run-20250508_212154-is8630ik/files/output.log new file mode 100644 index 0000000000000000000000000000000000000000..373ed12d67671ffe62f212d5f9e083983a8a79dc --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/files/output.log @@ -0,0 +1,15 @@ +***** Running training ***** +Training 3/3 epoch (loss 0.0331): 100%|██████████| 16950/16950 [3:06:43<00:00, 1.51it/s] + +***** Evaluating at the beginning ***** + +***** Evaluating at epoch 1/3 ***** + +***** Evaluating at epoch 2/3 ***** + +***** Evaluating at epoch 3/3 ***** +Saving model to "/home/panda/pda-llm/output/sft-tools/run-false-1-10" ... +Saving Hugging Face Checkpoints... +/home/panda/miniconda3/envs/pda/lib/python3.11/site-packages/peft/utils/save_and_load.py:220: UserWarning: Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`. + warnings.warn("Setting `save_embedding_layers` to `True` as embedding layers found in `target_modules`.") +Model saved! diff --git a/wandb/run-20250508_212154-is8630ik/files/requirements.txt b/wandb/run-20250508_212154-is8630ik/files/requirements.txt new file mode 100644 index 0000000000000000000000000000000000000000..71b302373526f7bd332d0d4bad620c97b4da950d --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/files/requirements.txt @@ -0,0 +1,137 @@ +PySocks==1.7.1 +pip==25.1.1 +certifi==2025.4.26 +parso==0.8.4 +wcwidth==0.2.13 +nvidia-ml-py==12.535.133 +mkl_fft==1.3.11 +urllib3==2.3.0 +charset-normalizer==3.3.2 +transformers==4.51.3 +smmap==4.0.0 +xxhash==3.5.0 +platformdirs==4.3.8 +tzdata==2025.2 +ipython==9.2.0 +deepspeed==0.16.7 +gmpy2==2.2.1 +pydantic==2.10.3 +tqdm==4.67.1 +pyparsing==3.2.0 +exceptiongroup==1.2.2 +tokenizers==0.21.1 +Markdown==3.8 +matplotlib-inline==0.1.7 +jupyter_core==5.7.2 +pydantic_core==2.27.1 +hjson==3.1.0 +mkl_random==1.2.8 +ipykernel==6.29.5 +sentry-sdk==2.18.0 +wandb==0.19.10 +einops==0.8.1 +prompt_toolkit==3.0.51 +Bottleneck==1.4.2 +msgpack==1.1.0 +fsspec==2024.12.0 +mdurl==0.1.0 +executing==2.2.0 +wheel==0.45.1 +bitsandbytes==0.45.5 +networkx==3.4.2 +comm==0.2.2 +asttokens==3.0.0 +nvitop==1.5.0 +sympy==1.13.1 +MarkupSafe==3.0.2 +numexpr==2.10.1 +tensorboard_data_server==0.7.0 +markdown-it-py==2.2.0 +six==1.17.0 +idna==3.7 +docker-pycreds==0.4.0 +jedi==0.19.2 +GitPython==3.1.43 +pure_eval==0.2.3 +ninja==1.11.1.4 +setuptools==80.1.0 +fonttools==4.55.3 +annotated-types==0.6.0 +psutil==7.0.0 +cycler==0.11.0 +py-cpuinfo==9.0.0 +accelerate==1.6.0 +filelock==3.17.0 +pytz==2024.1 +rich==13.9.4 +appdirs==1.4.4 +click==8.1.8 +Jinja2==3.1.6 +unicodedata2==15.1.0 +pillow==11.1.0 +traitlets==5.14.3 +stack_data==0.6.3 +peft==0.15.2 +python-dateutil==2.9.0.post0 +requests==2.32.3 +Werkzeug==3.1.3 +gitdb==4.0.7 +torch==2.5.1 +scipy==1.15.2 +sentencepiece==0.2.0 +huggingface-hub==0.31.1 +dill==0.3.8 +pexpect==4.9.0 +pickleshare==0.7.5 +ptyprocess==0.7.0 +contourpy==1.3.1 +grpcio==1.71.0 +optree==0.14.1 +safetensors==0.5.3 +mpmath==1.3.0 +nest_asyncio==1.6.0 +pyarrow==19.0.0 +PyYAML==6.0.2 +decorator==5.2.1 +pandas==2.2.3 +tensorboard==2.19.0 +zipp==3.21.0 +regex==2024.11.6 +Brotli==1.0.9 +mkl-service==2.4.0 +matplotlib==3.10.0 +Pygments==2.19.1 +absl-py==2.1.0 +hf-xet==1.1.0 +packaging==25.0 +jupyter_client==8.6.3 +triton==3.1.0 +multiprocess==0.70.15 +debugpy==1.8.14 +numpy==2.0.1 +protobuf==5.29.3 +tornado==6.4.2 +datasets==3.6.0 +eval_type_backport==0.2.2 +typing_extensions==4.12.2 +kiwisolver==1.4.8 +pyzmq==26.4.0 +ipython_pygments_lexers==1.1.1 +setproctitle==1.2.2 +importlib_metadata==8.6.1 +jaraco.text==3.12.1 +backports.tarfile==1.2.0 +importlib_metadata==8.0.0 +jaraco.context==5.3.0 +more-itertools==10.3.0 +wheel==0.45.1 +jaraco.collections==5.1.0 +platformdirs==4.2.2 +autocommand==2.2.2 +zipp==3.19.2 +tomli==2.0.1 +typeguard==4.3.0 +packaging==24.2 +inflect==7.3.1 +typing_extensions==4.12.2 +jaraco.functools==4.0.1 diff --git a/wandb/run-20250508_212154-is8630ik/files/wandb-metadata.json b/wandb/run-20250508_212154-is8630ik/files/wandb-metadata.json new file mode 100644 index 0000000000000000000000000000000000000000..cd2a9114655a855a988bfed89dcfe949567e9263 --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/files/wandb-metadata.json @@ -0,0 +1,106 @@ +{ + "os": "Linux-5.15.0-138-generic-x86_64-with-glibc2.31", + "python": "CPython 3.11.11", + "startedAt": "2025-05-09T01:21:54.558480Z", + "args": [ + "--local_rank=0", + "--train_datasets", + "tools", + "--model_name_or_path", + "meta-llama/Llama-3.1-8B-Instruct", + "--cache_dir", + "/home/panda/pda-llm/cache/sft-tools", + "--important_sft", + "false", + "--max_length", + "2048", + "--trust_remote_code", + "True", + "--epochs", + "3", + "--per_device_train_batch_size", + "1", + "--per_device_eval_batch_size", + "1", + "--gradient_accumulation_steps", + "48", + "--gradient_checkpointing", + "--learning_rate", + "1e-4", + "--lr_scheduler_type", + "cosine", + "--lr_warmup_ratio", + "0.1", + "--weight_decay", + "0.0", + "--seed", + "42", + "--output_dir", + "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "--log_type", + "wandb", + "--log_project", + "TOOLS-SFT", + "--zero_stage", + "0", + "--offload", + "none", + "--safety_ratio_tol", + "10", + "--resilient_coeff", + "1", + "--lora_r", + "16", + "--lora_alpha", + "32", + "--lora_dropout", + "0.05", + "--bf16", + "True", + "--fp16", + "False", + "--tf32", + "False" + ], + "program": "-m safe_rlhf.algorithms.tools_ft.__main__", + "git": { + "remote": "git@github.com:IgnacioBoero/pda-llm.git", + "commit": "3e25c5a329b7cfcf30b4af51d00c42fd38478aba" + }, + "email": "iboero@upenn.edu", + "root": "/home/panda/pda-llm/output/sft-tools/run-false-1-10", + "host": "plaza", + "executable": "/home/panda/miniconda3/envs/pda/bin/python", + "cpu_count": 24, + "cpu_count_logical": 48, + "gpu": "NVIDIA RTX A6000", + "gpu_count": 2, + "disk": { + "/": { + "total": "982820896768", + "used": "889025499136" + } + }, + "memory": { + "total": "270090010624" + }, + "cpu": { + "count": 24, + "countLogical": 48 + }, + "gpu_nvidia": [ + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + }, + { + "name": "NVIDIA RTX A6000", + "memoryTotal": "51527024640", + "cudaCores": 10752, + "architecture": "Ampere" + } + ], + "cudaVersion": "12.4" +} \ No newline at end of file diff --git a/wandb/run-20250508_212154-is8630ik/files/wandb-summary.json b/wandb/run-20250508_212154-is8630ik/files/wandb-summary.json new file mode 100644 index 0000000000000000000000000000000000000000..24e2f76aa0afe1ec007e513e230825b879efc2a5 --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/files/wandb-summary.json @@ -0,0 +1 @@ +{"eval/hist_log_ratio":{"bins":[0.017333984375,3.017578125,6.015625,9.015625,12.015625,15.015625,18.015625,21.015625,24.015625,27.015625,30.015625,33.03125,36.03125,39.03125,42.03125,45.03125,48.03125,51.03125,54.03125,57.03125,60.03125,63.03125,66,69,72,75,78,81,84,87,90,93,96,99,102,105,108,111,114,117,120,123,126,129,132,135,138,141,144,147,150,153,156,159,162,165,168,171,174,177,180,183,186,189,192],"_type":"histogram","values":[1,1,1,1,7,6,4,12,13,22,35,42,28,38,37,25,32,23,14,18,12,9,26,13,24,15,24,13,29,20,29,11,21,16,29,11,13,14,14,5,19,11,12,1,6,6,3,1,9,4,3,1,1,2,2,0,3,1,2,0,0,1,0,1]},"train/lr":2.4246824991525087e-09,"eval/min_log_ratio":0.017333984375,"_timestamp":1.7467649187762368e+09,"_runtime":11206.604924269,"_step":16950,"train/epoch":3,"eval/step":16950,"_wandb":{"runtime":11206},"eval/mean_important_log_ratio":71.5,"eval/max_log_ratio":192,"train/step":16950,"train/loss":0.03305719792842865} \ No newline at end of file diff --git a/wandb/run-20250508_212154-is8630ik/logs/debug-core.log b/wandb/run-20250508_212154-is8630ik/logs/debug-core.log new file mode 100644 index 0000000000000000000000000000000000000000..860c5805585c2e11f073c67a20682476a780a0fd --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/logs/debug-core.log @@ -0,0 +1,15 @@ +{"time":"2025-05-08T21:21:54.231261668-04:00","level":"INFO","msg":"main: starting server","port-filename":"/tmp/tmpy4cwdbnf/port-1548317.txt","pid":1548317,"log-level":0,"disable-analytics":false,"shutdown-on-parent-exit":false} +{"time":"2025-05-08T21:21:54.237187926-04:00","level":"INFO","msg":"Will exit if parent process dies.","ppid":1548317} +{"time":"2025-05-08T21:21:54.237146955-04:00","level":"INFO","msg":"server is running","addr":{"IP":"127.0.0.1","Port":45581,"Zone":""}} +{"time":"2025-05-08T21:21:54.417525407-04:00","level":"INFO","msg":"connection: ManageConnectionData: new connection created","id":"127.0.0.1:60288"} +{"time":"2025-05-08T21:21:54.561577635-04:00","level":"INFO","msg":"handleInformInit: received","streamId":"is8630ik","id":"127.0.0.1:60288"} +{"time":"2025-05-08T21:21:54.740496215-04:00","level":"INFO","msg":"handleInformInit: stream started","streamId":"is8630ik","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:41.773667114-04:00","level":"INFO","msg":"handleInformFinish: finish message received","streamId":"is8630ik","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:41.773796316-04:00","level":"INFO","msg":"handleInformFinish: stream closed","streamId":"is8630ik","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:42.76336349-04:00","level":"INFO","msg":"handleInformTeardown: server teardown initiated","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:42.76339462-04:00","level":"INFO","msg":"handleInformTeardown: server shutdown complete","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:42.763405881-04:00","level":"INFO","msg":"server is shutting down"} +{"time":"2025-05-09T00:28:42.763424251-04:00","level":"INFO","msg":"connection: closing","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:42.763530093-04:00","level":"INFO","msg":"connection: closed successfully","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:42.763549613-04:00","level":"INFO","msg":"connection: ManageConnectionData: connection closed","id":"127.0.0.1:60288"} +{"time":"2025-05-09T00:28:42.763559364-04:00","level":"INFO","msg":"server is closed"} diff --git a/wandb/run-20250508_212154-is8630ik/logs/debug-internal.log b/wandb/run-20250508_212154-is8630ik/logs/debug-internal.log new file mode 100644 index 0000000000000000000000000000000000000000..d37033d7f8e5ba7af95d0fef4bc3dd9187388fa4 --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/logs/debug-internal.log @@ -0,0 +1,18 @@ +{"time":"2025-05-08T21:21:54.561866061-04:00","level":"INFO","msg":"stream: starting","core version":"0.19.10","symlink path":"/home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_212154-is8630ik/logs/debug-core.log"} +{"time":"2025-05-08T21:21:54.740453474-04:00","level":"INFO","msg":"created new stream","id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740487984-04:00","level":"INFO","msg":"stream: started","id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740504345-04:00","level":"INFO","msg":"writer: Do: started","stream_id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740556706-04:00","level":"INFO","msg":"handler: started","stream_id":"is8630ik"} +{"time":"2025-05-08T21:21:54.740592257-04:00","level":"INFO","msg":"sender: started","stream_id":"is8630ik"} +{"time":"2025-05-08T21:21:54.884261607-04:00","level":"INFO","msg":"Starting system monitor"} +{"time":"2025-05-08T21:21:54.884293377-04:00","level":"WARN","msg":"handleCodeSave: program relative path is empty"} +{"time":"2025-05-08T21:21:54.884348759-04:00","level":"ERROR","msg":"git repo not found","error":"repository does not exist"} +{"time":"2025-05-09T00:28:41.16342178-04:00","level":"INFO","msg":"Stopping system monitor"} +{"time":"2025-05-09T00:28:41.163499301-04:00","level":"INFO","msg":"Stopped system monitor"} +{"time":"2025-05-09T00:28:41.635621981-04:00","level":"INFO","msg":"fileTransfer: Close: file transfer manager closed"} +{"time":"2025-05-09T00:28:41.760403458-04:00","level":"INFO","msg":"handler: operation stats","stats":{}} +{"time":"2025-05-09T00:28:41.773696754-04:00","level":"INFO","msg":"stream: closing","id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773713434-04:00","level":"INFO","msg":"handler: closed","stream_id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773723895-04:00","level":"INFO","msg":"writer: Close: closed","stream_id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773731765-04:00","level":"INFO","msg":"sender: closed","stream_id":"is8630ik"} +{"time":"2025-05-09T00:28:41.773786876-04:00","level":"INFO","msg":"stream: closed","id":"is8630ik"} diff --git a/wandb/run-20250508_212154-is8630ik/logs/debug.log b/wandb/run-20250508_212154-is8630ik/logs/debug.log new file mode 100644 index 0000000000000000000000000000000000000000..c0d34bc4185ae49d7b75ac0313612476f5540c1c --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/logs/debug.log @@ -0,0 +1,29 @@ +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Current SDK version is 0.19.10 +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Configure stats pid to 1548317 +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Loading settings from /home/panda/.config/wandb/settings +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Loading settings from /home/panda/pda-llm/scripts/wandb/settings +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_setup.py:_flush():68] Loading settings from environment variables +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:setup_run_log_directory():724] Logging user logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_212154-is8630ik/logs/debug.log +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:setup_run_log_directory():725] Logging internal logs to /home/panda/pda-llm/output/sft-tools/run-false-1-10/wandb/run-20250508_212154-is8630ik/logs/debug-internal.log +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():852] calling init triggers +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():857] wandb.init called with sweep_config: {} +config: {'model_name_or_path': 'meta-llama/Llama-3.1-8B-Instruct', 'recompute_baseline': False, 'cache_dir': '/home/panda/pda-llm/cache/sft-tools', 'max_length': 2048, 'trust_remote_code': True, 'train_datasets': [('tools', {'proportion': 1.0})], 'eval_datasets': None, 'safety_ratio_tol': 10.0, 'important_sft': False, 'resilient_coeff': 1.0, 'epochs': 3, 'per_device_train_batch_size': 1, 'per_device_eval_batch_size': 1, 'gradient_accumulation_steps': 48, 'gradient_checkpointing': True, 'lr': 0.0001, 'lr_scheduler_type': , 'lr_warmup_ratio': 0.1, 'weight_decay': 0.0, 'seed': 42, 'fp16': False, 'bf16': True, 'tf32': False, 'lora_r': 16, 'lora_alpha': 32, 'lora_dropout': 0.05, 'eval_strategy': 'epoch', 'eval_interval': 1000000, 'need_eval': True, 'eval_split_ratio': None, 'output_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_type': 'wandb', 'log_dir': '/home/panda/pda-llm/output/sft-tools/run-false-1-10', 'log_project': 'TOOLS-SFT', 'log_run_name': 'tools-sft-2025-05-08-21-21-54', 'save_16bit': False, 'save_interval': 1000000, 'local_rank': 0, 'zero_stage': 0, 'offload': 'none', 'deepspeed': False, 'deepspeed_config': None, 'deepscale': False, 'deepscale_config': None, 'global_rank': 0, 'device': device(type='cuda', index=0), 'num_update_steps_per_epoch': 118, 'total_training_steps': 354, '_wandb': {}} +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():893] starting backend +2025-05-08 21:21:54,556 INFO MainThread:1548317 [wandb_init.py:init():897] sending inform_init request +2025-05-08 21:21:54,558 INFO MainThread:1548317 [backend.py:_multiprocessing_setup():101] multiprocessing start_methods=fork,spawn,forkserver, using: spawn +2025-05-08 21:21:54,558 INFO MainThread:1548317 [wandb_init.py:init():907] backend started and connected +2025-05-08 21:21:54,559 INFO MainThread:1548317 [wandb_init.py:init():1002] updated telemetry +2025-05-08 21:21:54,564 INFO MainThread:1548317 [wandb_init.py:init():1026] communicating run to backend with 90.0 second timeout +2025-05-08 21:21:54,882 INFO MainThread:1548317 [wandb_init.py:init():1101] starting run threads in backend +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_console_start():2566] atexit reg +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_redirect():2414] redirect: wrap_raw +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_redirect():2483] Wrapping output streams. +2025-05-08 21:21:54,947 INFO MainThread:1548317 [wandb_run.py:_redirect():2506] Redirects installed. +2025-05-08 21:21:54,948 INFO MainThread:1548317 [wandb_init.py:init():1147] run started, returning control to user process +2025-05-09 00:28:41,160 INFO MainThread:1548317 [wandb_run.py:_finish():2314] finishing run alelab/TOOLS-SFT/is8630ik +2025-05-09 00:28:41,162 INFO MainThread:1548317 [wandb_run.py:_atexit_cleanup():2531] got exitcode: 0 +2025-05-09 00:28:41,162 INFO MainThread:1548317 [wandb_run.py:_restore():2513] restore +2025-05-09 00:28:41,162 INFO MainThread:1548317 [wandb_run.py:_restore():2519] restore done +2025-05-09 00:28:41,762 INFO MainThread:1548317 [wandb_run.py:_footer_history_summary_info():4160] rendering history +2025-05-09 00:28:41,765 INFO MainThread:1548317 [wandb_run.py:_footer_history_summary_info():4192] rendering summary +2025-05-09 00:28:41,773 INFO MainThread:1548317 [wandb_run.py:_footer_sync_info():4121] logging synced files diff --git a/wandb/run-20250508_212154-is8630ik/run-is8630ik.wandb b/wandb/run-20250508_212154-is8630ik/run-is8630ik.wandb new file mode 100644 index 0000000000000000000000000000000000000000..6a569b1512558004fca4b166ee0ffaf404d777e0 --- /dev/null +++ b/wandb/run-20250508_212154-is8630ik/run-is8630ik.wandb @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67d65b4dc2493dac7ae1ddb04d23cfbac6e5a5a21bc3eccadf7336bf25261f72 +size 14778518