diff --git a/.gitattributes b/.gitattributes index a6344aac8c09253b3b630fb776ae94478aa0275b..e1ef2c46b8d03ade7702395bdb3ab135929610df 100644 --- a/.gitattributes +++ b/.gitattributes @@ -33,3 +33,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text *.zip filter=lfs diff=lfs merge=lfs -text *.zst filter=lfs diff=lfs merge=lfs -text *tfevents* filter=lfs diff=lfs merge=lfs -text +not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer.json filter=lfs diff=lfs merge=lfs -text +not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer.json filter=lfs diff=lfs merge=lfs -text +not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer.json filter=lfs diff=lfs merge=lfs -text +not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer.json filter=lfs diff=lfs merge=lfs -text +not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer.json filter=lfs diff=lfs merge=lfs -text diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/README.md b/not_grounded/0_128_e5_3e-5/checkpoint-1028/README.md new file mode 100644 index 0000000000000000000000000000000000000000..0994074add6527289eb93335320fbd12270c4e7f --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.12.0 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/adapter_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-1028/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2afdcf8899170963495ad14c67c715c56d8fd480 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/adapter_config.json @@ -0,0 +1,34 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 256, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 128, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "gate_proj", + "o_proj", + "v_proj", + "q_proj", + "down_proj", + "up_proj", + "k_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/adapter_model.safetensors b/not_grounded/0_128_e5_3e-5/checkpoint-1028/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b9e8c460d94ed8c935c9fcebbd8c5e5a01426a4 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:779fdbb3ebedfcd949ee1b6e0d38d772f8aff0044058a08627efa3579d2ea4b1 +size 671150064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/latest b/not_grounded/0_128_e5_3e-5/checkpoint-1028/latest new file mode 100644 index 0000000000000000000000000000000000000000..a8c19c91ca53931f4fac64b04c2f6c6d6e7f7cb3 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/latest @@ -0,0 +1 @@ +global_step1026 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/rng_state_0.pth b/not_grounded/0_128_e5_3e-5/checkpoint-1028/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..be290c01be235f06aa42ee37567eaa5cda94a766 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:be3ec1ce382b55ee0b40545251ad78363265ed23784c139bddb5293011c3969c +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/rng_state_1.pth b/not_grounded/0_128_e5_3e-5/checkpoint-1028/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..29c7cbe61c650a4acb462ea5b59093fb3ec27d7a --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3fc2354354f4c346cdea89bb1566c61c393d764ce3a31d8c36682258a12575b9 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/scheduler.pt b/not_grounded/0_128_e5_3e-5/checkpoint-1028/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..c09ec4ba0af08df17801d42a710e6d57571a132a --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:613316938e00be59ad361c40f96116a824f4991d399908b2bd99ff679482e6d5 +size 1064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/special_tokens_map.json b/not_grounded/0_128_e5_3e-5/checkpoint-1028/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..b43be96621d147110fb8a18b5776ec6e38516127 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer.json b/not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6ac83ff885a7c13c549808db25dd230af9d67648 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384a7e7c676f7be2e5d2e8449c508be9b00e5b18c5b3c39ebc626e96b3f4b988 +size 17210019 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..8ae1bc63bd6e5ca8a863628311061c143679ff93 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/tokenizer_config.json @@ -0,0 +1,2064 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizer" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/trainer_state.json b/not_grounded/0_128_e5_3e-5/checkpoint-1028/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..729e1f4d381cc77f70b9677c7b38f5b534cd5295 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/trainer_state.json @@ -0,0 +1,1469 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.0, + "eval_steps": 500, + "global_step": 1028, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.01949317738791423, + "grad_norm": 0.6081938743591309, + "learning_rate": 1.875e-06, + "loss": 1.7943, + "step": 5 + }, + { + "epoch": 0.03898635477582846, + "grad_norm": 0.5703465938568115, + "learning_rate": 4.21875e-06, + "loss": 1.7931, + "step": 10 + }, + { + "epoch": 0.05847953216374269, + "grad_norm": 0.4802420139312744, + "learning_rate": 6.5625e-06, + "loss": 1.8677, + "step": 15 + }, + { + "epoch": 0.07797270955165692, + "grad_norm": 0.4527171552181244, + "learning_rate": 8.90625e-06, + "loss": 1.7498, + "step": 20 + }, + { + "epoch": 0.09746588693957114, + "grad_norm": 0.41348591446876526, + "learning_rate": 1.125e-05, + "loss": 1.7402, + "step": 25 + }, + { + "epoch": 0.11695906432748537, + "grad_norm": 0.44569578766822815, + "learning_rate": 1.359375e-05, + "loss": 1.6984, + "step": 30 + }, + { + "epoch": 0.1364522417153996, + "grad_norm": 0.439203143119812, + "learning_rate": 1.59375e-05, + "loss": 1.6548, + "step": 35 + }, + { + "epoch": 0.15594541910331383, + "grad_norm": 0.4098866879940033, + "learning_rate": 1.828125e-05, + "loss": 1.7276, + "step": 40 + }, + { + "epoch": 0.17543859649122806, + "grad_norm": 0.42010200023651123, + "learning_rate": 2.0625e-05, + "loss": 1.7037, + "step": 45 + }, + { + "epoch": 0.1949317738791423, + "grad_norm": 0.4497705399990082, + "learning_rate": 2.296875e-05, + "loss": 1.6829, + "step": 50 + }, + { + "epoch": 0.21442495126705652, + "grad_norm": 0.4349888563156128, + "learning_rate": 2.5312500000000002e-05, + "loss": 1.696, + "step": 55 + }, + { + "epoch": 0.23391812865497075, + "grad_norm": 0.45078787207603455, + "learning_rate": 2.765625e-05, + "loss": 1.6246, + "step": 60 + }, + { + "epoch": 0.253411306042885, + "grad_norm": 0.44585877656936646, + "learning_rate": 3e-05, + "loss": 1.6324, + "step": 65 + }, + { + "epoch": 0.2729044834307992, + "grad_norm": 0.4684946537017822, + "learning_rate": 2.9998748508718575e-05, + "loss": 1.5858, + "step": 70 + }, + { + "epoch": 0.29239766081871343, + "grad_norm": 0.5270784497261047, + "learning_rate": 2.9994994243705013e-05, + "loss": 1.5719, + "step": 75 + }, + { + "epoch": 0.31189083820662766, + "grad_norm": 0.5215119123458862, + "learning_rate": 2.9988737831416642e-05, + "loss": 1.5858, + "step": 80 + }, + { + "epoch": 0.3313840155945419, + "grad_norm": 0.5029568672180176, + "learning_rate": 2.9979980315832853e-05, + "loss": 1.4428, + "step": 85 + }, + { + "epoch": 0.3508771929824561, + "grad_norm": 0.5102682709693909, + "learning_rate": 2.9968723158280906e-05, + "loss": 1.5211, + "step": 90 + }, + { + "epoch": 0.37037037037037035, + "grad_norm": 0.5617806911468506, + "learning_rate": 2.995496823719206e-05, + "loss": 1.4377, + "step": 95 + }, + { + "epoch": 0.3898635477582846, + "grad_norm": 0.5236561298370361, + "learning_rate": 2.9938717847788167e-05, + "loss": 1.464, + "step": 100 + }, + { + "epoch": 0.4093567251461988, + "grad_norm": 0.6214772462844849, + "learning_rate": 2.9919974701698638e-05, + "loss": 1.4009, + "step": 105 + }, + { + "epoch": 0.42884990253411304, + "grad_norm": 0.5746711492538452, + "learning_rate": 2.989874192650801e-05, + "loss": 1.3927, + "step": 110 + }, + { + "epoch": 0.44834307992202727, + "grad_norm": 0.6362848877906799, + "learning_rate": 2.9875023065234003e-05, + "loss": 1.4212, + "step": 115 + }, + { + "epoch": 0.4678362573099415, + "grad_norm": 0.6465691328048706, + "learning_rate": 2.984882207573638e-05, + "loss": 1.4462, + "step": 120 + }, + { + "epoch": 0.4873294346978557, + "grad_norm": 0.6339296698570251, + "learning_rate": 2.982014333005645e-05, + "loss": 1.337, + "step": 125 + }, + { + "epoch": 0.50682261208577, + "grad_norm": 0.6140757203102112, + "learning_rate": 2.9788991613687575e-05, + "loss": 1.3289, + "step": 130 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 0.6720002293586731, + "learning_rate": 2.9755372124776616e-05, + "loss": 1.3323, + "step": 135 + }, + { + "epoch": 0.5458089668615984, + "grad_norm": 0.6524502038955688, + "learning_rate": 2.971929047325654e-05, + "loss": 1.3166, + "step": 140 + }, + { + "epoch": 0.5653021442495126, + "grad_norm": 0.7752301692962646, + "learning_rate": 2.968075267991032e-05, + "loss": 1.3543, + "step": 145 + }, + { + "epoch": 0.5847953216374269, + "grad_norm": 0.6975963711738586, + "learning_rate": 2.963976517536627e-05, + "loss": 1.2675, + "step": 150 + }, + { + "epoch": 0.6042884990253411, + "grad_norm": 0.7260017991065979, + "learning_rate": 2.9596334799025003e-05, + "loss": 1.3213, + "step": 155 + }, + { + "epoch": 0.6237816764132553, + "grad_norm": 0.8624387979507446, + "learning_rate": 2.9550468797918162e-05, + "loss": 1.2492, + "step": 160 + }, + { + "epoch": 0.6432748538011696, + "grad_norm": 0.7931470274925232, + "learning_rate": 2.950217482549915e-05, + "loss": 1.2353, + "step": 165 + }, + { + "epoch": 0.6627680311890838, + "grad_norm": 0.7223592400550842, + "learning_rate": 2.9451460940366025e-05, + "loss": 1.1868, + "step": 170 + }, + { + "epoch": 0.682261208576998, + "grad_norm": 0.7910372614860535, + "learning_rate": 2.9398335604916797e-05, + "loss": 1.2148, + "step": 175 + }, + { + "epoch": 0.7017543859649122, + "grad_norm": 0.851186990737915, + "learning_rate": 2.9342807683937352e-05, + "loss": 1.2595, + "step": 180 + }, + { + "epoch": 0.7212475633528265, + "grad_norm": 0.8052693009376526, + "learning_rate": 2.928488644312222e-05, + "loss": 1.2406, + "step": 185 + }, + { + "epoch": 0.7407407407407407, + "grad_norm": 0.8178960680961609, + "learning_rate": 2.9224581547528453e-05, + "loss": 1.2197, + "step": 190 + }, + { + "epoch": 0.7602339181286549, + "grad_norm": 0.8441738486289978, + "learning_rate": 2.916190305996286e-05, + "loss": 1.1726, + "step": 195 + }, + { + "epoch": 0.7797270955165692, + "grad_norm": 0.9862117767333984, + "learning_rate": 2.909686143930287e-05, + "loss": 1.1745, + "step": 200 + }, + { + "epoch": 0.7992202729044834, + "grad_norm": 0.8886749148368835, + "learning_rate": 2.902946753875131e-05, + "loss": 1.1447, + "step": 205 + }, + { + "epoch": 0.8187134502923976, + "grad_norm": 0.8914117813110352, + "learning_rate": 2.895973260402537e-05, + "loss": 1.1248, + "step": 210 + }, + { + "epoch": 0.8382066276803118, + "grad_norm": 1.1335160732269287, + "learning_rate": 2.8887668271480098e-05, + "loss": 1.1291, + "step": 215 + }, + { + "epoch": 0.8576998050682261, + "grad_norm": 1.0726529359817505, + "learning_rate": 2.8813286566166674e-05, + "loss": 1.1646, + "step": 220 + }, + { + "epoch": 0.8771929824561403, + "grad_norm": 1.0396595001220703, + "learning_rate": 2.873659989982586e-05, + "loss": 1.0894, + "step": 225 + }, + { + "epoch": 0.8966861598440545, + "grad_norm": 0.9956820011138916, + "learning_rate": 2.8657621068816903e-05, + "loss": 1.0984, + "step": 230 + }, + { + "epoch": 0.9161793372319688, + "grad_norm": 0.9483380317687988, + "learning_rate": 2.857636325198225e-05, + "loss": 1.0216, + "step": 235 + }, + { + "epoch": 0.935672514619883, + "grad_norm": 0.8558198809623718, + "learning_rate": 2.849284000844849e-05, + "loss": 1.105, + "step": 240 + }, + { + "epoch": 0.9551656920077972, + "grad_norm": 1.005858063697815, + "learning_rate": 2.8407065275363756e-05, + "loss": 1.0714, + "step": 245 + }, + { + "epoch": 0.9746588693957114, + "grad_norm": 0.9553005695343018, + "learning_rate": 2.8319053365572135e-05, + "loss": 1.0339, + "step": 250 + }, + { + "epoch": 0.9941520467836257, + "grad_norm": 0.9755716919898987, + "learning_rate": 2.8228818965225325e-05, + "loss": 1.0532, + "step": 255 + }, + { + "epoch": 1.0116959064327484, + "grad_norm": 1.0532433986663818, + "learning_rate": 2.8136377131332043e-05, + "loss": 1.0005, + "step": 260 + }, + { + "epoch": 1.0311890838206628, + "grad_norm": 0.9859170913696289, + "learning_rate": 2.8041743289245503e-05, + "loss": 0.9084, + "step": 265 + }, + { + "epoch": 1.050682261208577, + "grad_norm": 0.9481168389320374, + "learning_rate": 2.7944933230089484e-05, + "loss": 0.9431, + "step": 270 + }, + { + "epoch": 1.0701754385964912, + "grad_norm": 0.9678940773010254, + "learning_rate": 2.784596310812331e-05, + "loss": 0.9262, + "step": 275 + }, + { + "epoch": 1.0896686159844053, + "grad_norm": 0.9351955652236938, + "learning_rate": 2.774484943804629e-05, + "loss": 0.9044, + "step": 280 + }, + { + "epoch": 1.1091617933723197, + "grad_norm": 1.0261086225509644, + "learning_rate": 2.764160909224196e-05, + "loss": 0.9579, + "step": 285 + }, + { + "epoch": 1.128654970760234, + "grad_norm": 0.9796629548072815, + "learning_rate": 2.7536259297962674e-05, + "loss": 0.9283, + "step": 290 + }, + { + "epoch": 1.1481481481481481, + "grad_norm": 0.9690019488334656, + "learning_rate": 2.7428817634454973e-05, + "loss": 0.9017, + "step": 295 + }, + { + "epoch": 1.1676413255360623, + "grad_norm": 1.1676615476608276, + "learning_rate": 2.7319302030026207e-05, + "loss": 0.8919, + "step": 300 + }, + { + "epoch": 1.1871345029239766, + "grad_norm": 1.1467899084091187, + "learning_rate": 2.7207730759052925e-05, + "loss": 0.871, + "step": 305 + }, + { + "epoch": 1.206627680311891, + "grad_norm": 1.0315958261489868, + "learning_rate": 2.7094122438931513e-05, + "loss": 0.8388, + "step": 310 + }, + { + "epoch": 1.226120857699805, + "grad_norm": 1.0231361389160156, + "learning_rate": 2.697849602697159e-05, + "loss": 0.8841, + "step": 315 + }, + { + "epoch": 1.2456140350877192, + "grad_norm": 0.995548665523529, + "learning_rate": 2.6860870817232682e-05, + "loss": 0.8879, + "step": 320 + }, + { + "epoch": 1.2651072124756335, + "grad_norm": 1.102622151374817, + "learning_rate": 2.6741266437304718e-05, + "loss": 0.857, + "step": 325 + }, + { + "epoch": 1.2846003898635479, + "grad_norm": 1.3752007484436035, + "learning_rate": 2.661970284503286e-05, + "loss": 0.8027, + "step": 330 + }, + { + "epoch": 1.304093567251462, + "grad_norm": 1.129936933517456, + "learning_rate": 2.6496200325187223e-05, + "loss": 0.855, + "step": 335 + }, + { + "epoch": 1.323586744639376, + "grad_norm": 1.2070459127426147, + "learning_rate": 2.6370779486078047e-05, + "loss": 0.8544, + "step": 340 + }, + { + "epoch": 1.3430799220272904, + "grad_norm": 1.1400156021118164, + "learning_rate": 2.6243461256116892e-05, + "loss": 0.9026, + "step": 345 + }, + { + "epoch": 1.3625730994152048, + "grad_norm": 0.9957541823387146, + "learning_rate": 2.611426688032439e-05, + "loss": 0.8115, + "step": 350 + }, + { + "epoch": 1.3820662768031189, + "grad_norm": 1.0340908765792847, + "learning_rate": 2.598321791678519e-05, + "loss": 0.8314, + "step": 355 + }, + { + "epoch": 1.401559454191033, + "grad_norm": 1.1327755451202393, + "learning_rate": 2.5850336233050677e-05, + "loss": 0.8319, + "step": 360 + }, + { + "epoch": 1.4210526315789473, + "grad_norm": 1.0419026613235474, + "learning_rate": 2.5715644002489998e-05, + "loss": 0.8115, + "step": 365 + }, + { + "epoch": 1.4405458089668617, + "grad_norm": 1.080306053161621, + "learning_rate": 2.557916370059012e-05, + "loss": 0.7734, + "step": 370 + }, + { + "epoch": 1.4600389863547758, + "grad_norm": 1.1247351169586182, + "learning_rate": 2.544091810120543e-05, + "loss": 0.7968, + "step": 375 + }, + { + "epoch": 1.47953216374269, + "grad_norm": 1.1105760335922241, + "learning_rate": 2.530093027275757e-05, + "loss": 0.7994, + "step": 380 + }, + { + "epoch": 1.4990253411306043, + "grad_norm": 1.3396145105361938, + "learning_rate": 2.5159223574386117e-05, + "loss": 0.7891, + "step": 385 + }, + { + "epoch": 1.5185185185185186, + "grad_norm": 1.2513933181762695, + "learning_rate": 2.501582165205074e-05, + "loss": 0.7825, + "step": 390 + }, + { + "epoch": 1.5380116959064327, + "grad_norm": 1.139088749885559, + "learning_rate": 2.4870748434585514e-05, + "loss": 0.7875, + "step": 395 + }, + { + "epoch": 1.5575048732943468, + "grad_norm": 1.2570606470108032, + "learning_rate": 2.4724028129706012e-05, + "loss": 0.7524, + "step": 400 + }, + { + "epoch": 1.5769980506822612, + "grad_norm": 1.115220069885254, + "learning_rate": 2.4575685219969884e-05, + "loss": 0.7855, + "step": 405 + }, + { + "epoch": 1.5964912280701755, + "grad_norm": 1.2705992460250854, + "learning_rate": 2.442574445869156e-05, + "loss": 0.7338, + "step": 410 + }, + { + "epoch": 1.6159844054580896, + "grad_norm": 1.1325453519821167, + "learning_rate": 2.4274230865811763e-05, + "loss": 0.7528, + "step": 415 + }, + { + "epoch": 1.6354775828460038, + "grad_norm": 1.301644206047058, + "learning_rate": 2.4121169723722566e-05, + "loss": 0.7445, + "step": 420 + }, + { + "epoch": 1.654970760233918, + "grad_norm": 1.4577105045318604, + "learning_rate": 2.396658657304861e-05, + "loss": 0.7335, + "step": 425 + }, + { + "epoch": 1.6744639376218324, + "grad_norm": 1.1624274253845215, + "learning_rate": 2.381050720838528e-05, + "loss": 0.7146, + "step": 430 + }, + { + "epoch": 1.6939571150097466, + "grad_norm": 1.4543778896331787, + "learning_rate": 2.3652957673994448e-05, + "loss": 0.7338, + "step": 435 + }, + { + "epoch": 1.7134502923976607, + "grad_norm": 1.208083987236023, + "learning_rate": 2.3493964259458603e-05, + "loss": 0.7187, + "step": 440 + }, + { + "epoch": 1.732943469785575, + "grad_norm": 1.1645222902297974, + "learning_rate": 2.3333553495294033e-05, + "loss": 0.7325, + "step": 445 + }, + { + "epoch": 1.7524366471734893, + "grad_norm": 1.2736948728561401, + "learning_rate": 2.317175214852377e-05, + "loss": 0.7152, + "step": 450 + }, + { + "epoch": 1.7719298245614035, + "grad_norm": 1.2995411157608032, + "learning_rate": 2.3008587218211127e-05, + "loss": 0.6953, + "step": 455 + }, + { + "epoch": 1.7914230019493176, + "grad_norm": 1.4111621379852295, + "learning_rate": 2.284408593095446e-05, + "loss": 0.7023, + "step": 460 + }, + { + "epoch": 1.810916179337232, + "grad_norm": 1.2578918933868408, + "learning_rate": 2.2678275736344014e-05, + "loss": 0.6771, + "step": 465 + }, + { + "epoch": 1.8304093567251463, + "grad_norm": 1.3310747146606445, + "learning_rate": 2.251118430238151e-05, + "loss": 0.6792, + "step": 470 + }, + { + "epoch": 1.8499025341130604, + "grad_norm": 1.1377298831939697, + "learning_rate": 2.2342839510863323e-05, + "loss": 0.7322, + "step": 475 + }, + { + "epoch": 1.8693957115009745, + "grad_norm": 1.2897000312805176, + "learning_rate": 2.2173269452727965e-05, + "loss": 0.6468, + "step": 480 + }, + { + "epoch": 1.8888888888888888, + "grad_norm": 1.1680046319961548, + "learning_rate": 2.200250242336868e-05, + "loss": 0.6785, + "step": 485 + }, + { + "epoch": 1.9083820662768032, + "grad_norm": 1.2389404773712158, + "learning_rate": 2.183056691791193e-05, + "loss": 0.6521, + "step": 490 + }, + { + "epoch": 1.9278752436647173, + "grad_norm": 1.1207393407821655, + "learning_rate": 2.1657491626462514e-05, + "loss": 0.6607, + "step": 495 + }, + { + "epoch": 1.9473684210526314, + "grad_norm": 1.1037189960479736, + "learning_rate": 2.1483305429316208e-05, + "loss": 0.6033, + "step": 500 + }, + { + "epoch": 1.9668615984405458, + "grad_norm": 1.1787821054458618, + "learning_rate": 2.1308037392140613e-05, + "loss": 0.6254, + "step": 505 + }, + { + "epoch": 1.98635477582846, + "grad_norm": 1.266696572303772, + "learning_rate": 2.113171676112513e-05, + "loss": 0.5962, + "step": 510 + }, + { + "epoch": 2.003898635477583, + "grad_norm": 1.4790624380111694, + "learning_rate": 2.0954372958100746e-05, + "loss": 0.6104, + "step": 515 + }, + { + "epoch": 2.023391812865497, + "grad_norm": 1.6624841690063477, + "learning_rate": 2.0776035575630575e-05, + "loss": 0.5632, + "step": 520 + }, + { + "epoch": 2.042884990253411, + "grad_norm": 1.243015170097351, + "learning_rate": 2.0596734372071855e-05, + "loss": 0.5019, + "step": 525 + }, + { + "epoch": 2.0623781676413255, + "grad_norm": 1.4598456621170044, + "learning_rate": 2.0416499266610316e-05, + "loss": 0.5299, + "step": 530 + }, + { + "epoch": 2.08187134502924, + "grad_norm": 1.2799503803253174, + "learning_rate": 2.023536033426771e-05, + "loss": 0.5432, + "step": 535 + }, + { + "epoch": 2.101364522417154, + "grad_norm": 1.3306210041046143, + "learning_rate": 2.00533478008833e-05, + "loss": 0.5639, + "step": 540 + }, + { + "epoch": 2.120857699805068, + "grad_norm": 1.5138165950775146, + "learning_rate": 1.9870492038070255e-05, + "loss": 0.519, + "step": 545 + }, + { + "epoch": 2.1403508771929824, + "grad_norm": 1.283997893333435, + "learning_rate": 1.9686823558147616e-05, + "loss": 0.5276, + "step": 550 + }, + { + "epoch": 2.159844054580897, + "grad_norm": 1.2089860439300537, + "learning_rate": 1.9502373009048898e-05, + "loss": 0.5384, + "step": 555 + }, + { + "epoch": 2.1793372319688107, + "grad_norm": 1.327548861503601, + "learning_rate": 1.9317171169207973e-05, + "loss": 0.5297, + "step": 560 + }, + { + "epoch": 2.198830409356725, + "grad_norm": 1.3738446235656738, + "learning_rate": 1.913124894242322e-05, + "loss": 0.5324, + "step": 565 + }, + { + "epoch": 2.2183235867446394, + "grad_norm": 1.2523514032363892, + "learning_rate": 1.8944637352700754e-05, + "loss": 0.4629, + "step": 570 + }, + { + "epoch": 2.2378167641325537, + "grad_norm": 1.4739748239517212, + "learning_rate": 1.875736753907758e-05, + "loss": 0.487, + "step": 575 + }, + { + "epoch": 2.257309941520468, + "grad_norm": 1.3324633836746216, + "learning_rate": 1.856947075042557e-05, + "loss": 0.4968, + "step": 580 + }, + { + "epoch": 2.276803118908382, + "grad_norm": 1.3165260553359985, + "learning_rate": 1.8380978340237094e-05, + "loss": 0.5223, + "step": 585 + }, + { + "epoch": 2.2962962962962963, + "grad_norm": 1.3404366970062256, + "learning_rate": 1.8191921761393222e-05, + "loss": 0.4838, + "step": 590 + }, + { + "epoch": 2.3157894736842106, + "grad_norm": 1.3298143148422241, + "learning_rate": 1.8002332560915297e-05, + "loss": 0.4742, + "step": 595 + }, + { + "epoch": 2.3352826510721245, + "grad_norm": 1.2816650867462158, + "learning_rate": 1.7812242374700847e-05, + "loss": 0.4775, + "step": 600 + }, + { + "epoch": 2.354775828460039, + "grad_norm": 1.406812310218811, + "learning_rate": 1.7621682922244637e-05, + "loss": 0.5127, + "step": 605 + }, + { + "epoch": 2.374269005847953, + "grad_norm": 1.3161938190460205, + "learning_rate": 1.7430686001345777e-05, + "loss": 0.4723, + "step": 610 + }, + { + "epoch": 2.3937621832358675, + "grad_norm": 1.4261293411254883, + "learning_rate": 1.7239283482801783e-05, + "loss": 0.5121, + "step": 615 + }, + { + "epoch": 2.413255360623782, + "grad_norm": 1.2458994388580322, + "learning_rate": 1.7047507305090397e-05, + "loss": 0.4821, + "step": 620 + }, + { + "epoch": 2.4327485380116958, + "grad_norm": 1.3558402061462402, + "learning_rate": 1.6855389469040217e-05, + "loss": 0.4423, + "step": 625 + }, + { + "epoch": 2.45224171539961, + "grad_norm": 1.2846639156341553, + "learning_rate": 1.6662962032490814e-05, + "loss": 0.4964, + "step": 630 + }, + { + "epoch": 2.4717348927875245, + "grad_norm": 1.3657177686691284, + "learning_rate": 1.6470257104943414e-05, + "loss": 0.4906, + "step": 635 + }, + { + "epoch": 2.4912280701754383, + "grad_norm": 1.3313744068145752, + "learning_rate": 1.6277306842202902e-05, + "loss": 0.4824, + "step": 640 + }, + { + "epoch": 2.5107212475633527, + "grad_norm": 1.427259922027588, + "learning_rate": 1.6084143441012158e-05, + "loss": 0.5067, + "step": 645 + }, + { + "epoch": 2.530214424951267, + "grad_norm": 1.3357858657836914, + "learning_rate": 1.5890799133679523e-05, + "loss": 0.5128, + "step": 650 + }, + { + "epoch": 2.5497076023391814, + "grad_norm": 1.6301915645599365, + "learning_rate": 1.569730618270031e-05, + "loss": 0.4331, + "step": 655 + }, + { + "epoch": 2.5692007797270957, + "grad_norm": 1.4900562763214111, + "learning_rate": 1.5503696875373353e-05, + "loss": 0.4391, + "step": 660 + }, + { + "epoch": 2.5886939571150096, + "grad_norm": 1.402565836906433, + "learning_rate": 1.5310003518413318e-05, + "loss": 0.4457, + "step": 665 + }, + { + "epoch": 2.608187134502924, + "grad_norm": 1.3605144023895264, + "learning_rate": 1.5116258432559883e-05, + "loss": 0.4546, + "step": 670 + }, + { + "epoch": 2.6276803118908383, + "grad_norm": 1.3472214937210083, + "learning_rate": 1.4922493947184482e-05, + "loss": 0.4791, + "step": 675 + }, + { + "epoch": 2.647173489278752, + "grad_norm": 1.4113649129867554, + "learning_rate": 1.4728742394895666e-05, + "loss": 0.4364, + "step": 680 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 1.3126051425933838, + "learning_rate": 1.4535036106143895e-05, + "loss": 0.4309, + "step": 685 + }, + { + "epoch": 2.686159844054581, + "grad_norm": 1.4124526977539062, + "learning_rate": 1.4341407403826695e-05, + "loss": 0.4346, + "step": 690 + }, + { + "epoch": 2.705653021442495, + "grad_norm": 1.4493342638015747, + "learning_rate": 1.4147888597895117e-05, + "loss": 0.4175, + "step": 695 + }, + { + "epoch": 2.7251461988304095, + "grad_norm": 1.4753097295761108, + "learning_rate": 1.3954511979962271e-05, + "loss": 0.4275, + "step": 700 + }, + { + "epoch": 2.7446393762183234, + "grad_norm": 1.2626641988754272, + "learning_rate": 1.3761309817915017e-05, + "loss": 0.4182, + "step": 705 + }, + { + "epoch": 2.7641325536062378, + "grad_norm": 1.526242971420288, + "learning_rate": 1.3568314350529534e-05, + "loss": 0.3954, + "step": 710 + }, + { + "epoch": 2.783625730994152, + "grad_norm": 1.5078595876693726, + "learning_rate": 1.3375557782091789e-05, + "loss": 0.4406, + "step": 715 + }, + { + "epoch": 2.803118908382066, + "grad_norm": 1.398832082748413, + "learning_rate": 1.3183072277023766e-05, + "loss": 0.4096, + "step": 720 + }, + { + "epoch": 2.8226120857699804, + "grad_norm": 1.4593217372894287, + "learning_rate": 1.2990889954516313e-05, + "loss": 0.4015, + "step": 725 + }, + { + "epoch": 2.8421052631578947, + "grad_norm": 1.5111234188079834, + "learning_rate": 1.2799042883169576e-05, + "loss": 0.4074, + "step": 730 + }, + { + "epoch": 2.861598440545809, + "grad_norm": 1.4062628746032715, + "learning_rate": 1.2607563075641838e-05, + "loss": 0.4055, + "step": 735 + }, + { + "epoch": 2.8810916179337234, + "grad_norm": 1.3948218822479248, + "learning_rate": 1.2416482483307728e-05, + "loss": 0.4129, + "step": 740 + }, + { + "epoch": 2.9005847953216373, + "grad_norm": 1.408527135848999, + "learning_rate": 1.2225832990926625e-05, + "loss": 0.4073, + "step": 745 + }, + { + "epoch": 2.9200779727095516, + "grad_norm": 1.4006417989730835, + "learning_rate": 1.20356464113222e-05, + "loss": 0.4147, + "step": 750 + }, + { + "epoch": 2.939571150097466, + "grad_norm": 1.3896782398223877, + "learning_rate": 1.184595448007395e-05, + "loss": 0.3975, + "step": 755 + }, + { + "epoch": 2.95906432748538, + "grad_norm": 1.4923348426818848, + "learning_rate": 1.1656788850221617e-05, + "loss": 0.3933, + "step": 760 + }, + { + "epoch": 2.978557504873294, + "grad_norm": 1.3745965957641602, + "learning_rate": 1.1468181086983414e-05, + "loss": 0.395, + "step": 765 + }, + { + "epoch": 2.9980506822612085, + "grad_norm": 1.429971694946289, + "learning_rate": 1.1280162662488834e-05, + "loss": 0.4257, + "step": 770 + }, + { + "epoch": 3.0155945419103314, + "grad_norm": 1.4495044946670532, + "learning_rate": 1.1092764950527088e-05, + "loss": 0.3595, + "step": 775 + }, + { + "epoch": 3.0350877192982457, + "grad_norm": 1.3142398595809937, + "learning_rate": 1.090601922131187e-05, + "loss": 0.278, + "step": 780 + }, + { + "epoch": 3.0545808966861596, + "grad_norm": 1.3927865028381348, + "learning_rate": 1.0719956636263424e-05, + "loss": 0.2914, + "step": 785 + }, + { + "epoch": 3.074074074074074, + "grad_norm": 1.6307252645492554, + "learning_rate": 1.0534608242808827e-05, + "loss": 0.3096, + "step": 790 + }, + { + "epoch": 3.0935672514619883, + "grad_norm": 1.4899730682373047, + "learning_rate": 1.03500049692012e-05, + "loss": 0.3015, + "step": 795 + }, + { + "epoch": 3.1130604288499026, + "grad_norm": 1.6715621948242188, + "learning_rate": 1.0166177619358873e-05, + "loss": 0.316, + "step": 800 + }, + { + "epoch": 3.132553606237817, + "grad_norm": 1.5180559158325195, + "learning_rate": 9.983156867725257e-06, + "loss": 0.3194, + "step": 805 + }, + { + "epoch": 3.152046783625731, + "grad_norm": 1.4555518627166748, + "learning_rate": 9.800973254150351e-06, + "loss": 0.3059, + "step": 810 + }, + { + "epoch": 3.171539961013645, + "grad_norm": 1.3086048364639282, + "learning_rate": 9.619657178794692e-06, + "loss": 0.3327, + "step": 815 + }, + { + "epoch": 3.1910331384015596, + "grad_norm": 1.4091737270355225, + "learning_rate": 9.439238897056612e-06, + "loss": 0.3313, + "step": 820 + }, + { + "epoch": 3.2105263157894735, + "grad_norm": 1.4301328659057617, + "learning_rate": 9.259748514523654e-06, + "loss": 0.3003, + "step": 825 + }, + { + "epoch": 3.230019493177388, + "grad_norm": 1.4087915420532227, + "learning_rate": 9.081215981949002e-06, + "loss": 0.3249, + "step": 830 + }, + { + "epoch": 3.249512670565302, + "grad_norm": 1.4424235820770264, + "learning_rate": 8.903671090253717e-06, + "loss": 0.3117, + "step": 835 + }, + { + "epoch": 3.2690058479532165, + "grad_norm": 1.3572005033493042, + "learning_rate": 8.727143465555668e-06, + "loss": 0.3062, + "step": 840 + }, + { + "epoch": 3.288499025341131, + "grad_norm": 1.4524186849594116, + "learning_rate": 8.551662564225952e-06, + "loss": 0.3053, + "step": 845 + }, + { + "epoch": 3.3079922027290447, + "grad_norm": 1.3756184577941895, + "learning_rate": 8.377257667973644e-06, + "loss": 0.333, + "step": 850 + }, + { + "epoch": 3.327485380116959, + "grad_norm": 1.5087087154388428, + "learning_rate": 8.203957878959693e-06, + "loss": 0.2984, + "step": 855 + }, + { + "epoch": 3.3469785575048734, + "grad_norm": 1.7474857568740845, + "learning_rate": 8.031792114940777e-06, + "loss": 0.3344, + "step": 860 + }, + { + "epoch": 3.3664717348927873, + "grad_norm": 1.4694979190826416, + "learning_rate": 7.860789104443897e-06, + "loss": 0.2761, + "step": 865 + }, + { + "epoch": 3.3859649122807016, + "grad_norm": 1.578460931777954, + "learning_rate": 7.690977381972626e-06, + "loss": 0.2994, + "step": 870 + }, + { + "epoch": 3.405458089668616, + "grad_norm": 1.3219410181045532, + "learning_rate": 7.5223852832456585e-06, + "loss": 0.273, + "step": 875 + }, + { + "epoch": 3.4249512670565303, + "grad_norm": 1.3564660549163818, + "learning_rate": 7.3550409404685386e-06, + "loss": 0.2994, + "step": 880 + }, + { + "epoch": 3.4444444444444446, + "grad_norm": 1.562962293624878, + "learning_rate": 7.188972277639406e-06, + "loss": 0.3037, + "step": 885 + }, + { + "epoch": 3.4639376218323585, + "grad_norm": 1.477868914604187, + "learning_rate": 7.024207005889406e-06, + "loss": 0.2856, + "step": 890 + }, + { + "epoch": 3.483430799220273, + "grad_norm": 1.406184434890747, + "learning_rate": 6.8607726188586864e-06, + "loss": 0.2949, + "step": 895 + }, + { + "epoch": 3.502923976608187, + "grad_norm": 1.4179635047912598, + "learning_rate": 6.698696388108642e-06, + "loss": 0.2876, + "step": 900 + }, + { + "epoch": 3.522417153996101, + "grad_norm": 1.3572025299072266, + "learning_rate": 6.538005358571235e-06, + "loss": 0.2922, + "step": 905 + }, + { + "epoch": 3.5419103313840155, + "grad_norm": 1.3908215761184692, + "learning_rate": 6.37872634403613e-06, + "loss": 0.2695, + "step": 910 + }, + { + "epoch": 3.56140350877193, + "grad_norm": 1.5078115463256836, + "learning_rate": 6.220885922676385e-06, + "loss": 0.3221, + "step": 915 + }, + { + "epoch": 3.580896686159844, + "grad_norm": 1.4563665390014648, + "learning_rate": 6.0645104326135e-06, + "loss": 0.2814, + "step": 920 + }, + { + "epoch": 3.6003898635477585, + "grad_norm": 1.452856421470642, + "learning_rate": 5.909625967522465e-06, + "loss": 0.288, + "step": 925 + }, + { + "epoch": 3.6198830409356724, + "grad_norm": 1.5082298517227173, + "learning_rate": 5.7562583722776435e-06, + "loss": 0.2762, + "step": 930 + }, + { + "epoch": 3.6393762183235867, + "grad_norm": 1.5691982507705688, + "learning_rate": 5.604433238640134e-06, + "loss": 0.2858, + "step": 935 + }, + { + "epoch": 3.658869395711501, + "grad_norm": 1.399603247642517, + "learning_rate": 5.454175900987411e-06, + "loss": 0.2698, + "step": 940 + }, + { + "epoch": 3.678362573099415, + "grad_norm": 1.2988371849060059, + "learning_rate": 5.305511432085885e-06, + "loss": 0.2611, + "step": 945 + }, + { + "epoch": 3.6978557504873293, + "grad_norm": 1.612922191619873, + "learning_rate": 5.158464638907113e-06, + "loss": 0.2704, + "step": 950 + }, + { + "epoch": 3.7173489278752436, + "grad_norm": 1.4260296821594238, + "learning_rate": 5.01306005848838e-06, + "loss": 0.2624, + "step": 955 + }, + { + "epoch": 3.736842105263158, + "grad_norm": 1.299452304840088, + "learning_rate": 4.869321953838305e-06, + "loss": 0.2503, + "step": 960 + }, + { + "epoch": 3.7563352826510723, + "grad_norm": 1.5614985227584839, + "learning_rate": 4.7272743098881915e-06, + "loss": 0.2642, + "step": 965 + }, + { + "epoch": 3.775828460038986, + "grad_norm": 1.5323383808135986, + "learning_rate": 4.586940829489775e-06, + "loss": 0.307, + "step": 970 + }, + { + "epoch": 3.7953216374269005, + "grad_norm": 1.7373608350753784, + "learning_rate": 4.448344929460004e-06, + "loss": 0.2796, + "step": 975 + }, + { + "epoch": 3.814814814814815, + "grad_norm": 1.4814600944519043, + "learning_rate": 4.311509736673627e-06, + "loss": 0.2774, + "step": 980 + }, + { + "epoch": 3.8343079922027288, + "grad_norm": 1.3454452753067017, + "learning_rate": 4.17645808420407e-06, + "loss": 0.2528, + "step": 985 + }, + { + "epoch": 3.853801169590643, + "grad_norm": 1.4341703653335571, + "learning_rate": 4.043212507513419e-06, + "loss": 0.2618, + "step": 990 + }, + { + "epoch": 3.8732943469785575, + "grad_norm": 1.3091856241226196, + "learning_rate": 3.911795240692008e-06, + "loss": 0.259, + "step": 995 + }, + { + "epoch": 3.892787524366472, + "grad_norm": 1.3750048875808716, + "learning_rate": 3.7822282127483323e-06, + "loss": 0.2385, + "step": 1000 + }, + { + "epoch": 3.912280701754386, + "grad_norm": 1.217860460281372, + "learning_rate": 3.654533043949824e-06, + "loss": 0.2555, + "step": 1005 + }, + { + "epoch": 3.9317738791423, + "grad_norm": 1.5768376588821411, + "learning_rate": 3.5287310422152124e-06, + "loss": 0.2305, + "step": 1010 + }, + { + "epoch": 3.9512670565302144, + "grad_norm": 1.4825247526168823, + "learning_rate": 3.4048431995589453e-06, + "loss": 0.2575, + "step": 1015 + }, + { + "epoch": 3.9707602339181287, + "grad_norm": 1.3332613706588745, + "learning_rate": 3.282890188588351e-06, + "loss": 0.2698, + "step": 1020 + }, + { + "epoch": 3.9902534113060426, + "grad_norm": 1.3274041414260864, + "learning_rate": 3.162892359054098e-06, + "loss": 0.2688, + "step": 1025 + } + ], + "logging_steps": 5, + "max_steps": 1280, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 20000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.781944817895342e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/training_args.bin b/not_grounded/0_128_e5_3e-5/checkpoint-1028/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4ce5ba5a4e8b2f8e0972e0e67dc1879011fda1a2 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa02dca237b948f004cb8d3e65da8c4d5a5b94ef5684cb38197d8ffa2ad8f171 +size 7800 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1028/zero_to_fp32.py b/not_grounded/0_128_e5_3e-5/checkpoint-1028/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1028/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/README.md b/not_grounded/0_128_e5_3e-5/checkpoint-1280/README.md new file mode 100644 index 0000000000000000000000000000000000000000..0994074add6527289eb93335320fbd12270c4e7f --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.12.0 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/adapter_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-1280/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2afdcf8899170963495ad14c67c715c56d8fd480 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/adapter_config.json @@ -0,0 +1,34 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 256, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 128, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "gate_proj", + "o_proj", + "v_proj", + "q_proj", + "down_proj", + "up_proj", + "k_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/adapter_model.safetensors b/not_grounded/0_128_e5_3e-5/checkpoint-1280/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ead20001ebe65d23fb503e1e7bce6202ddc19be7 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c165af254b73087fae0132a22c447163060cace354a89c59ed05addc2b82080 +size 671150064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/latest b/not_grounded/0_128_e5_3e-5/checkpoint-1280/latest new file mode 100644 index 0000000000000000000000000000000000000000..380bcdedc8b93542ba12b8a0168d1e6cfb7732a4 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/latest @@ -0,0 +1 @@ +global_step1278 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/rng_state_0.pth b/not_grounded/0_128_e5_3e-5/checkpoint-1280/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..2b6d6f190c12aec1394b479d5eaefd636cf3d0a2 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:72f5937dd88cf0c6ff98d18b6044a2a7ec4f202c93ede9d3307d018b1b770ec2 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/rng_state_1.pth b/not_grounded/0_128_e5_3e-5/checkpoint-1280/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..e41f03b1dae1bff031b49b09034814a84513bfa0 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:284678eba718698ee7db95a9de33fbd46a304b6f97d78c5caca147fea474fe70 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/scheduler.pt b/not_grounded/0_128_e5_3e-5/checkpoint-1280/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1052cfde047c132578f29b0e3c2ae2f856aeba4e --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f4e927bdf4126e4c5e082619dfd8a9fc963d4a19aa3004def570dc37a58fa43 +size 1064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/special_tokens_map.json b/not_grounded/0_128_e5_3e-5/checkpoint-1280/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..b43be96621d147110fb8a18b5776ec6e38516127 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer.json b/not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6ac83ff885a7c13c549808db25dd230af9d67648 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384a7e7c676f7be2e5d2e8449c508be9b00e5b18c5b3c39ebc626e96b3f4b988 +size 17210019 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..8ae1bc63bd6e5ca8a863628311061c143679ff93 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/tokenizer_config.json @@ -0,0 +1,2064 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizer" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/trainer_state.json b/not_grounded/0_128_e5_3e-5/checkpoint-1280/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..19c113bb27ee1e19c533d9767296c84190f72b52 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/trainer_state.json @@ -0,0 +1,1826 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 4.982456140350877, + "eval_steps": 500, + "global_step": 1280, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.01949317738791423, + "grad_norm": 0.6081938743591309, + "learning_rate": 1.875e-06, + "loss": 1.7943, + "step": 5 + }, + { + "epoch": 0.03898635477582846, + "grad_norm": 0.5703465938568115, + "learning_rate": 4.21875e-06, + "loss": 1.7931, + "step": 10 + }, + { + "epoch": 0.05847953216374269, + "grad_norm": 0.4802420139312744, + "learning_rate": 6.5625e-06, + "loss": 1.8677, + "step": 15 + }, + { + "epoch": 0.07797270955165692, + "grad_norm": 0.4527171552181244, + "learning_rate": 8.90625e-06, + "loss": 1.7498, + "step": 20 + }, + { + "epoch": 0.09746588693957114, + "grad_norm": 0.41348591446876526, + "learning_rate": 1.125e-05, + "loss": 1.7402, + "step": 25 + }, + { + "epoch": 0.11695906432748537, + "grad_norm": 0.44569578766822815, + "learning_rate": 1.359375e-05, + "loss": 1.6984, + "step": 30 + }, + { + "epoch": 0.1364522417153996, + "grad_norm": 0.439203143119812, + "learning_rate": 1.59375e-05, + "loss": 1.6548, + "step": 35 + }, + { + "epoch": 0.15594541910331383, + "grad_norm": 0.4098866879940033, + "learning_rate": 1.828125e-05, + "loss": 1.7276, + "step": 40 + }, + { + "epoch": 0.17543859649122806, + "grad_norm": 0.42010200023651123, + "learning_rate": 2.0625e-05, + "loss": 1.7037, + "step": 45 + }, + { + "epoch": 0.1949317738791423, + "grad_norm": 0.4497705399990082, + "learning_rate": 2.296875e-05, + "loss": 1.6829, + "step": 50 + }, + { + "epoch": 0.21442495126705652, + "grad_norm": 0.4349888563156128, + "learning_rate": 2.5312500000000002e-05, + "loss": 1.696, + "step": 55 + }, + { + "epoch": 0.23391812865497075, + "grad_norm": 0.45078787207603455, + "learning_rate": 2.765625e-05, + "loss": 1.6246, + "step": 60 + }, + { + "epoch": 0.253411306042885, + "grad_norm": 0.44585877656936646, + "learning_rate": 3e-05, + "loss": 1.6324, + "step": 65 + }, + { + "epoch": 0.2729044834307992, + "grad_norm": 0.4684946537017822, + "learning_rate": 2.9998748508718575e-05, + "loss": 1.5858, + "step": 70 + }, + { + "epoch": 0.29239766081871343, + "grad_norm": 0.5270784497261047, + "learning_rate": 2.9994994243705013e-05, + "loss": 1.5719, + "step": 75 + }, + { + "epoch": 0.31189083820662766, + "grad_norm": 0.5215119123458862, + "learning_rate": 2.9988737831416642e-05, + "loss": 1.5858, + "step": 80 + }, + { + "epoch": 0.3313840155945419, + "grad_norm": 0.5029568672180176, + "learning_rate": 2.9979980315832853e-05, + "loss": 1.4428, + "step": 85 + }, + { + "epoch": 0.3508771929824561, + "grad_norm": 0.5102682709693909, + "learning_rate": 2.9968723158280906e-05, + "loss": 1.5211, + "step": 90 + }, + { + "epoch": 0.37037037037037035, + "grad_norm": 0.5617806911468506, + "learning_rate": 2.995496823719206e-05, + "loss": 1.4377, + "step": 95 + }, + { + "epoch": 0.3898635477582846, + "grad_norm": 0.5236561298370361, + "learning_rate": 2.9938717847788167e-05, + "loss": 1.464, + "step": 100 + }, + { + "epoch": 0.4093567251461988, + "grad_norm": 0.6214772462844849, + "learning_rate": 2.9919974701698638e-05, + "loss": 1.4009, + "step": 105 + }, + { + "epoch": 0.42884990253411304, + "grad_norm": 0.5746711492538452, + "learning_rate": 2.989874192650801e-05, + "loss": 1.3927, + "step": 110 + }, + { + "epoch": 0.44834307992202727, + "grad_norm": 0.6362848877906799, + "learning_rate": 2.9875023065234003e-05, + "loss": 1.4212, + "step": 115 + }, + { + "epoch": 0.4678362573099415, + "grad_norm": 0.6465691328048706, + "learning_rate": 2.984882207573638e-05, + "loss": 1.4462, + "step": 120 + }, + { + "epoch": 0.4873294346978557, + "grad_norm": 0.6339296698570251, + "learning_rate": 2.982014333005645e-05, + "loss": 1.337, + "step": 125 + }, + { + "epoch": 0.50682261208577, + "grad_norm": 0.6140757203102112, + "learning_rate": 2.9788991613687575e-05, + "loss": 1.3289, + "step": 130 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 0.6720002293586731, + "learning_rate": 2.9755372124776616e-05, + "loss": 1.3323, + "step": 135 + }, + { + "epoch": 0.5458089668615984, + "grad_norm": 0.6524502038955688, + "learning_rate": 2.971929047325654e-05, + "loss": 1.3166, + "step": 140 + }, + { + "epoch": 0.5653021442495126, + "grad_norm": 0.7752301692962646, + "learning_rate": 2.968075267991032e-05, + "loss": 1.3543, + "step": 145 + }, + { + "epoch": 0.5847953216374269, + "grad_norm": 0.6975963711738586, + "learning_rate": 2.963976517536627e-05, + "loss": 1.2675, + "step": 150 + }, + { + "epoch": 0.6042884990253411, + "grad_norm": 0.7260017991065979, + "learning_rate": 2.9596334799025003e-05, + "loss": 1.3213, + "step": 155 + }, + { + "epoch": 0.6237816764132553, + "grad_norm": 0.8624387979507446, + "learning_rate": 2.9550468797918162e-05, + "loss": 1.2492, + "step": 160 + }, + { + "epoch": 0.6432748538011696, + "grad_norm": 0.7931470274925232, + "learning_rate": 2.950217482549915e-05, + "loss": 1.2353, + "step": 165 + }, + { + "epoch": 0.6627680311890838, + "grad_norm": 0.7223592400550842, + "learning_rate": 2.9451460940366025e-05, + "loss": 1.1868, + "step": 170 + }, + { + "epoch": 0.682261208576998, + "grad_norm": 0.7910372614860535, + "learning_rate": 2.9398335604916797e-05, + "loss": 1.2148, + "step": 175 + }, + { + "epoch": 0.7017543859649122, + "grad_norm": 0.851186990737915, + "learning_rate": 2.9342807683937352e-05, + "loss": 1.2595, + "step": 180 + }, + { + "epoch": 0.7212475633528265, + "grad_norm": 0.8052693009376526, + "learning_rate": 2.928488644312222e-05, + "loss": 1.2406, + "step": 185 + }, + { + "epoch": 0.7407407407407407, + "grad_norm": 0.8178960680961609, + "learning_rate": 2.9224581547528453e-05, + "loss": 1.2197, + "step": 190 + }, + { + "epoch": 0.7602339181286549, + "grad_norm": 0.8441738486289978, + "learning_rate": 2.916190305996286e-05, + "loss": 1.1726, + "step": 195 + }, + { + "epoch": 0.7797270955165692, + "grad_norm": 0.9862117767333984, + "learning_rate": 2.909686143930287e-05, + "loss": 1.1745, + "step": 200 + }, + { + "epoch": 0.7992202729044834, + "grad_norm": 0.8886749148368835, + "learning_rate": 2.902946753875131e-05, + "loss": 1.1447, + "step": 205 + }, + { + "epoch": 0.8187134502923976, + "grad_norm": 0.8914117813110352, + "learning_rate": 2.895973260402537e-05, + "loss": 1.1248, + "step": 210 + }, + { + "epoch": 0.8382066276803118, + "grad_norm": 1.1335160732269287, + "learning_rate": 2.8887668271480098e-05, + "loss": 1.1291, + "step": 215 + }, + { + "epoch": 0.8576998050682261, + "grad_norm": 1.0726529359817505, + "learning_rate": 2.8813286566166674e-05, + "loss": 1.1646, + "step": 220 + }, + { + "epoch": 0.8771929824561403, + "grad_norm": 1.0396595001220703, + "learning_rate": 2.873659989982586e-05, + "loss": 1.0894, + "step": 225 + }, + { + "epoch": 0.8966861598440545, + "grad_norm": 0.9956820011138916, + "learning_rate": 2.8657621068816903e-05, + "loss": 1.0984, + "step": 230 + }, + { + "epoch": 0.9161793372319688, + "grad_norm": 0.9483380317687988, + "learning_rate": 2.857636325198225e-05, + "loss": 1.0216, + "step": 235 + }, + { + "epoch": 0.935672514619883, + "grad_norm": 0.8558198809623718, + "learning_rate": 2.849284000844849e-05, + "loss": 1.105, + "step": 240 + }, + { + "epoch": 0.9551656920077972, + "grad_norm": 1.005858063697815, + "learning_rate": 2.8407065275363756e-05, + "loss": 1.0714, + "step": 245 + }, + { + "epoch": 0.9746588693957114, + "grad_norm": 0.9553005695343018, + "learning_rate": 2.8319053365572135e-05, + "loss": 1.0339, + "step": 250 + }, + { + "epoch": 0.9941520467836257, + "grad_norm": 0.9755716919898987, + "learning_rate": 2.8228818965225325e-05, + "loss": 1.0532, + "step": 255 + }, + { + "epoch": 1.0116959064327484, + "grad_norm": 1.0532433986663818, + "learning_rate": 2.8136377131332043e-05, + "loss": 1.0005, + "step": 260 + }, + { + "epoch": 1.0311890838206628, + "grad_norm": 0.9859170913696289, + "learning_rate": 2.8041743289245503e-05, + "loss": 0.9084, + "step": 265 + }, + { + "epoch": 1.050682261208577, + "grad_norm": 0.9481168389320374, + "learning_rate": 2.7944933230089484e-05, + "loss": 0.9431, + "step": 270 + }, + { + "epoch": 1.0701754385964912, + "grad_norm": 0.9678940773010254, + "learning_rate": 2.784596310812331e-05, + "loss": 0.9262, + "step": 275 + }, + { + "epoch": 1.0896686159844053, + "grad_norm": 0.9351955652236938, + "learning_rate": 2.774484943804629e-05, + "loss": 0.9044, + "step": 280 + }, + { + "epoch": 1.1091617933723197, + "grad_norm": 1.0261086225509644, + "learning_rate": 2.764160909224196e-05, + "loss": 0.9579, + "step": 285 + }, + { + "epoch": 1.128654970760234, + "grad_norm": 0.9796629548072815, + "learning_rate": 2.7536259297962674e-05, + "loss": 0.9283, + "step": 290 + }, + { + "epoch": 1.1481481481481481, + "grad_norm": 0.9690019488334656, + "learning_rate": 2.7428817634454973e-05, + "loss": 0.9017, + "step": 295 + }, + { + "epoch": 1.1676413255360623, + "grad_norm": 1.1676615476608276, + "learning_rate": 2.7319302030026207e-05, + "loss": 0.8919, + "step": 300 + }, + { + "epoch": 1.1871345029239766, + "grad_norm": 1.1467899084091187, + "learning_rate": 2.7207730759052925e-05, + "loss": 0.871, + "step": 305 + }, + { + "epoch": 1.206627680311891, + "grad_norm": 1.0315958261489868, + "learning_rate": 2.7094122438931513e-05, + "loss": 0.8388, + "step": 310 + }, + { + "epoch": 1.226120857699805, + "grad_norm": 1.0231361389160156, + "learning_rate": 2.697849602697159e-05, + "loss": 0.8841, + "step": 315 + }, + { + "epoch": 1.2456140350877192, + "grad_norm": 0.995548665523529, + "learning_rate": 2.6860870817232682e-05, + "loss": 0.8879, + "step": 320 + }, + { + "epoch": 1.2651072124756335, + "grad_norm": 1.102622151374817, + "learning_rate": 2.6741266437304718e-05, + "loss": 0.857, + "step": 325 + }, + { + "epoch": 1.2846003898635479, + "grad_norm": 1.3752007484436035, + "learning_rate": 2.661970284503286e-05, + "loss": 0.8027, + "step": 330 + }, + { + "epoch": 1.304093567251462, + "grad_norm": 1.129936933517456, + "learning_rate": 2.6496200325187223e-05, + "loss": 0.855, + "step": 335 + }, + { + "epoch": 1.323586744639376, + "grad_norm": 1.2070459127426147, + "learning_rate": 2.6370779486078047e-05, + "loss": 0.8544, + "step": 340 + }, + { + "epoch": 1.3430799220272904, + "grad_norm": 1.1400156021118164, + "learning_rate": 2.6243461256116892e-05, + "loss": 0.9026, + "step": 345 + }, + { + "epoch": 1.3625730994152048, + "grad_norm": 0.9957541823387146, + "learning_rate": 2.611426688032439e-05, + "loss": 0.8115, + "step": 350 + }, + { + "epoch": 1.3820662768031189, + "grad_norm": 1.0340908765792847, + "learning_rate": 2.598321791678519e-05, + "loss": 0.8314, + "step": 355 + }, + { + "epoch": 1.401559454191033, + "grad_norm": 1.1327755451202393, + "learning_rate": 2.5850336233050677e-05, + "loss": 0.8319, + "step": 360 + }, + { + "epoch": 1.4210526315789473, + "grad_norm": 1.0419026613235474, + "learning_rate": 2.5715644002489998e-05, + "loss": 0.8115, + "step": 365 + }, + { + "epoch": 1.4405458089668617, + "grad_norm": 1.080306053161621, + "learning_rate": 2.557916370059012e-05, + "loss": 0.7734, + "step": 370 + }, + { + "epoch": 1.4600389863547758, + "grad_norm": 1.1247351169586182, + "learning_rate": 2.544091810120543e-05, + "loss": 0.7968, + "step": 375 + }, + { + "epoch": 1.47953216374269, + "grad_norm": 1.1105760335922241, + "learning_rate": 2.530093027275757e-05, + "loss": 0.7994, + "step": 380 + }, + { + "epoch": 1.4990253411306043, + "grad_norm": 1.3396145105361938, + "learning_rate": 2.5159223574386117e-05, + "loss": 0.7891, + "step": 385 + }, + { + "epoch": 1.5185185185185186, + "grad_norm": 1.2513933181762695, + "learning_rate": 2.501582165205074e-05, + "loss": 0.7825, + "step": 390 + }, + { + "epoch": 1.5380116959064327, + "grad_norm": 1.139088749885559, + "learning_rate": 2.4870748434585514e-05, + "loss": 0.7875, + "step": 395 + }, + { + "epoch": 1.5575048732943468, + "grad_norm": 1.2570606470108032, + "learning_rate": 2.4724028129706012e-05, + "loss": 0.7524, + "step": 400 + }, + { + "epoch": 1.5769980506822612, + "grad_norm": 1.115220069885254, + "learning_rate": 2.4575685219969884e-05, + "loss": 0.7855, + "step": 405 + }, + { + "epoch": 1.5964912280701755, + "grad_norm": 1.2705992460250854, + "learning_rate": 2.442574445869156e-05, + "loss": 0.7338, + "step": 410 + }, + { + "epoch": 1.6159844054580896, + "grad_norm": 1.1325453519821167, + "learning_rate": 2.4274230865811763e-05, + "loss": 0.7528, + "step": 415 + }, + { + "epoch": 1.6354775828460038, + "grad_norm": 1.301644206047058, + "learning_rate": 2.4121169723722566e-05, + "loss": 0.7445, + "step": 420 + }, + { + "epoch": 1.654970760233918, + "grad_norm": 1.4577105045318604, + "learning_rate": 2.396658657304861e-05, + "loss": 0.7335, + "step": 425 + }, + { + "epoch": 1.6744639376218324, + "grad_norm": 1.1624274253845215, + "learning_rate": 2.381050720838528e-05, + "loss": 0.7146, + "step": 430 + }, + { + "epoch": 1.6939571150097466, + "grad_norm": 1.4543778896331787, + "learning_rate": 2.3652957673994448e-05, + "loss": 0.7338, + "step": 435 + }, + { + "epoch": 1.7134502923976607, + "grad_norm": 1.208083987236023, + "learning_rate": 2.3493964259458603e-05, + "loss": 0.7187, + "step": 440 + }, + { + "epoch": 1.732943469785575, + "grad_norm": 1.1645222902297974, + "learning_rate": 2.3333553495294033e-05, + "loss": 0.7325, + "step": 445 + }, + { + "epoch": 1.7524366471734893, + "grad_norm": 1.2736948728561401, + "learning_rate": 2.317175214852377e-05, + "loss": 0.7152, + "step": 450 + }, + { + "epoch": 1.7719298245614035, + "grad_norm": 1.2995411157608032, + "learning_rate": 2.3008587218211127e-05, + "loss": 0.6953, + "step": 455 + }, + { + "epoch": 1.7914230019493176, + "grad_norm": 1.4111621379852295, + "learning_rate": 2.284408593095446e-05, + "loss": 0.7023, + "step": 460 + }, + { + "epoch": 1.810916179337232, + "grad_norm": 1.2578918933868408, + "learning_rate": 2.2678275736344014e-05, + "loss": 0.6771, + "step": 465 + }, + { + "epoch": 1.8304093567251463, + "grad_norm": 1.3310747146606445, + "learning_rate": 2.251118430238151e-05, + "loss": 0.6792, + "step": 470 + }, + { + "epoch": 1.8499025341130604, + "grad_norm": 1.1377298831939697, + "learning_rate": 2.2342839510863323e-05, + "loss": 0.7322, + "step": 475 + }, + { + "epoch": 1.8693957115009745, + "grad_norm": 1.2897000312805176, + "learning_rate": 2.2173269452727965e-05, + "loss": 0.6468, + "step": 480 + }, + { + "epoch": 1.8888888888888888, + "grad_norm": 1.1680046319961548, + "learning_rate": 2.200250242336868e-05, + "loss": 0.6785, + "step": 485 + }, + { + "epoch": 1.9083820662768032, + "grad_norm": 1.2389404773712158, + "learning_rate": 2.183056691791193e-05, + "loss": 0.6521, + "step": 490 + }, + { + "epoch": 1.9278752436647173, + "grad_norm": 1.1207393407821655, + "learning_rate": 2.1657491626462514e-05, + "loss": 0.6607, + "step": 495 + }, + { + "epoch": 1.9473684210526314, + "grad_norm": 1.1037189960479736, + "learning_rate": 2.1483305429316208e-05, + "loss": 0.6033, + "step": 500 + }, + { + "epoch": 1.9668615984405458, + "grad_norm": 1.1787821054458618, + "learning_rate": 2.1308037392140613e-05, + "loss": 0.6254, + "step": 505 + }, + { + "epoch": 1.98635477582846, + "grad_norm": 1.266696572303772, + "learning_rate": 2.113171676112513e-05, + "loss": 0.5962, + "step": 510 + }, + { + "epoch": 2.003898635477583, + "grad_norm": 1.4790624380111694, + "learning_rate": 2.0954372958100746e-05, + "loss": 0.6104, + "step": 515 + }, + { + "epoch": 2.023391812865497, + "grad_norm": 1.6624841690063477, + "learning_rate": 2.0776035575630575e-05, + "loss": 0.5632, + "step": 520 + }, + { + "epoch": 2.042884990253411, + "grad_norm": 1.243015170097351, + "learning_rate": 2.0596734372071855e-05, + "loss": 0.5019, + "step": 525 + }, + { + "epoch": 2.0623781676413255, + "grad_norm": 1.4598456621170044, + "learning_rate": 2.0416499266610316e-05, + "loss": 0.5299, + "step": 530 + }, + { + "epoch": 2.08187134502924, + "grad_norm": 1.2799503803253174, + "learning_rate": 2.023536033426771e-05, + "loss": 0.5432, + "step": 535 + }, + { + "epoch": 2.101364522417154, + "grad_norm": 1.3306210041046143, + "learning_rate": 2.00533478008833e-05, + "loss": 0.5639, + "step": 540 + }, + { + "epoch": 2.120857699805068, + "grad_norm": 1.5138165950775146, + "learning_rate": 1.9870492038070255e-05, + "loss": 0.519, + "step": 545 + }, + { + "epoch": 2.1403508771929824, + "grad_norm": 1.283997893333435, + "learning_rate": 1.9686823558147616e-05, + "loss": 0.5276, + "step": 550 + }, + { + "epoch": 2.159844054580897, + "grad_norm": 1.2089860439300537, + "learning_rate": 1.9502373009048898e-05, + "loss": 0.5384, + "step": 555 + }, + { + "epoch": 2.1793372319688107, + "grad_norm": 1.327548861503601, + "learning_rate": 1.9317171169207973e-05, + "loss": 0.5297, + "step": 560 + }, + { + "epoch": 2.198830409356725, + "grad_norm": 1.3738446235656738, + "learning_rate": 1.913124894242322e-05, + "loss": 0.5324, + "step": 565 + }, + { + "epoch": 2.2183235867446394, + "grad_norm": 1.2523514032363892, + "learning_rate": 1.8944637352700754e-05, + "loss": 0.4629, + "step": 570 + }, + { + "epoch": 2.2378167641325537, + "grad_norm": 1.4739748239517212, + "learning_rate": 1.875736753907758e-05, + "loss": 0.487, + "step": 575 + }, + { + "epoch": 2.257309941520468, + "grad_norm": 1.3324633836746216, + "learning_rate": 1.856947075042557e-05, + "loss": 0.4968, + "step": 580 + }, + { + "epoch": 2.276803118908382, + "grad_norm": 1.3165260553359985, + "learning_rate": 1.8380978340237094e-05, + "loss": 0.5223, + "step": 585 + }, + { + "epoch": 2.2962962962962963, + "grad_norm": 1.3404366970062256, + "learning_rate": 1.8191921761393222e-05, + "loss": 0.4838, + "step": 590 + }, + { + "epoch": 2.3157894736842106, + "grad_norm": 1.3298143148422241, + "learning_rate": 1.8002332560915297e-05, + "loss": 0.4742, + "step": 595 + }, + { + "epoch": 2.3352826510721245, + "grad_norm": 1.2816650867462158, + "learning_rate": 1.7812242374700847e-05, + "loss": 0.4775, + "step": 600 + }, + { + "epoch": 2.354775828460039, + "grad_norm": 1.406812310218811, + "learning_rate": 1.7621682922244637e-05, + "loss": 0.5127, + "step": 605 + }, + { + "epoch": 2.374269005847953, + "grad_norm": 1.3161938190460205, + "learning_rate": 1.7430686001345777e-05, + "loss": 0.4723, + "step": 610 + }, + { + "epoch": 2.3937621832358675, + "grad_norm": 1.4261293411254883, + "learning_rate": 1.7239283482801783e-05, + "loss": 0.5121, + "step": 615 + }, + { + "epoch": 2.413255360623782, + "grad_norm": 1.2458994388580322, + "learning_rate": 1.7047507305090397e-05, + "loss": 0.4821, + "step": 620 + }, + { + "epoch": 2.4327485380116958, + "grad_norm": 1.3558402061462402, + "learning_rate": 1.6855389469040217e-05, + "loss": 0.4423, + "step": 625 + }, + { + "epoch": 2.45224171539961, + "grad_norm": 1.2846639156341553, + "learning_rate": 1.6662962032490814e-05, + "loss": 0.4964, + "step": 630 + }, + { + "epoch": 2.4717348927875245, + "grad_norm": 1.3657177686691284, + "learning_rate": 1.6470257104943414e-05, + "loss": 0.4906, + "step": 635 + }, + { + "epoch": 2.4912280701754383, + "grad_norm": 1.3313744068145752, + "learning_rate": 1.6277306842202902e-05, + "loss": 0.4824, + "step": 640 + }, + { + "epoch": 2.5107212475633527, + "grad_norm": 1.427259922027588, + "learning_rate": 1.6084143441012158e-05, + "loss": 0.5067, + "step": 645 + }, + { + "epoch": 2.530214424951267, + "grad_norm": 1.3357858657836914, + "learning_rate": 1.5890799133679523e-05, + "loss": 0.5128, + "step": 650 + }, + { + "epoch": 2.5497076023391814, + "grad_norm": 1.6301915645599365, + "learning_rate": 1.569730618270031e-05, + "loss": 0.4331, + "step": 655 + }, + { + "epoch": 2.5692007797270957, + "grad_norm": 1.4900562763214111, + "learning_rate": 1.5503696875373353e-05, + "loss": 0.4391, + "step": 660 + }, + { + "epoch": 2.5886939571150096, + "grad_norm": 1.402565836906433, + "learning_rate": 1.5310003518413318e-05, + "loss": 0.4457, + "step": 665 + }, + { + "epoch": 2.608187134502924, + "grad_norm": 1.3605144023895264, + "learning_rate": 1.5116258432559883e-05, + "loss": 0.4546, + "step": 670 + }, + { + "epoch": 2.6276803118908383, + "grad_norm": 1.3472214937210083, + "learning_rate": 1.4922493947184482e-05, + "loss": 0.4791, + "step": 675 + }, + { + "epoch": 2.647173489278752, + "grad_norm": 1.4113649129867554, + "learning_rate": 1.4728742394895666e-05, + "loss": 0.4364, + "step": 680 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 1.3126051425933838, + "learning_rate": 1.4535036106143895e-05, + "loss": 0.4309, + "step": 685 + }, + { + "epoch": 2.686159844054581, + "grad_norm": 1.4124526977539062, + "learning_rate": 1.4341407403826695e-05, + "loss": 0.4346, + "step": 690 + }, + { + "epoch": 2.705653021442495, + "grad_norm": 1.4493342638015747, + "learning_rate": 1.4147888597895117e-05, + "loss": 0.4175, + "step": 695 + }, + { + "epoch": 2.7251461988304095, + "grad_norm": 1.4753097295761108, + "learning_rate": 1.3954511979962271e-05, + "loss": 0.4275, + "step": 700 + }, + { + "epoch": 2.7446393762183234, + "grad_norm": 1.2626641988754272, + "learning_rate": 1.3761309817915017e-05, + "loss": 0.4182, + "step": 705 + }, + { + "epoch": 2.7641325536062378, + "grad_norm": 1.526242971420288, + "learning_rate": 1.3568314350529534e-05, + "loss": 0.3954, + "step": 710 + }, + { + "epoch": 2.783625730994152, + "grad_norm": 1.5078595876693726, + "learning_rate": 1.3375557782091789e-05, + "loss": 0.4406, + "step": 715 + }, + { + "epoch": 2.803118908382066, + "grad_norm": 1.398832082748413, + "learning_rate": 1.3183072277023766e-05, + "loss": 0.4096, + "step": 720 + }, + { + "epoch": 2.8226120857699804, + "grad_norm": 1.4593217372894287, + "learning_rate": 1.2990889954516313e-05, + "loss": 0.4015, + "step": 725 + }, + { + "epoch": 2.8421052631578947, + "grad_norm": 1.5111234188079834, + "learning_rate": 1.2799042883169576e-05, + "loss": 0.4074, + "step": 730 + }, + { + "epoch": 2.861598440545809, + "grad_norm": 1.4062628746032715, + "learning_rate": 1.2607563075641838e-05, + "loss": 0.4055, + "step": 735 + }, + { + "epoch": 2.8810916179337234, + "grad_norm": 1.3948218822479248, + "learning_rate": 1.2416482483307728e-05, + "loss": 0.4129, + "step": 740 + }, + { + "epoch": 2.9005847953216373, + "grad_norm": 1.408527135848999, + "learning_rate": 1.2225832990926625e-05, + "loss": 0.4073, + "step": 745 + }, + { + "epoch": 2.9200779727095516, + "grad_norm": 1.4006417989730835, + "learning_rate": 1.20356464113222e-05, + "loss": 0.4147, + "step": 750 + }, + { + "epoch": 2.939571150097466, + "grad_norm": 1.3896782398223877, + "learning_rate": 1.184595448007395e-05, + "loss": 0.3975, + "step": 755 + }, + { + "epoch": 2.95906432748538, + "grad_norm": 1.4923348426818848, + "learning_rate": 1.1656788850221617e-05, + "loss": 0.3933, + "step": 760 + }, + { + "epoch": 2.978557504873294, + "grad_norm": 1.3745965957641602, + "learning_rate": 1.1468181086983414e-05, + "loss": 0.395, + "step": 765 + }, + { + "epoch": 2.9980506822612085, + "grad_norm": 1.429971694946289, + "learning_rate": 1.1280162662488834e-05, + "loss": 0.4257, + "step": 770 + }, + { + "epoch": 3.0155945419103314, + "grad_norm": 1.4495044946670532, + "learning_rate": 1.1092764950527088e-05, + "loss": 0.3595, + "step": 775 + }, + { + "epoch": 3.0350877192982457, + "grad_norm": 1.3142398595809937, + "learning_rate": 1.090601922131187e-05, + "loss": 0.278, + "step": 780 + }, + { + "epoch": 3.0545808966861596, + "grad_norm": 1.3927865028381348, + "learning_rate": 1.0719956636263424e-05, + "loss": 0.2914, + "step": 785 + }, + { + "epoch": 3.074074074074074, + "grad_norm": 1.6307252645492554, + "learning_rate": 1.0534608242808827e-05, + "loss": 0.3096, + "step": 790 + }, + { + "epoch": 3.0935672514619883, + "grad_norm": 1.4899730682373047, + "learning_rate": 1.03500049692012e-05, + "loss": 0.3015, + "step": 795 + }, + { + "epoch": 3.1130604288499026, + "grad_norm": 1.6715621948242188, + "learning_rate": 1.0166177619358873e-05, + "loss": 0.316, + "step": 800 + }, + { + "epoch": 3.132553606237817, + "grad_norm": 1.5180559158325195, + "learning_rate": 9.983156867725257e-06, + "loss": 0.3194, + "step": 805 + }, + { + "epoch": 3.152046783625731, + "grad_norm": 1.4555518627166748, + "learning_rate": 9.800973254150351e-06, + "loss": 0.3059, + "step": 810 + }, + { + "epoch": 3.171539961013645, + "grad_norm": 1.3086048364639282, + "learning_rate": 9.619657178794692e-06, + "loss": 0.3327, + "step": 815 + }, + { + "epoch": 3.1910331384015596, + "grad_norm": 1.4091737270355225, + "learning_rate": 9.439238897056612e-06, + "loss": 0.3313, + "step": 820 + }, + { + "epoch": 3.2105263157894735, + "grad_norm": 1.4301328659057617, + "learning_rate": 9.259748514523654e-06, + "loss": 0.3003, + "step": 825 + }, + { + "epoch": 3.230019493177388, + "grad_norm": 1.4087915420532227, + "learning_rate": 9.081215981949002e-06, + "loss": 0.3249, + "step": 830 + }, + { + "epoch": 3.249512670565302, + "grad_norm": 1.4424235820770264, + "learning_rate": 8.903671090253717e-06, + "loss": 0.3117, + "step": 835 + }, + { + "epoch": 3.2690058479532165, + "grad_norm": 1.3572005033493042, + "learning_rate": 8.727143465555668e-06, + "loss": 0.3062, + "step": 840 + }, + { + "epoch": 3.288499025341131, + "grad_norm": 1.4524186849594116, + "learning_rate": 8.551662564225952e-06, + "loss": 0.3053, + "step": 845 + }, + { + "epoch": 3.3079922027290447, + "grad_norm": 1.3756184577941895, + "learning_rate": 8.377257667973644e-06, + "loss": 0.333, + "step": 850 + }, + { + "epoch": 3.327485380116959, + "grad_norm": 1.5087087154388428, + "learning_rate": 8.203957878959693e-06, + "loss": 0.2984, + "step": 855 + }, + { + "epoch": 3.3469785575048734, + "grad_norm": 1.7474857568740845, + "learning_rate": 8.031792114940777e-06, + "loss": 0.3344, + "step": 860 + }, + { + "epoch": 3.3664717348927873, + "grad_norm": 1.4694979190826416, + "learning_rate": 7.860789104443897e-06, + "loss": 0.2761, + "step": 865 + }, + { + "epoch": 3.3859649122807016, + "grad_norm": 1.578460931777954, + "learning_rate": 7.690977381972626e-06, + "loss": 0.2994, + "step": 870 + }, + { + "epoch": 3.405458089668616, + "grad_norm": 1.3219410181045532, + "learning_rate": 7.5223852832456585e-06, + "loss": 0.273, + "step": 875 + }, + { + "epoch": 3.4249512670565303, + "grad_norm": 1.3564660549163818, + "learning_rate": 7.3550409404685386e-06, + "loss": 0.2994, + "step": 880 + }, + { + "epoch": 3.4444444444444446, + "grad_norm": 1.562962293624878, + "learning_rate": 7.188972277639406e-06, + "loss": 0.3037, + "step": 885 + }, + { + "epoch": 3.4639376218323585, + "grad_norm": 1.477868914604187, + "learning_rate": 7.024207005889406e-06, + "loss": 0.2856, + "step": 890 + }, + { + "epoch": 3.483430799220273, + "grad_norm": 1.406184434890747, + "learning_rate": 6.8607726188586864e-06, + "loss": 0.2949, + "step": 895 + }, + { + "epoch": 3.502923976608187, + "grad_norm": 1.4179635047912598, + "learning_rate": 6.698696388108642e-06, + "loss": 0.2876, + "step": 900 + }, + { + "epoch": 3.522417153996101, + "grad_norm": 1.3572025299072266, + "learning_rate": 6.538005358571235e-06, + "loss": 0.2922, + "step": 905 + }, + { + "epoch": 3.5419103313840155, + "grad_norm": 1.3908215761184692, + "learning_rate": 6.37872634403613e-06, + "loss": 0.2695, + "step": 910 + }, + { + "epoch": 3.56140350877193, + "grad_norm": 1.5078115463256836, + "learning_rate": 6.220885922676385e-06, + "loss": 0.3221, + "step": 915 + }, + { + "epoch": 3.580896686159844, + "grad_norm": 1.4563665390014648, + "learning_rate": 6.0645104326135e-06, + "loss": 0.2814, + "step": 920 + }, + { + "epoch": 3.6003898635477585, + "grad_norm": 1.452856421470642, + "learning_rate": 5.909625967522465e-06, + "loss": 0.288, + "step": 925 + }, + { + "epoch": 3.6198830409356724, + "grad_norm": 1.5082298517227173, + "learning_rate": 5.7562583722776435e-06, + "loss": 0.2762, + "step": 930 + }, + { + "epoch": 3.6393762183235867, + "grad_norm": 1.5691982507705688, + "learning_rate": 5.604433238640134e-06, + "loss": 0.2858, + "step": 935 + }, + { + "epoch": 3.658869395711501, + "grad_norm": 1.399603247642517, + "learning_rate": 5.454175900987411e-06, + "loss": 0.2698, + "step": 940 + }, + { + "epoch": 3.678362573099415, + "grad_norm": 1.2988371849060059, + "learning_rate": 5.305511432085885e-06, + "loss": 0.2611, + "step": 945 + }, + { + "epoch": 3.6978557504873293, + "grad_norm": 1.612922191619873, + "learning_rate": 5.158464638907113e-06, + "loss": 0.2704, + "step": 950 + }, + { + "epoch": 3.7173489278752436, + "grad_norm": 1.4260296821594238, + "learning_rate": 5.01306005848838e-06, + "loss": 0.2624, + "step": 955 + }, + { + "epoch": 3.736842105263158, + "grad_norm": 1.299452304840088, + "learning_rate": 4.869321953838305e-06, + "loss": 0.2503, + "step": 960 + }, + { + "epoch": 3.7563352826510723, + "grad_norm": 1.5614985227584839, + "learning_rate": 4.7272743098881915e-06, + "loss": 0.2642, + "step": 965 + }, + { + "epoch": 3.775828460038986, + "grad_norm": 1.5323383808135986, + "learning_rate": 4.586940829489775e-06, + "loss": 0.307, + "step": 970 + }, + { + "epoch": 3.7953216374269005, + "grad_norm": 1.7373608350753784, + "learning_rate": 4.448344929460004e-06, + "loss": 0.2796, + "step": 975 + }, + { + "epoch": 3.814814814814815, + "grad_norm": 1.4814600944519043, + "learning_rate": 4.311509736673627e-06, + "loss": 0.2774, + "step": 980 + }, + { + "epoch": 3.8343079922027288, + "grad_norm": 1.3454452753067017, + "learning_rate": 4.17645808420407e-06, + "loss": 0.2528, + "step": 985 + }, + { + "epoch": 3.853801169590643, + "grad_norm": 1.4341703653335571, + "learning_rate": 4.043212507513419e-06, + "loss": 0.2618, + "step": 990 + }, + { + "epoch": 3.8732943469785575, + "grad_norm": 1.3091856241226196, + "learning_rate": 3.911795240692008e-06, + "loss": 0.259, + "step": 995 + }, + { + "epoch": 3.892787524366472, + "grad_norm": 1.3750048875808716, + "learning_rate": 3.7822282127483323e-06, + "loss": 0.2385, + "step": 1000 + }, + { + "epoch": 3.912280701754386, + "grad_norm": 1.217860460281372, + "learning_rate": 3.654533043949824e-06, + "loss": 0.2555, + "step": 1005 + }, + { + "epoch": 3.9317738791423, + "grad_norm": 1.5768376588821411, + "learning_rate": 3.5287310422152124e-06, + "loss": 0.2305, + "step": 1010 + }, + { + "epoch": 3.9512670565302144, + "grad_norm": 1.4825247526168823, + "learning_rate": 3.4048431995589453e-06, + "loss": 0.2575, + "step": 1015 + }, + { + "epoch": 3.9707602339181287, + "grad_norm": 1.3332613706588745, + "learning_rate": 3.282890188588351e-06, + "loss": 0.2698, + "step": 1020 + }, + { + "epoch": 3.9902534113060426, + "grad_norm": 1.3274041414260864, + "learning_rate": 3.162892359054098e-06, + "loss": 0.2688, + "step": 1025 + }, + { + "epoch": 4.007797270955166, + "grad_norm": 1.3027465343475342, + "learning_rate": 3.044869734454502e-06, + "loss": 0.2368, + "step": 1030 + }, + { + "epoch": 4.02729044834308, + "grad_norm": 1.4599330425262451, + "learning_rate": 2.928842008694331e-06, + "loss": 0.237, + "step": 1035 + }, + { + "epoch": 4.046783625730994, + "grad_norm": 1.3548412322998047, + "learning_rate": 2.8148285427985383e-06, + "loss": 0.1993, + "step": 1040 + }, + { + "epoch": 4.0662768031189085, + "grad_norm": 1.404860496520996, + "learning_rate": 2.702848361681605e-06, + "loss": 0.2193, + "step": 1045 + }, + { + "epoch": 4.085769980506822, + "grad_norm": 1.2915431261062622, + "learning_rate": 2.5929201509729006e-06, + "loss": 0.2059, + "step": 1050 + }, + { + "epoch": 4.105263157894737, + "grad_norm": 1.3350541591644287, + "learning_rate": 2.4850622538987357e-06, + "loss": 0.2089, + "step": 1055 + }, + { + "epoch": 4.124756335282651, + "grad_norm": 1.285410761833191, + "learning_rate": 2.3792926682214855e-06, + "loss": 0.2033, + "step": 1060 + }, + { + "epoch": 4.144249512670565, + "grad_norm": 1.2443702220916748, + "learning_rate": 2.2756290432363958e-06, + "loss": 0.2204, + "step": 1065 + }, + { + "epoch": 4.16374269005848, + "grad_norm": 1.2086948156356812, + "learning_rate": 2.1740886768265124e-06, + "loss": 0.2146, + "step": 1070 + }, + { + "epoch": 4.183235867446394, + "grad_norm": 1.2621572017669678, + "learning_rate": 2.074688512576267e-06, + "loss": 0.2012, + "step": 1075 + }, + { + "epoch": 4.202729044834308, + "grad_norm": 1.3186023235321045, + "learning_rate": 1.977445136944192e-06, + "loss": 0.2057, + "step": 1080 + }, + { + "epoch": 4.222222222222222, + "grad_norm": 1.4103914499282837, + "learning_rate": 1.882374776495187e-06, + "loss": 0.2273, + "step": 1085 + }, + { + "epoch": 4.241715399610136, + "grad_norm": 1.3912384510040283, + "learning_rate": 1.7894932951928833e-06, + "loss": 0.216, + "step": 1090 + }, + { + "epoch": 4.261208576998051, + "grad_norm": 1.5405287742614746, + "learning_rate": 1.6988161917524902e-06, + "loss": 0.2315, + "step": 1095 + }, + { + "epoch": 4.280701754385965, + "grad_norm": 1.2610150575637817, + "learning_rate": 1.610358597054588e-06, + "loss": 0.2166, + "step": 1100 + }, + { + "epoch": 4.300194931773879, + "grad_norm": 1.282578706741333, + "learning_rate": 1.5241352716203173e-06, + "loss": 0.2059, + "step": 1105 + }, + { + "epoch": 4.319688109161794, + "grad_norm": 1.2637677192687988, + "learning_rate": 1.44016060314835e-06, + "loss": 0.2055, + "step": 1110 + }, + { + "epoch": 4.3391812865497075, + "grad_norm": 1.2789418697357178, + "learning_rate": 1.3584486041140815e-06, + "loss": 0.1914, + "step": 1115 + }, + { + "epoch": 4.358674463937621, + "grad_norm": 1.160399079322815, + "learning_rate": 1.2790129094314247e-06, + "loss": 0.2172, + "step": 1120 + }, + { + "epoch": 4.378167641325536, + "grad_norm": 1.2635594606399536, + "learning_rate": 1.2018667741776269e-06, + "loss": 0.2184, + "step": 1125 + }, + { + "epoch": 4.39766081871345, + "grad_norm": 1.3619000911712646, + "learning_rate": 1.127023071381441e-06, + "loss": 0.2091, + "step": 1130 + }, + { + "epoch": 4.417153996101365, + "grad_norm": 1.2586579322814941, + "learning_rate": 1.0544942898750758e-06, + "loss": 0.2024, + "step": 1135 + }, + { + "epoch": 4.436647173489279, + "grad_norm": 1.450278878211975, + "learning_rate": 9.84292532210217e-07, + "loss": 0.2119, + "step": 1140 + }, + { + "epoch": 4.456140350877193, + "grad_norm": 1.2155523300170898, + "learning_rate": 9.164295126385564e-07, + "loss": 0.1923, + "step": 1145 + }, + { + "epoch": 4.475633528265107, + "grad_norm": 1.3330084085464478, + "learning_rate": 8.509165551570724e-07, + "loss": 0.1957, + "step": 1150 + }, + { + "epoch": 4.495126705653021, + "grad_norm": 1.2223491668701172, + "learning_rate": 7.877645916184434e-07, + "loss": 0.2036, + "step": 1155 + }, + { + "epoch": 4.514619883040936, + "grad_norm": 1.337951421737671, + "learning_rate": 7.269841599069094e-07, + "loss": 0.2036, + "step": 1160 + }, + { + "epoch": 4.53411306042885, + "grad_norm": 1.3998116254806519, + "learning_rate": 6.68585402179851e-07, + "loss": 0.2241, + "step": 1165 + }, + { + "epoch": 4.553606237816764, + "grad_norm": 1.2945024967193604, + "learning_rate": 6.125780631754163e-07, + "loss": 0.2127, + "step": 1170 + }, + { + "epoch": 4.573099415204679, + "grad_norm": 1.2616260051727295, + "learning_rate": 5.589714885864716e-07, + "loss": 0.1986, + "step": 1175 + }, + { + "epoch": 4.592592592592593, + "grad_norm": 1.448205590248108, + "learning_rate": 5.077746235011122e-07, + "loss": 0.2123, + "step": 1180 + }, + { + "epoch": 4.6120857699805065, + "grad_norm": 1.518507480621338, + "learning_rate": 4.589960109100444e-07, + "loss": 0.2093, + "step": 1185 + }, + { + "epoch": 4.631578947368421, + "grad_norm": 1.3452050685882568, + "learning_rate": 4.1264379028105147e-07, + "loss": 0.1927, + "step": 1190 + }, + { + "epoch": 4.651072124756335, + "grad_norm": 1.320759892463684, + "learning_rate": 3.687256962007962e-07, + "loss": 0.2179, + "step": 1195 + }, + { + "epoch": 4.670565302144249, + "grad_norm": 1.398247480392456, + "learning_rate": 3.2724905708419517e-07, + "loss": 0.2136, + "step": 1200 + }, + { + "epoch": 4.690058479532164, + "grad_norm": 1.3366625308990479, + "learning_rate": 2.8822079395154357e-07, + "loss": 0.201, + "step": 1205 + }, + { + "epoch": 4.709551656920078, + "grad_norm": 1.2992753982543945, + "learning_rate": 2.516474192736473e-07, + "loss": 0.204, + "step": 1210 + }, + { + "epoch": 4.7290448343079925, + "grad_norm": 1.391621708869934, + "learning_rate": 2.1753503588511247e-07, + "loss": 0.2317, + "step": 1215 + }, + { + "epoch": 4.748538011695906, + "grad_norm": 1.2902519702911377, + "learning_rate": 1.8588933596599134e-07, + "loss": 0.205, + "step": 1220 + }, + { + "epoch": 4.76803118908382, + "grad_norm": 1.2960290908813477, + "learning_rate": 1.5671560009195894e-07, + "loss": 0.2118, + "step": 1225 + }, + { + "epoch": 4.787524366471735, + "grad_norm": 1.376568078994751, + "learning_rate": 1.30018696353158e-07, + "loss": 0.2157, + "step": 1230 + }, + { + "epoch": 4.807017543859649, + "grad_norm": 1.4016090631484985, + "learning_rate": 1.0580307954189184e-07, + "loss": 0.1956, + "step": 1235 + }, + { + "epoch": 4.826510721247564, + "grad_norm": 1.2883591651916504, + "learning_rate": 8.407279040926919e-08, + "loss": 0.2051, + "step": 1240 + }, + { + "epoch": 4.846003898635478, + "grad_norm": 1.2717663049697876, + "learning_rate": 6.483145499094345e-08, + "loss": 0.2094, + "step": 1245 + }, + { + "epoch": 4.8654970760233915, + "grad_norm": 1.3355618715286255, + "learning_rate": 4.8082284002050056e-08, + "loss": 0.1952, + "step": 1250 + }, + { + "epoch": 4.884990253411306, + "grad_norm": 1.3054178953170776, + "learning_rate": 3.382807230145002e-08, + "loss": 0.2088, + "step": 1255 + }, + { + "epoch": 4.90448343079922, + "grad_norm": 1.2859562635421753, + "learning_rate": 2.2071198425366314e-08, + "loss": 0.1884, + "step": 1260 + }, + { + "epoch": 4.923976608187134, + "grad_norm": 1.4428141117095947, + "learning_rate": 1.2813624190484707e-08, + "loss": 0.2352, + "step": 1265 + }, + { + "epoch": 4.943469785575049, + "grad_norm": 1.3250340223312378, + "learning_rate": 6.056894366598953e-09, + "loss": 0.2082, + "step": 1270 + }, + { + "epoch": 4.962962962962963, + "grad_norm": 1.2995291948318481, + "learning_rate": 1.8021364188369749e-09, + "loss": 0.2173, + "step": 1275 + }, + { + "epoch": 4.982456140350877, + "grad_norm": 1.2339032888412476, + "learning_rate": 5.006031952969892e-11, + "loss": 0.2294, + "step": 1280 + } + ], + "logging_steps": 5, + "max_steps": 1280, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 20000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 2.2211210999895163e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/training_args.bin b/not_grounded/0_128_e5_3e-5/checkpoint-1280/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4ce5ba5a4e8b2f8e0972e0e67dc1879011fda1a2 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa02dca237b948f004cb8d3e65da8c4d5a5b94ef5684cb38197d8ffa2ad8f171 +size 7800 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-1280/zero_to_fp32.py b/not_grounded/0_128_e5_3e-5/checkpoint-1280/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-1280/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/README.md b/not_grounded/0_128_e5_3e-5/checkpoint-257/README.md new file mode 100644 index 0000000000000000000000000000000000000000..0994074add6527289eb93335320fbd12270c4e7f --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.12.0 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/adapter_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-257/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2afdcf8899170963495ad14c67c715c56d8fd480 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/adapter_config.json @@ -0,0 +1,34 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 256, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 128, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "gate_proj", + "o_proj", + "v_proj", + "q_proj", + "down_proj", + "up_proj", + "k_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/adapter_model.safetensors b/not_grounded/0_128_e5_3e-5/checkpoint-257/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b13173f3f2830bb9d93a1756ed7dd5ecccdbefda --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b5bfa18a6deaf765ea0f34852f0f5b33e2db7e0065688fad196dd2b8d624b624 +size 671150064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/latest b/not_grounded/0_128_e5_3e-5/checkpoint-257/latest new file mode 100644 index 0000000000000000000000000000000000000000..b747f9725067064e241a7a3bed90583971af8ad1 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/latest @@ -0,0 +1 @@ +global_step256 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/rng_state_0.pth b/not_grounded/0_128_e5_3e-5/checkpoint-257/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..459bbacae1b7ac859e9589043cdd9687db254ce8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d67af647462a2969771f1e4d4fe4ea0abb26883fc7985f8f20007cd1cd70f596 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/rng_state_1.pth b/not_grounded/0_128_e5_3e-5/checkpoint-257/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..1b77b2f43e7afdce4df424aab7c5262e0f6964d1 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb2233ad164435e2ce6aa90e2da6219ec6a44ee55807659d8f82a31454c521cf +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/scheduler.pt b/not_grounded/0_128_e5_3e-5/checkpoint-257/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..a2bc227815bd1ab2dfd370e1fbad139fe7d91ecb --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67eec0c18f43b7f436b5fc94546c1ce92a9edbe5b7af11490de7bd422e596d2f +size 1064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/special_tokens_map.json b/not_grounded/0_128_e5_3e-5/checkpoint-257/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..b43be96621d147110fb8a18b5776ec6e38516127 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer.json b/not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6ac83ff885a7c13c549808db25dd230af9d67648 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384a7e7c676f7be2e5d2e8449c508be9b00e5b18c5b3c39ebc626e96b3f4b988 +size 17210019 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..8ae1bc63bd6e5ca8a863628311061c143679ff93 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/tokenizer_config.json @@ -0,0 +1,2064 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizer" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/trainer_state.json b/not_grounded/0_128_e5_3e-5/checkpoint-257/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..111fedad1acd12605fb367f37d070f5c58d3e9b8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/trainer_state.json @@ -0,0 +1,391 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 1.0, + "eval_steps": 500, + "global_step": 257, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.01949317738791423, + "grad_norm": 0.6081938743591309, + "learning_rate": 1.875e-06, + "loss": 1.7943, + "step": 5 + }, + { + "epoch": 0.03898635477582846, + "grad_norm": 0.5703465938568115, + "learning_rate": 4.21875e-06, + "loss": 1.7931, + "step": 10 + }, + { + "epoch": 0.05847953216374269, + "grad_norm": 0.4802420139312744, + "learning_rate": 6.5625e-06, + "loss": 1.8677, + "step": 15 + }, + { + "epoch": 0.07797270955165692, + "grad_norm": 0.4527171552181244, + "learning_rate": 8.90625e-06, + "loss": 1.7498, + "step": 20 + }, + { + "epoch": 0.09746588693957114, + "grad_norm": 0.41348591446876526, + "learning_rate": 1.125e-05, + "loss": 1.7402, + "step": 25 + }, + { + "epoch": 0.11695906432748537, + "grad_norm": 0.44569578766822815, + "learning_rate": 1.359375e-05, + "loss": 1.6984, + "step": 30 + }, + { + "epoch": 0.1364522417153996, + "grad_norm": 0.439203143119812, + "learning_rate": 1.59375e-05, + "loss": 1.6548, + "step": 35 + }, + { + "epoch": 0.15594541910331383, + "grad_norm": 0.4098866879940033, + "learning_rate": 1.828125e-05, + "loss": 1.7276, + "step": 40 + }, + { + "epoch": 0.17543859649122806, + "grad_norm": 0.42010200023651123, + "learning_rate": 2.0625e-05, + "loss": 1.7037, + "step": 45 + }, + { + "epoch": 0.1949317738791423, + "grad_norm": 0.4497705399990082, + "learning_rate": 2.296875e-05, + "loss": 1.6829, + "step": 50 + }, + { + "epoch": 0.21442495126705652, + "grad_norm": 0.4349888563156128, + "learning_rate": 2.5312500000000002e-05, + "loss": 1.696, + "step": 55 + }, + { + "epoch": 0.23391812865497075, + "grad_norm": 0.45078787207603455, + "learning_rate": 2.765625e-05, + "loss": 1.6246, + "step": 60 + }, + { + "epoch": 0.253411306042885, + "grad_norm": 0.44585877656936646, + "learning_rate": 3e-05, + "loss": 1.6324, + "step": 65 + }, + { + "epoch": 0.2729044834307992, + "grad_norm": 0.4684946537017822, + "learning_rate": 2.9998748508718575e-05, + "loss": 1.5858, + "step": 70 + }, + { + "epoch": 0.29239766081871343, + "grad_norm": 0.5270784497261047, + "learning_rate": 2.9994994243705013e-05, + "loss": 1.5719, + "step": 75 + }, + { + "epoch": 0.31189083820662766, + "grad_norm": 0.5215119123458862, + "learning_rate": 2.9988737831416642e-05, + "loss": 1.5858, + "step": 80 + }, + { + "epoch": 0.3313840155945419, + "grad_norm": 0.5029568672180176, + "learning_rate": 2.9979980315832853e-05, + "loss": 1.4428, + "step": 85 + }, + { + "epoch": 0.3508771929824561, + "grad_norm": 0.5102682709693909, + "learning_rate": 2.9968723158280906e-05, + "loss": 1.5211, + "step": 90 + }, + { + "epoch": 0.37037037037037035, + "grad_norm": 0.5617806911468506, + "learning_rate": 2.995496823719206e-05, + "loss": 1.4377, + "step": 95 + }, + { + "epoch": 0.3898635477582846, + "grad_norm": 0.5236561298370361, + "learning_rate": 2.9938717847788167e-05, + "loss": 1.464, + "step": 100 + }, + { + "epoch": 0.4093567251461988, + "grad_norm": 0.6214772462844849, + "learning_rate": 2.9919974701698638e-05, + "loss": 1.4009, + "step": 105 + }, + { + "epoch": 0.42884990253411304, + "grad_norm": 0.5746711492538452, + "learning_rate": 2.989874192650801e-05, + "loss": 1.3927, + "step": 110 + }, + { + "epoch": 0.44834307992202727, + "grad_norm": 0.6362848877906799, + "learning_rate": 2.9875023065234003e-05, + "loss": 1.4212, + "step": 115 + }, + { + "epoch": 0.4678362573099415, + "grad_norm": 0.6465691328048706, + "learning_rate": 2.984882207573638e-05, + "loss": 1.4462, + "step": 120 + }, + { + "epoch": 0.4873294346978557, + "grad_norm": 0.6339296698570251, + "learning_rate": 2.982014333005645e-05, + "loss": 1.337, + "step": 125 + }, + { + "epoch": 0.50682261208577, + "grad_norm": 0.6140757203102112, + "learning_rate": 2.9788991613687575e-05, + "loss": 1.3289, + "step": 130 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 0.6720002293586731, + "learning_rate": 2.9755372124776616e-05, + "loss": 1.3323, + "step": 135 + }, + { + "epoch": 0.5458089668615984, + "grad_norm": 0.6524502038955688, + "learning_rate": 2.971929047325654e-05, + "loss": 1.3166, + "step": 140 + }, + { + "epoch": 0.5653021442495126, + "grad_norm": 0.7752301692962646, + "learning_rate": 2.968075267991032e-05, + "loss": 1.3543, + "step": 145 + }, + { + "epoch": 0.5847953216374269, + "grad_norm": 0.6975963711738586, + "learning_rate": 2.963976517536627e-05, + "loss": 1.2675, + "step": 150 + }, + { + "epoch": 0.6042884990253411, + "grad_norm": 0.7260017991065979, + "learning_rate": 2.9596334799025003e-05, + "loss": 1.3213, + "step": 155 + }, + { + "epoch": 0.6237816764132553, + "grad_norm": 0.8624387979507446, + "learning_rate": 2.9550468797918162e-05, + "loss": 1.2492, + "step": 160 + }, + { + "epoch": 0.6432748538011696, + "grad_norm": 0.7931470274925232, + "learning_rate": 2.950217482549915e-05, + "loss": 1.2353, + "step": 165 + }, + { + "epoch": 0.6627680311890838, + "grad_norm": 0.7223592400550842, + "learning_rate": 2.9451460940366025e-05, + "loss": 1.1868, + "step": 170 + }, + { + "epoch": 0.682261208576998, + "grad_norm": 0.7910372614860535, + "learning_rate": 2.9398335604916797e-05, + "loss": 1.2148, + "step": 175 + }, + { + "epoch": 0.7017543859649122, + "grad_norm": 0.851186990737915, + "learning_rate": 2.9342807683937352e-05, + "loss": 1.2595, + "step": 180 + }, + { + "epoch": 0.7212475633528265, + "grad_norm": 0.8052693009376526, + "learning_rate": 2.928488644312222e-05, + "loss": 1.2406, + "step": 185 + }, + { + "epoch": 0.7407407407407407, + "grad_norm": 0.8178960680961609, + "learning_rate": 2.9224581547528453e-05, + "loss": 1.2197, + "step": 190 + }, + { + "epoch": 0.7602339181286549, + "grad_norm": 0.8441738486289978, + "learning_rate": 2.916190305996286e-05, + "loss": 1.1726, + "step": 195 + }, + { + "epoch": 0.7797270955165692, + "grad_norm": 0.9862117767333984, + "learning_rate": 2.909686143930287e-05, + "loss": 1.1745, + "step": 200 + }, + { + "epoch": 0.7992202729044834, + "grad_norm": 0.8886749148368835, + "learning_rate": 2.902946753875131e-05, + "loss": 1.1447, + "step": 205 + }, + { + "epoch": 0.8187134502923976, + "grad_norm": 0.8914117813110352, + "learning_rate": 2.895973260402537e-05, + "loss": 1.1248, + "step": 210 + }, + { + "epoch": 0.8382066276803118, + "grad_norm": 1.1335160732269287, + "learning_rate": 2.8887668271480098e-05, + "loss": 1.1291, + "step": 215 + }, + { + "epoch": 0.8576998050682261, + "grad_norm": 1.0726529359817505, + "learning_rate": 2.8813286566166674e-05, + "loss": 1.1646, + "step": 220 + }, + { + "epoch": 0.8771929824561403, + "grad_norm": 1.0396595001220703, + "learning_rate": 2.873659989982586e-05, + "loss": 1.0894, + "step": 225 + }, + { + "epoch": 0.8966861598440545, + "grad_norm": 0.9956820011138916, + "learning_rate": 2.8657621068816903e-05, + "loss": 1.0984, + "step": 230 + }, + { + "epoch": 0.9161793372319688, + "grad_norm": 0.9483380317687988, + "learning_rate": 2.857636325198225e-05, + "loss": 1.0216, + "step": 235 + }, + { + "epoch": 0.935672514619883, + "grad_norm": 0.8558198809623718, + "learning_rate": 2.849284000844849e-05, + "loss": 1.105, + "step": 240 + }, + { + "epoch": 0.9551656920077972, + "grad_norm": 1.005858063697815, + "learning_rate": 2.8407065275363756e-05, + "loss": 1.0714, + "step": 245 + }, + { + "epoch": 0.9746588693957114, + "grad_norm": 0.9553005695343018, + "learning_rate": 2.8319053365572135e-05, + "loss": 1.0339, + "step": 250 + }, + { + "epoch": 0.9941520467836257, + "grad_norm": 0.9755716919898987, + "learning_rate": 2.8228818965225325e-05, + "loss": 1.0532, + "step": 255 + } + ], + "logging_steps": 5, + "max_steps": 1280, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 20000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 4.4492096218228326e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/training_args.bin b/not_grounded/0_128_e5_3e-5/checkpoint-257/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4ce5ba5a4e8b2f8e0972e0e67dc1879011fda1a2 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa02dca237b948f004cb8d3e65da8c4d5a5b94ef5684cb38197d8ffa2ad8f171 +size 7800 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-257/zero_to_fp32.py b/not_grounded/0_128_e5_3e-5/checkpoint-257/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-257/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/README.md b/not_grounded/0_128_e5_3e-5/checkpoint-514/README.md new file mode 100644 index 0000000000000000000000000000000000000000..0994074add6527289eb93335320fbd12270c4e7f --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.12.0 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/adapter_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-514/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2afdcf8899170963495ad14c67c715c56d8fd480 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/adapter_config.json @@ -0,0 +1,34 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 256, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 128, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "gate_proj", + "o_proj", + "v_proj", + "q_proj", + "down_proj", + "up_proj", + "k_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/adapter_model.safetensors b/not_grounded/0_128_e5_3e-5/checkpoint-514/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eda1a653bfc916e2b3606d5dc8739dee3e21354a --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd8ac42a1ec76c69433b85d0c4fce3c4a946fcb34953aecbe71a077f38acf045 +size 671150064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/latest b/not_grounded/0_128_e5_3e-5/checkpoint-514/latest new file mode 100644 index 0000000000000000000000000000000000000000..2990f81fc6ef312b3c6ef531f6aaeb8e1a3fb967 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/latest @@ -0,0 +1 @@ +global_step513 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/rng_state_0.pth b/not_grounded/0_128_e5_3e-5/checkpoint-514/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..47cc72ba9bf0f971a4f5d843e6127832c2b0aa53 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e3031018b9287cd949469f9d20870ad459b0cdf84c026fa64197d1b61099436 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/rng_state_1.pth b/not_grounded/0_128_e5_3e-5/checkpoint-514/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..b090bbf1fda0298a4c83513650fa19feeedb84e7 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2afbaaeb7156e14ce17e2ea0990fa0cdb5485548ab1f7c494114e3b65f2e7ce2 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/scheduler.pt b/not_grounded/0_128_e5_3e-5/checkpoint-514/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..647ff87f21a11f1f9fd5aed97572c8cade5bc396 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08bb599f8f80d068049d45be04c8657de2a8ecc9cfa9fadcb4c0e6bf9802d841 +size 1064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/special_tokens_map.json b/not_grounded/0_128_e5_3e-5/checkpoint-514/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..b43be96621d147110fb8a18b5776ec6e38516127 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer.json b/not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6ac83ff885a7c13c549808db25dd230af9d67648 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384a7e7c676f7be2e5d2e8449c508be9b00e5b18c5b3c39ebc626e96b3f4b988 +size 17210019 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..8ae1bc63bd6e5ca8a863628311061c143679ff93 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/tokenizer_config.json @@ -0,0 +1,2064 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizer" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/trainer_state.json b/not_grounded/0_128_e5_3e-5/checkpoint-514/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..f1f950d030f3125424cc184eb1a93889ea36790a --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/trainer_state.json @@ -0,0 +1,748 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 2.0, + "eval_steps": 500, + "global_step": 514, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.01949317738791423, + "grad_norm": 0.6081938743591309, + "learning_rate": 1.875e-06, + "loss": 1.7943, + "step": 5 + }, + { + "epoch": 0.03898635477582846, + "grad_norm": 0.5703465938568115, + "learning_rate": 4.21875e-06, + "loss": 1.7931, + "step": 10 + }, + { + "epoch": 0.05847953216374269, + "grad_norm": 0.4802420139312744, + "learning_rate": 6.5625e-06, + "loss": 1.8677, + "step": 15 + }, + { + "epoch": 0.07797270955165692, + "grad_norm": 0.4527171552181244, + "learning_rate": 8.90625e-06, + "loss": 1.7498, + "step": 20 + }, + { + "epoch": 0.09746588693957114, + "grad_norm": 0.41348591446876526, + "learning_rate": 1.125e-05, + "loss": 1.7402, + "step": 25 + }, + { + "epoch": 0.11695906432748537, + "grad_norm": 0.44569578766822815, + "learning_rate": 1.359375e-05, + "loss": 1.6984, + "step": 30 + }, + { + "epoch": 0.1364522417153996, + "grad_norm": 0.439203143119812, + "learning_rate": 1.59375e-05, + "loss": 1.6548, + "step": 35 + }, + { + "epoch": 0.15594541910331383, + "grad_norm": 0.4098866879940033, + "learning_rate": 1.828125e-05, + "loss": 1.7276, + "step": 40 + }, + { + "epoch": 0.17543859649122806, + "grad_norm": 0.42010200023651123, + "learning_rate": 2.0625e-05, + "loss": 1.7037, + "step": 45 + }, + { + "epoch": 0.1949317738791423, + "grad_norm": 0.4497705399990082, + "learning_rate": 2.296875e-05, + "loss": 1.6829, + "step": 50 + }, + { + "epoch": 0.21442495126705652, + "grad_norm": 0.4349888563156128, + "learning_rate": 2.5312500000000002e-05, + "loss": 1.696, + "step": 55 + }, + { + "epoch": 0.23391812865497075, + "grad_norm": 0.45078787207603455, + "learning_rate": 2.765625e-05, + "loss": 1.6246, + "step": 60 + }, + { + "epoch": 0.253411306042885, + "grad_norm": 0.44585877656936646, + "learning_rate": 3e-05, + "loss": 1.6324, + "step": 65 + }, + { + "epoch": 0.2729044834307992, + "grad_norm": 0.4684946537017822, + "learning_rate": 2.9998748508718575e-05, + "loss": 1.5858, + "step": 70 + }, + { + "epoch": 0.29239766081871343, + "grad_norm": 0.5270784497261047, + "learning_rate": 2.9994994243705013e-05, + "loss": 1.5719, + "step": 75 + }, + { + "epoch": 0.31189083820662766, + "grad_norm": 0.5215119123458862, + "learning_rate": 2.9988737831416642e-05, + "loss": 1.5858, + "step": 80 + }, + { + "epoch": 0.3313840155945419, + "grad_norm": 0.5029568672180176, + "learning_rate": 2.9979980315832853e-05, + "loss": 1.4428, + "step": 85 + }, + { + "epoch": 0.3508771929824561, + "grad_norm": 0.5102682709693909, + "learning_rate": 2.9968723158280906e-05, + "loss": 1.5211, + "step": 90 + }, + { + "epoch": 0.37037037037037035, + "grad_norm": 0.5617806911468506, + "learning_rate": 2.995496823719206e-05, + "loss": 1.4377, + "step": 95 + }, + { + "epoch": 0.3898635477582846, + "grad_norm": 0.5236561298370361, + "learning_rate": 2.9938717847788167e-05, + "loss": 1.464, + "step": 100 + }, + { + "epoch": 0.4093567251461988, + "grad_norm": 0.6214772462844849, + "learning_rate": 2.9919974701698638e-05, + "loss": 1.4009, + "step": 105 + }, + { + "epoch": 0.42884990253411304, + "grad_norm": 0.5746711492538452, + "learning_rate": 2.989874192650801e-05, + "loss": 1.3927, + "step": 110 + }, + { + "epoch": 0.44834307992202727, + "grad_norm": 0.6362848877906799, + "learning_rate": 2.9875023065234003e-05, + "loss": 1.4212, + "step": 115 + }, + { + "epoch": 0.4678362573099415, + "grad_norm": 0.6465691328048706, + "learning_rate": 2.984882207573638e-05, + "loss": 1.4462, + "step": 120 + }, + { + "epoch": 0.4873294346978557, + "grad_norm": 0.6339296698570251, + "learning_rate": 2.982014333005645e-05, + "loss": 1.337, + "step": 125 + }, + { + "epoch": 0.50682261208577, + "grad_norm": 0.6140757203102112, + "learning_rate": 2.9788991613687575e-05, + "loss": 1.3289, + "step": 130 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 0.6720002293586731, + "learning_rate": 2.9755372124776616e-05, + "loss": 1.3323, + "step": 135 + }, + { + "epoch": 0.5458089668615984, + "grad_norm": 0.6524502038955688, + "learning_rate": 2.971929047325654e-05, + "loss": 1.3166, + "step": 140 + }, + { + "epoch": 0.5653021442495126, + "grad_norm": 0.7752301692962646, + "learning_rate": 2.968075267991032e-05, + "loss": 1.3543, + "step": 145 + }, + { + "epoch": 0.5847953216374269, + "grad_norm": 0.6975963711738586, + "learning_rate": 2.963976517536627e-05, + "loss": 1.2675, + "step": 150 + }, + { + "epoch": 0.6042884990253411, + "grad_norm": 0.7260017991065979, + "learning_rate": 2.9596334799025003e-05, + "loss": 1.3213, + "step": 155 + }, + { + "epoch": 0.6237816764132553, + "grad_norm": 0.8624387979507446, + "learning_rate": 2.9550468797918162e-05, + "loss": 1.2492, + "step": 160 + }, + { + "epoch": 0.6432748538011696, + "grad_norm": 0.7931470274925232, + "learning_rate": 2.950217482549915e-05, + "loss": 1.2353, + "step": 165 + }, + { + "epoch": 0.6627680311890838, + "grad_norm": 0.7223592400550842, + "learning_rate": 2.9451460940366025e-05, + "loss": 1.1868, + "step": 170 + }, + { + "epoch": 0.682261208576998, + "grad_norm": 0.7910372614860535, + "learning_rate": 2.9398335604916797e-05, + "loss": 1.2148, + "step": 175 + }, + { + "epoch": 0.7017543859649122, + "grad_norm": 0.851186990737915, + "learning_rate": 2.9342807683937352e-05, + "loss": 1.2595, + "step": 180 + }, + { + "epoch": 0.7212475633528265, + "grad_norm": 0.8052693009376526, + "learning_rate": 2.928488644312222e-05, + "loss": 1.2406, + "step": 185 + }, + { + "epoch": 0.7407407407407407, + "grad_norm": 0.8178960680961609, + "learning_rate": 2.9224581547528453e-05, + "loss": 1.2197, + "step": 190 + }, + { + "epoch": 0.7602339181286549, + "grad_norm": 0.8441738486289978, + "learning_rate": 2.916190305996286e-05, + "loss": 1.1726, + "step": 195 + }, + { + "epoch": 0.7797270955165692, + "grad_norm": 0.9862117767333984, + "learning_rate": 2.909686143930287e-05, + "loss": 1.1745, + "step": 200 + }, + { + "epoch": 0.7992202729044834, + "grad_norm": 0.8886749148368835, + "learning_rate": 2.902946753875131e-05, + "loss": 1.1447, + "step": 205 + }, + { + "epoch": 0.8187134502923976, + "grad_norm": 0.8914117813110352, + "learning_rate": 2.895973260402537e-05, + "loss": 1.1248, + "step": 210 + }, + { + "epoch": 0.8382066276803118, + "grad_norm": 1.1335160732269287, + "learning_rate": 2.8887668271480098e-05, + "loss": 1.1291, + "step": 215 + }, + { + "epoch": 0.8576998050682261, + "grad_norm": 1.0726529359817505, + "learning_rate": 2.8813286566166674e-05, + "loss": 1.1646, + "step": 220 + }, + { + "epoch": 0.8771929824561403, + "grad_norm": 1.0396595001220703, + "learning_rate": 2.873659989982586e-05, + "loss": 1.0894, + "step": 225 + }, + { + "epoch": 0.8966861598440545, + "grad_norm": 0.9956820011138916, + "learning_rate": 2.8657621068816903e-05, + "loss": 1.0984, + "step": 230 + }, + { + "epoch": 0.9161793372319688, + "grad_norm": 0.9483380317687988, + "learning_rate": 2.857636325198225e-05, + "loss": 1.0216, + "step": 235 + }, + { + "epoch": 0.935672514619883, + "grad_norm": 0.8558198809623718, + "learning_rate": 2.849284000844849e-05, + "loss": 1.105, + "step": 240 + }, + { + "epoch": 0.9551656920077972, + "grad_norm": 1.005858063697815, + "learning_rate": 2.8407065275363756e-05, + "loss": 1.0714, + "step": 245 + }, + { + "epoch": 0.9746588693957114, + "grad_norm": 0.9553005695343018, + "learning_rate": 2.8319053365572135e-05, + "loss": 1.0339, + "step": 250 + }, + { + "epoch": 0.9941520467836257, + "grad_norm": 0.9755716919898987, + "learning_rate": 2.8228818965225325e-05, + "loss": 1.0532, + "step": 255 + }, + { + "epoch": 1.0116959064327484, + "grad_norm": 1.0532433986663818, + "learning_rate": 2.8136377131332043e-05, + "loss": 1.0005, + "step": 260 + }, + { + "epoch": 1.0311890838206628, + "grad_norm": 0.9859170913696289, + "learning_rate": 2.8041743289245503e-05, + "loss": 0.9084, + "step": 265 + }, + { + "epoch": 1.050682261208577, + "grad_norm": 0.9481168389320374, + "learning_rate": 2.7944933230089484e-05, + "loss": 0.9431, + "step": 270 + }, + { + "epoch": 1.0701754385964912, + "grad_norm": 0.9678940773010254, + "learning_rate": 2.784596310812331e-05, + "loss": 0.9262, + "step": 275 + }, + { + "epoch": 1.0896686159844053, + "grad_norm": 0.9351955652236938, + "learning_rate": 2.774484943804629e-05, + "loss": 0.9044, + "step": 280 + }, + { + "epoch": 1.1091617933723197, + "grad_norm": 1.0261086225509644, + "learning_rate": 2.764160909224196e-05, + "loss": 0.9579, + "step": 285 + }, + { + "epoch": 1.128654970760234, + "grad_norm": 0.9796629548072815, + "learning_rate": 2.7536259297962674e-05, + "loss": 0.9283, + "step": 290 + }, + { + "epoch": 1.1481481481481481, + "grad_norm": 0.9690019488334656, + "learning_rate": 2.7428817634454973e-05, + "loss": 0.9017, + "step": 295 + }, + { + "epoch": 1.1676413255360623, + "grad_norm": 1.1676615476608276, + "learning_rate": 2.7319302030026207e-05, + "loss": 0.8919, + "step": 300 + }, + { + "epoch": 1.1871345029239766, + "grad_norm": 1.1467899084091187, + "learning_rate": 2.7207730759052925e-05, + "loss": 0.871, + "step": 305 + }, + { + "epoch": 1.206627680311891, + "grad_norm": 1.0315958261489868, + "learning_rate": 2.7094122438931513e-05, + "loss": 0.8388, + "step": 310 + }, + { + "epoch": 1.226120857699805, + "grad_norm": 1.0231361389160156, + "learning_rate": 2.697849602697159e-05, + "loss": 0.8841, + "step": 315 + }, + { + "epoch": 1.2456140350877192, + "grad_norm": 0.995548665523529, + "learning_rate": 2.6860870817232682e-05, + "loss": 0.8879, + "step": 320 + }, + { + "epoch": 1.2651072124756335, + "grad_norm": 1.102622151374817, + "learning_rate": 2.6741266437304718e-05, + "loss": 0.857, + "step": 325 + }, + { + "epoch": 1.2846003898635479, + "grad_norm": 1.3752007484436035, + "learning_rate": 2.661970284503286e-05, + "loss": 0.8027, + "step": 330 + }, + { + "epoch": 1.304093567251462, + "grad_norm": 1.129936933517456, + "learning_rate": 2.6496200325187223e-05, + "loss": 0.855, + "step": 335 + }, + { + "epoch": 1.323586744639376, + "grad_norm": 1.2070459127426147, + "learning_rate": 2.6370779486078047e-05, + "loss": 0.8544, + "step": 340 + }, + { + "epoch": 1.3430799220272904, + "grad_norm": 1.1400156021118164, + "learning_rate": 2.6243461256116892e-05, + "loss": 0.9026, + "step": 345 + }, + { + "epoch": 1.3625730994152048, + "grad_norm": 0.9957541823387146, + "learning_rate": 2.611426688032439e-05, + "loss": 0.8115, + "step": 350 + }, + { + "epoch": 1.3820662768031189, + "grad_norm": 1.0340908765792847, + "learning_rate": 2.598321791678519e-05, + "loss": 0.8314, + "step": 355 + }, + { + "epoch": 1.401559454191033, + "grad_norm": 1.1327755451202393, + "learning_rate": 2.5850336233050677e-05, + "loss": 0.8319, + "step": 360 + }, + { + "epoch": 1.4210526315789473, + "grad_norm": 1.0419026613235474, + "learning_rate": 2.5715644002489998e-05, + "loss": 0.8115, + "step": 365 + }, + { + "epoch": 1.4405458089668617, + "grad_norm": 1.080306053161621, + "learning_rate": 2.557916370059012e-05, + "loss": 0.7734, + "step": 370 + }, + { + "epoch": 1.4600389863547758, + "grad_norm": 1.1247351169586182, + "learning_rate": 2.544091810120543e-05, + "loss": 0.7968, + "step": 375 + }, + { + "epoch": 1.47953216374269, + "grad_norm": 1.1105760335922241, + "learning_rate": 2.530093027275757e-05, + "loss": 0.7994, + "step": 380 + }, + { + "epoch": 1.4990253411306043, + "grad_norm": 1.3396145105361938, + "learning_rate": 2.5159223574386117e-05, + "loss": 0.7891, + "step": 385 + }, + { + "epoch": 1.5185185185185186, + "grad_norm": 1.2513933181762695, + "learning_rate": 2.501582165205074e-05, + "loss": 0.7825, + "step": 390 + }, + { + "epoch": 1.5380116959064327, + "grad_norm": 1.139088749885559, + "learning_rate": 2.4870748434585514e-05, + "loss": 0.7875, + "step": 395 + }, + { + "epoch": 1.5575048732943468, + "grad_norm": 1.2570606470108032, + "learning_rate": 2.4724028129706012e-05, + "loss": 0.7524, + "step": 400 + }, + { + "epoch": 1.5769980506822612, + "grad_norm": 1.115220069885254, + "learning_rate": 2.4575685219969884e-05, + "loss": 0.7855, + "step": 405 + }, + { + "epoch": 1.5964912280701755, + "grad_norm": 1.2705992460250854, + "learning_rate": 2.442574445869156e-05, + "loss": 0.7338, + "step": 410 + }, + { + "epoch": 1.6159844054580896, + "grad_norm": 1.1325453519821167, + "learning_rate": 2.4274230865811763e-05, + "loss": 0.7528, + "step": 415 + }, + { + "epoch": 1.6354775828460038, + "grad_norm": 1.301644206047058, + "learning_rate": 2.4121169723722566e-05, + "loss": 0.7445, + "step": 420 + }, + { + "epoch": 1.654970760233918, + "grad_norm": 1.4577105045318604, + "learning_rate": 2.396658657304861e-05, + "loss": 0.7335, + "step": 425 + }, + { + "epoch": 1.6744639376218324, + "grad_norm": 1.1624274253845215, + "learning_rate": 2.381050720838528e-05, + "loss": 0.7146, + "step": 430 + }, + { + "epoch": 1.6939571150097466, + "grad_norm": 1.4543778896331787, + "learning_rate": 2.3652957673994448e-05, + "loss": 0.7338, + "step": 435 + }, + { + "epoch": 1.7134502923976607, + "grad_norm": 1.208083987236023, + "learning_rate": 2.3493964259458603e-05, + "loss": 0.7187, + "step": 440 + }, + { + "epoch": 1.732943469785575, + "grad_norm": 1.1645222902297974, + "learning_rate": 2.3333553495294033e-05, + "loss": 0.7325, + "step": 445 + }, + { + "epoch": 1.7524366471734893, + "grad_norm": 1.2736948728561401, + "learning_rate": 2.317175214852377e-05, + "loss": 0.7152, + "step": 450 + }, + { + "epoch": 1.7719298245614035, + "grad_norm": 1.2995411157608032, + "learning_rate": 2.3008587218211127e-05, + "loss": 0.6953, + "step": 455 + }, + { + "epoch": 1.7914230019493176, + "grad_norm": 1.4111621379852295, + "learning_rate": 2.284408593095446e-05, + "loss": 0.7023, + "step": 460 + }, + { + "epoch": 1.810916179337232, + "grad_norm": 1.2578918933868408, + "learning_rate": 2.2678275736344014e-05, + "loss": 0.6771, + "step": 465 + }, + { + "epoch": 1.8304093567251463, + "grad_norm": 1.3310747146606445, + "learning_rate": 2.251118430238151e-05, + "loss": 0.6792, + "step": 470 + }, + { + "epoch": 1.8499025341130604, + "grad_norm": 1.1377298831939697, + "learning_rate": 2.2342839510863323e-05, + "loss": 0.7322, + "step": 475 + }, + { + "epoch": 1.8693957115009745, + "grad_norm": 1.2897000312805176, + "learning_rate": 2.2173269452727965e-05, + "loss": 0.6468, + "step": 480 + }, + { + "epoch": 1.8888888888888888, + "grad_norm": 1.1680046319961548, + "learning_rate": 2.200250242336868e-05, + "loss": 0.6785, + "step": 485 + }, + { + "epoch": 1.9083820662768032, + "grad_norm": 1.2389404773712158, + "learning_rate": 2.183056691791193e-05, + "loss": 0.6521, + "step": 490 + }, + { + "epoch": 1.9278752436647173, + "grad_norm": 1.1207393407821655, + "learning_rate": 2.1657491626462514e-05, + "loss": 0.6607, + "step": 495 + }, + { + "epoch": 1.9473684210526314, + "grad_norm": 1.1037189960479736, + "learning_rate": 2.1483305429316208e-05, + "loss": 0.6033, + "step": 500 + }, + { + "epoch": 1.9668615984405458, + "grad_norm": 1.1787821054458618, + "learning_rate": 2.1308037392140613e-05, + "loss": 0.6254, + "step": 505 + }, + { + "epoch": 1.98635477582846, + "grad_norm": 1.266696572303772, + "learning_rate": 2.113171676112513e-05, + "loss": 0.5962, + "step": 510 + } + ], + "logging_steps": 5, + "max_steps": 1280, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 20000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 8.89819342500266e+17, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/training_args.bin b/not_grounded/0_128_e5_3e-5/checkpoint-514/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4ce5ba5a4e8b2f8e0972e0e67dc1879011fda1a2 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa02dca237b948f004cb8d3e65da8c4d5a5b94ef5684cb38197d8ffa2ad8f171 +size 7800 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-514/zero_to_fp32.py b/not_grounded/0_128_e5_3e-5/checkpoint-514/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-514/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters) diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/README.md b/not_grounded/0_128_e5_3e-5/checkpoint-771/README.md new file mode 100644 index 0000000000000000000000000000000000000000..0994074add6527289eb93335320fbd12270c4e7f --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/README.md @@ -0,0 +1,202 @@ +--- +base_model: meta-llama/Llama-3.1-8B-Instruct +library_name: peft +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + + + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] +### Framework versions + +- PEFT 0.12.0 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/adapter_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-771/adapter_config.json new file mode 100644 index 0000000000000000000000000000000000000000..2afdcf8899170963495ad14c67c715c56d8fd480 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/adapter_config.json @@ -0,0 +1,34 @@ +{ + "alpha_pattern": {}, + "auto_mapping": null, + "base_model_name_or_path": "meta-llama/Llama-3.1-8B-Instruct", + "bias": "none", + "fan_in_fan_out": false, + "inference_mode": true, + "init_lora_weights": true, + "layer_replication": null, + "layers_pattern": null, + "layers_to_transform": null, + "loftq_config": {}, + "lora_alpha": 256, + "lora_dropout": 0.05, + "megatron_config": null, + "megatron_core": "megatron.core", + "modules_to_save": null, + "peft_type": "LORA", + "r": 128, + "rank_pattern": {}, + "revision": null, + "target_modules": [ + "gate_proj", + "o_proj", + "v_proj", + "q_proj", + "down_proj", + "up_proj", + "k_proj" + ], + "task_type": "CAUSAL_LM", + "use_dora": false, + "use_rslora": false +} \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/adapter_model.safetensors b/not_grounded/0_128_e5_3e-5/checkpoint-771/adapter_model.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3b720f0142e5461315bb657ba978ff6970fef946 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/adapter_model.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:606ff4eba3b8272e10098e8946090ac663b5fba17c0e5bb2aeb15b2b1e8ad4f3 +size 671150064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/latest b/not_grounded/0_128_e5_3e-5/checkpoint-771/latest new file mode 100644 index 0000000000000000000000000000000000000000..3ac51ce12972cbc46720f1d543f566d3c7557817 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/latest @@ -0,0 +1 @@ +global_step769 \ No newline at end of file diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/rng_state_0.pth b/not_grounded/0_128_e5_3e-5/checkpoint-771/rng_state_0.pth new file mode 100644 index 0000000000000000000000000000000000000000..fe702ab7a08fc5af18b0de11b54555329995fc61 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/rng_state_0.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f23f9c78521e414398c66c560db6526441078165dc87e68483544b4c14da28a4 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/rng_state_1.pth b/not_grounded/0_128_e5_3e-5/checkpoint-771/rng_state_1.pth new file mode 100644 index 0000000000000000000000000000000000000000..f5cf4edd3562df33c3f7be6acad72352ca503460 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/rng_state_1.pth @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8dbdbbde984e49f7c93e120bb4350f401e1432aaa699509d1568dac8f3890168 +size 14512 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/scheduler.pt b/not_grounded/0_128_e5_3e-5/checkpoint-771/scheduler.pt new file mode 100644 index 0000000000000000000000000000000000000000..1d4f77bd5a10208ddc02c3e220249476aba8ba00 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/scheduler.pt @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd14e9470d65cea84b3db363c5213e314f15d8ea73cc86e9effeb7cc2d6ac7df +size 1064 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/special_tokens_map.json b/not_grounded/0_128_e5_3e-5/checkpoint-771/special_tokens_map.json new file mode 100644 index 0000000000000000000000000000000000000000..b43be96621d147110fb8a18b5776ec6e38516127 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/special_tokens_map.json @@ -0,0 +1,17 @@ +{ + "bos_token": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "eos_token": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false + }, + "pad_token": "<|eot_id|>" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer.json b/not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer.json new file mode 100644 index 0000000000000000000000000000000000000000..6ac83ff885a7c13c549808db25dd230af9d67648 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer.json @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:384a7e7c676f7be2e5d2e8449c508be9b00e5b18c5b3c39ebc626e96b3f4b988 +size 17210019 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer_config.json b/not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer_config.json new file mode 100644 index 0000000000000000000000000000000000000000..8ae1bc63bd6e5ca8a863628311061c143679ff93 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/tokenizer_config.json @@ -0,0 +1,2064 @@ +{ + "added_tokens_decoder": { + "128000": { + "content": "<|begin_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128001": { + "content": "<|end_of_text|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128002": { + "content": "<|reserved_special_token_0|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128003": { + "content": "<|reserved_special_token_1|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128004": { + "content": "<|finetune_right_pad_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128005": { + "content": "<|reserved_special_token_2|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128006": { + "content": "<|start_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128007": { + "content": "<|end_header_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128008": { + "content": "<|eom_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128009": { + "content": "<|eot_id|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128010": { + "content": "<|python_tag|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128011": { + "content": "<|reserved_special_token_3|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128012": { + "content": "<|reserved_special_token_4|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128013": { + "content": "<|reserved_special_token_5|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128014": { + "content": "<|reserved_special_token_6|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128015": { + "content": "<|reserved_special_token_7|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128016": { + "content": "<|reserved_special_token_8|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128017": { + "content": "<|reserved_special_token_9|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128018": { + "content": "<|reserved_special_token_10|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128019": { + "content": "<|reserved_special_token_11|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128020": { + "content": "<|reserved_special_token_12|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128021": { + "content": "<|reserved_special_token_13|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128022": { + "content": "<|reserved_special_token_14|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128023": { + "content": "<|reserved_special_token_15|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128024": { + "content": "<|reserved_special_token_16|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128025": { + "content": "<|reserved_special_token_17|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128026": { + "content": "<|reserved_special_token_18|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128027": { + "content": "<|reserved_special_token_19|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128028": { + "content": "<|reserved_special_token_20|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128029": { + "content": "<|reserved_special_token_21|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128030": { + "content": "<|reserved_special_token_22|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128031": { + "content": "<|reserved_special_token_23|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128032": { + "content": "<|reserved_special_token_24|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128033": { + "content": "<|reserved_special_token_25|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128034": { + "content": "<|reserved_special_token_26|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128035": { + "content": "<|reserved_special_token_27|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128036": { + "content": "<|reserved_special_token_28|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128037": { + "content": "<|reserved_special_token_29|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128038": { + "content": "<|reserved_special_token_30|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128039": { + "content": "<|reserved_special_token_31|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128040": { + "content": "<|reserved_special_token_32|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128041": { + "content": "<|reserved_special_token_33|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128042": { + "content": "<|reserved_special_token_34|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128043": { + "content": "<|reserved_special_token_35|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128044": { + "content": "<|reserved_special_token_36|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128045": { + "content": "<|reserved_special_token_37|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128046": { + "content": "<|reserved_special_token_38|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128047": { + "content": "<|reserved_special_token_39|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128048": { + "content": "<|reserved_special_token_40|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128049": { + "content": "<|reserved_special_token_41|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128050": { + "content": "<|reserved_special_token_42|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128051": { + "content": "<|reserved_special_token_43|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128052": { + "content": "<|reserved_special_token_44|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128053": { + "content": "<|reserved_special_token_45|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128054": { + "content": "<|reserved_special_token_46|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128055": { + "content": "<|reserved_special_token_47|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128056": { + "content": "<|reserved_special_token_48|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128057": { + "content": "<|reserved_special_token_49|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128058": { + "content": "<|reserved_special_token_50|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128059": { + "content": "<|reserved_special_token_51|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128060": { + "content": "<|reserved_special_token_52|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128061": { + "content": "<|reserved_special_token_53|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128062": { + "content": "<|reserved_special_token_54|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128063": { + "content": "<|reserved_special_token_55|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128064": { + "content": "<|reserved_special_token_56|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128065": { + "content": "<|reserved_special_token_57|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128066": { + "content": "<|reserved_special_token_58|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128067": { + "content": "<|reserved_special_token_59|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128068": { + "content": "<|reserved_special_token_60|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128069": { + "content": "<|reserved_special_token_61|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128070": { + "content": "<|reserved_special_token_62|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128071": { + "content": "<|reserved_special_token_63|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128072": { + "content": "<|reserved_special_token_64|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128073": { + "content": "<|reserved_special_token_65|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128074": { + "content": "<|reserved_special_token_66|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128075": { + "content": "<|reserved_special_token_67|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128076": { + "content": "<|reserved_special_token_68|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128077": { + "content": "<|reserved_special_token_69|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128078": { + "content": "<|reserved_special_token_70|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128079": { + "content": "<|reserved_special_token_71|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128080": { + "content": "<|reserved_special_token_72|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128081": { + "content": "<|reserved_special_token_73|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128082": { + "content": "<|reserved_special_token_74|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128083": { + "content": "<|reserved_special_token_75|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128084": { + "content": "<|reserved_special_token_76|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128085": { + "content": "<|reserved_special_token_77|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128086": { + "content": "<|reserved_special_token_78|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128087": { + "content": "<|reserved_special_token_79|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128088": { + "content": "<|reserved_special_token_80|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128089": { + "content": "<|reserved_special_token_81|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128090": { + "content": "<|reserved_special_token_82|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128091": { + "content": "<|reserved_special_token_83|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128092": { + "content": "<|reserved_special_token_84|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128093": { + "content": "<|reserved_special_token_85|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128094": { + "content": "<|reserved_special_token_86|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128095": { + "content": "<|reserved_special_token_87|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128096": { + "content": "<|reserved_special_token_88|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128097": { + "content": "<|reserved_special_token_89|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128098": { + "content": "<|reserved_special_token_90|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128099": { + "content": "<|reserved_special_token_91|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128100": { + "content": "<|reserved_special_token_92|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128101": { + "content": "<|reserved_special_token_93|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128102": { + "content": "<|reserved_special_token_94|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128103": { + "content": "<|reserved_special_token_95|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128104": { + "content": "<|reserved_special_token_96|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128105": { + "content": "<|reserved_special_token_97|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128106": { + "content": "<|reserved_special_token_98|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128107": { + "content": "<|reserved_special_token_99|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128108": { + "content": "<|reserved_special_token_100|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128109": { + "content": "<|reserved_special_token_101|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128110": { + "content": "<|reserved_special_token_102|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128111": { + "content": "<|reserved_special_token_103|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128112": { + "content": "<|reserved_special_token_104|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128113": { + "content": "<|reserved_special_token_105|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128114": { + "content": "<|reserved_special_token_106|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128115": { + "content": "<|reserved_special_token_107|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128116": { + "content": "<|reserved_special_token_108|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128117": { + "content": "<|reserved_special_token_109|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128118": { + "content": "<|reserved_special_token_110|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128119": { + "content": "<|reserved_special_token_111|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128120": { + "content": "<|reserved_special_token_112|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128121": { + "content": "<|reserved_special_token_113|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128122": { + "content": "<|reserved_special_token_114|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128123": { + "content": "<|reserved_special_token_115|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128124": { + "content": "<|reserved_special_token_116|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128125": { + "content": "<|reserved_special_token_117|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128126": { + "content": "<|reserved_special_token_118|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128127": { + "content": "<|reserved_special_token_119|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128128": { + "content": "<|reserved_special_token_120|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128129": { + "content": "<|reserved_special_token_121|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128130": { + "content": "<|reserved_special_token_122|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128131": { + "content": "<|reserved_special_token_123|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128132": { + "content": "<|reserved_special_token_124|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128133": { + "content": "<|reserved_special_token_125|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128134": { + "content": "<|reserved_special_token_126|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128135": { + "content": "<|reserved_special_token_127|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128136": { + "content": "<|reserved_special_token_128|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128137": { + "content": "<|reserved_special_token_129|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128138": { + "content": "<|reserved_special_token_130|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128139": { + "content": "<|reserved_special_token_131|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128140": { + "content": "<|reserved_special_token_132|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128141": { + "content": "<|reserved_special_token_133|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128142": { + "content": "<|reserved_special_token_134|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128143": { + "content": "<|reserved_special_token_135|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128144": { + "content": "<|reserved_special_token_136|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128145": { + "content": "<|reserved_special_token_137|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128146": { + "content": "<|reserved_special_token_138|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128147": { + "content": "<|reserved_special_token_139|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128148": { + "content": "<|reserved_special_token_140|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128149": { + "content": "<|reserved_special_token_141|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128150": { + "content": "<|reserved_special_token_142|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128151": { + "content": "<|reserved_special_token_143|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128152": { + "content": "<|reserved_special_token_144|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128153": { + "content": "<|reserved_special_token_145|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128154": { + "content": "<|reserved_special_token_146|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128155": { + "content": "<|reserved_special_token_147|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128156": { + "content": "<|reserved_special_token_148|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128157": { + "content": "<|reserved_special_token_149|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128158": { + "content": "<|reserved_special_token_150|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128159": { + "content": "<|reserved_special_token_151|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128160": { + "content": "<|reserved_special_token_152|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128161": { + "content": "<|reserved_special_token_153|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128162": { + "content": "<|reserved_special_token_154|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128163": { + "content": "<|reserved_special_token_155|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128164": { + "content": "<|reserved_special_token_156|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128165": { + "content": "<|reserved_special_token_157|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128166": { + "content": "<|reserved_special_token_158|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128167": { + "content": "<|reserved_special_token_159|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128168": { + "content": "<|reserved_special_token_160|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128169": { + "content": "<|reserved_special_token_161|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128170": { + "content": "<|reserved_special_token_162|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128171": { + "content": "<|reserved_special_token_163|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128172": { + "content": "<|reserved_special_token_164|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128173": { + "content": "<|reserved_special_token_165|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128174": { + "content": "<|reserved_special_token_166|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128175": { + "content": "<|reserved_special_token_167|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128176": { + "content": "<|reserved_special_token_168|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128177": { + "content": "<|reserved_special_token_169|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128178": { + "content": "<|reserved_special_token_170|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128179": { + "content": "<|reserved_special_token_171|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128180": { + "content": "<|reserved_special_token_172|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128181": { + "content": "<|reserved_special_token_173|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128182": { + "content": "<|reserved_special_token_174|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128183": { + "content": "<|reserved_special_token_175|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128184": { + "content": "<|reserved_special_token_176|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128185": { + "content": "<|reserved_special_token_177|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128186": { + "content": "<|reserved_special_token_178|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128187": { + "content": "<|reserved_special_token_179|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128188": { + "content": "<|reserved_special_token_180|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128189": { + "content": "<|reserved_special_token_181|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128190": { + "content": "<|reserved_special_token_182|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128191": { + "content": "<|reserved_special_token_183|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128192": { + "content": "<|reserved_special_token_184|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128193": { + "content": "<|reserved_special_token_185|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128194": { + "content": "<|reserved_special_token_186|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128195": { + "content": "<|reserved_special_token_187|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128196": { + "content": "<|reserved_special_token_188|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128197": { + "content": "<|reserved_special_token_189|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128198": { + "content": "<|reserved_special_token_190|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128199": { + "content": "<|reserved_special_token_191|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128200": { + "content": "<|reserved_special_token_192|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128201": { + "content": "<|reserved_special_token_193|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128202": { + "content": "<|reserved_special_token_194|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128203": { + "content": "<|reserved_special_token_195|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128204": { + "content": "<|reserved_special_token_196|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128205": { + "content": "<|reserved_special_token_197|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128206": { + "content": "<|reserved_special_token_198|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128207": { + "content": "<|reserved_special_token_199|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128208": { + "content": "<|reserved_special_token_200|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128209": { + "content": "<|reserved_special_token_201|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128210": { + "content": "<|reserved_special_token_202|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128211": { + "content": "<|reserved_special_token_203|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128212": { + "content": "<|reserved_special_token_204|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128213": { + "content": "<|reserved_special_token_205|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128214": { + "content": "<|reserved_special_token_206|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128215": { + "content": "<|reserved_special_token_207|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128216": { + "content": "<|reserved_special_token_208|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128217": { + "content": "<|reserved_special_token_209|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128218": { + "content": "<|reserved_special_token_210|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128219": { + "content": "<|reserved_special_token_211|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128220": { + "content": "<|reserved_special_token_212|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128221": { + "content": "<|reserved_special_token_213|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128222": { + "content": "<|reserved_special_token_214|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128223": { + "content": "<|reserved_special_token_215|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128224": { + "content": "<|reserved_special_token_216|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128225": { + "content": "<|reserved_special_token_217|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128226": { + "content": "<|reserved_special_token_218|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128227": { + "content": "<|reserved_special_token_219|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128228": { + "content": "<|reserved_special_token_220|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128229": { + "content": "<|reserved_special_token_221|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128230": { + "content": "<|reserved_special_token_222|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128231": { + "content": "<|reserved_special_token_223|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128232": { + "content": "<|reserved_special_token_224|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128233": { + "content": "<|reserved_special_token_225|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128234": { + "content": "<|reserved_special_token_226|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128235": { + "content": "<|reserved_special_token_227|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128236": { + "content": "<|reserved_special_token_228|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128237": { + "content": "<|reserved_special_token_229|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128238": { + "content": "<|reserved_special_token_230|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128239": { + "content": "<|reserved_special_token_231|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128240": { + "content": "<|reserved_special_token_232|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128241": { + "content": "<|reserved_special_token_233|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128242": { + "content": "<|reserved_special_token_234|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128243": { + "content": "<|reserved_special_token_235|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128244": { + "content": "<|reserved_special_token_236|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128245": { + "content": "<|reserved_special_token_237|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128246": { + "content": "<|reserved_special_token_238|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128247": { + "content": "<|reserved_special_token_239|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128248": { + "content": "<|reserved_special_token_240|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128249": { + "content": "<|reserved_special_token_241|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128250": { + "content": "<|reserved_special_token_242|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128251": { + "content": "<|reserved_special_token_243|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128252": { + "content": "<|reserved_special_token_244|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128253": { + "content": "<|reserved_special_token_245|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128254": { + "content": "<|reserved_special_token_246|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + }, + "128255": { + "content": "<|reserved_special_token_247|>", + "lstrip": false, + "normalized": false, + "rstrip": false, + "single_word": false, + "special": true + } + }, + "bos_token": "<|begin_of_text|>", + "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", + "clean_up_tokenization_spaces": true, + "eos_token": "<|eot_id|>", + "extra_special_tokens": {}, + "model_input_names": [ + "input_ids", + "attention_mask" + ], + "model_max_length": 131072, + "pad_token": "<|eot_id|>", + "tokenizer_class": "PreTrainedTokenizer" +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/trainer_state.json b/not_grounded/0_128_e5_3e-5/checkpoint-771/trainer_state.json new file mode 100644 index 0000000000000000000000000000000000000000..9afdbcf4a6c5e52ff25288571cae3879eb315a02 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/trainer_state.json @@ -0,0 +1,1112 @@ +{ + "best_global_step": null, + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.0, + "eval_steps": 500, + "global_step": 771, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.01949317738791423, + "grad_norm": 0.6081938743591309, + "learning_rate": 1.875e-06, + "loss": 1.7943, + "step": 5 + }, + { + "epoch": 0.03898635477582846, + "grad_norm": 0.5703465938568115, + "learning_rate": 4.21875e-06, + "loss": 1.7931, + "step": 10 + }, + { + "epoch": 0.05847953216374269, + "grad_norm": 0.4802420139312744, + "learning_rate": 6.5625e-06, + "loss": 1.8677, + "step": 15 + }, + { + "epoch": 0.07797270955165692, + "grad_norm": 0.4527171552181244, + "learning_rate": 8.90625e-06, + "loss": 1.7498, + "step": 20 + }, + { + "epoch": 0.09746588693957114, + "grad_norm": 0.41348591446876526, + "learning_rate": 1.125e-05, + "loss": 1.7402, + "step": 25 + }, + { + "epoch": 0.11695906432748537, + "grad_norm": 0.44569578766822815, + "learning_rate": 1.359375e-05, + "loss": 1.6984, + "step": 30 + }, + { + "epoch": 0.1364522417153996, + "grad_norm": 0.439203143119812, + "learning_rate": 1.59375e-05, + "loss": 1.6548, + "step": 35 + }, + { + "epoch": 0.15594541910331383, + "grad_norm": 0.4098866879940033, + "learning_rate": 1.828125e-05, + "loss": 1.7276, + "step": 40 + }, + { + "epoch": 0.17543859649122806, + "grad_norm": 0.42010200023651123, + "learning_rate": 2.0625e-05, + "loss": 1.7037, + "step": 45 + }, + { + "epoch": 0.1949317738791423, + "grad_norm": 0.4497705399990082, + "learning_rate": 2.296875e-05, + "loss": 1.6829, + "step": 50 + }, + { + "epoch": 0.21442495126705652, + "grad_norm": 0.4349888563156128, + "learning_rate": 2.5312500000000002e-05, + "loss": 1.696, + "step": 55 + }, + { + "epoch": 0.23391812865497075, + "grad_norm": 0.45078787207603455, + "learning_rate": 2.765625e-05, + "loss": 1.6246, + "step": 60 + }, + { + "epoch": 0.253411306042885, + "grad_norm": 0.44585877656936646, + "learning_rate": 3e-05, + "loss": 1.6324, + "step": 65 + }, + { + "epoch": 0.2729044834307992, + "grad_norm": 0.4684946537017822, + "learning_rate": 2.9998748508718575e-05, + "loss": 1.5858, + "step": 70 + }, + { + "epoch": 0.29239766081871343, + "grad_norm": 0.5270784497261047, + "learning_rate": 2.9994994243705013e-05, + "loss": 1.5719, + "step": 75 + }, + { + "epoch": 0.31189083820662766, + "grad_norm": 0.5215119123458862, + "learning_rate": 2.9988737831416642e-05, + "loss": 1.5858, + "step": 80 + }, + { + "epoch": 0.3313840155945419, + "grad_norm": 0.5029568672180176, + "learning_rate": 2.9979980315832853e-05, + "loss": 1.4428, + "step": 85 + }, + { + "epoch": 0.3508771929824561, + "grad_norm": 0.5102682709693909, + "learning_rate": 2.9968723158280906e-05, + "loss": 1.5211, + "step": 90 + }, + { + "epoch": 0.37037037037037035, + "grad_norm": 0.5617806911468506, + "learning_rate": 2.995496823719206e-05, + "loss": 1.4377, + "step": 95 + }, + { + "epoch": 0.3898635477582846, + "grad_norm": 0.5236561298370361, + "learning_rate": 2.9938717847788167e-05, + "loss": 1.464, + "step": 100 + }, + { + "epoch": 0.4093567251461988, + "grad_norm": 0.6214772462844849, + "learning_rate": 2.9919974701698638e-05, + "loss": 1.4009, + "step": 105 + }, + { + "epoch": 0.42884990253411304, + "grad_norm": 0.5746711492538452, + "learning_rate": 2.989874192650801e-05, + "loss": 1.3927, + "step": 110 + }, + { + "epoch": 0.44834307992202727, + "grad_norm": 0.6362848877906799, + "learning_rate": 2.9875023065234003e-05, + "loss": 1.4212, + "step": 115 + }, + { + "epoch": 0.4678362573099415, + "grad_norm": 0.6465691328048706, + "learning_rate": 2.984882207573638e-05, + "loss": 1.4462, + "step": 120 + }, + { + "epoch": 0.4873294346978557, + "grad_norm": 0.6339296698570251, + "learning_rate": 2.982014333005645e-05, + "loss": 1.337, + "step": 125 + }, + { + "epoch": 0.50682261208577, + "grad_norm": 0.6140757203102112, + "learning_rate": 2.9788991613687575e-05, + "loss": 1.3289, + "step": 130 + }, + { + "epoch": 0.5263157894736842, + "grad_norm": 0.6720002293586731, + "learning_rate": 2.9755372124776616e-05, + "loss": 1.3323, + "step": 135 + }, + { + "epoch": 0.5458089668615984, + "grad_norm": 0.6524502038955688, + "learning_rate": 2.971929047325654e-05, + "loss": 1.3166, + "step": 140 + }, + { + "epoch": 0.5653021442495126, + "grad_norm": 0.7752301692962646, + "learning_rate": 2.968075267991032e-05, + "loss": 1.3543, + "step": 145 + }, + { + "epoch": 0.5847953216374269, + "grad_norm": 0.6975963711738586, + "learning_rate": 2.963976517536627e-05, + "loss": 1.2675, + "step": 150 + }, + { + "epoch": 0.6042884990253411, + "grad_norm": 0.7260017991065979, + "learning_rate": 2.9596334799025003e-05, + "loss": 1.3213, + "step": 155 + }, + { + "epoch": 0.6237816764132553, + "grad_norm": 0.8624387979507446, + "learning_rate": 2.9550468797918162e-05, + "loss": 1.2492, + "step": 160 + }, + { + "epoch": 0.6432748538011696, + "grad_norm": 0.7931470274925232, + "learning_rate": 2.950217482549915e-05, + "loss": 1.2353, + "step": 165 + }, + { + "epoch": 0.6627680311890838, + "grad_norm": 0.7223592400550842, + "learning_rate": 2.9451460940366025e-05, + "loss": 1.1868, + "step": 170 + }, + { + "epoch": 0.682261208576998, + "grad_norm": 0.7910372614860535, + "learning_rate": 2.9398335604916797e-05, + "loss": 1.2148, + "step": 175 + }, + { + "epoch": 0.7017543859649122, + "grad_norm": 0.851186990737915, + "learning_rate": 2.9342807683937352e-05, + "loss": 1.2595, + "step": 180 + }, + { + "epoch": 0.7212475633528265, + "grad_norm": 0.8052693009376526, + "learning_rate": 2.928488644312222e-05, + "loss": 1.2406, + "step": 185 + }, + { + "epoch": 0.7407407407407407, + "grad_norm": 0.8178960680961609, + "learning_rate": 2.9224581547528453e-05, + "loss": 1.2197, + "step": 190 + }, + { + "epoch": 0.7602339181286549, + "grad_norm": 0.8441738486289978, + "learning_rate": 2.916190305996286e-05, + "loss": 1.1726, + "step": 195 + }, + { + "epoch": 0.7797270955165692, + "grad_norm": 0.9862117767333984, + "learning_rate": 2.909686143930287e-05, + "loss": 1.1745, + "step": 200 + }, + { + "epoch": 0.7992202729044834, + "grad_norm": 0.8886749148368835, + "learning_rate": 2.902946753875131e-05, + "loss": 1.1447, + "step": 205 + }, + { + "epoch": 0.8187134502923976, + "grad_norm": 0.8914117813110352, + "learning_rate": 2.895973260402537e-05, + "loss": 1.1248, + "step": 210 + }, + { + "epoch": 0.8382066276803118, + "grad_norm": 1.1335160732269287, + "learning_rate": 2.8887668271480098e-05, + "loss": 1.1291, + "step": 215 + }, + { + "epoch": 0.8576998050682261, + "grad_norm": 1.0726529359817505, + "learning_rate": 2.8813286566166674e-05, + "loss": 1.1646, + "step": 220 + }, + { + "epoch": 0.8771929824561403, + "grad_norm": 1.0396595001220703, + "learning_rate": 2.873659989982586e-05, + "loss": 1.0894, + "step": 225 + }, + { + "epoch": 0.8966861598440545, + "grad_norm": 0.9956820011138916, + "learning_rate": 2.8657621068816903e-05, + "loss": 1.0984, + "step": 230 + }, + { + "epoch": 0.9161793372319688, + "grad_norm": 0.9483380317687988, + "learning_rate": 2.857636325198225e-05, + "loss": 1.0216, + "step": 235 + }, + { + "epoch": 0.935672514619883, + "grad_norm": 0.8558198809623718, + "learning_rate": 2.849284000844849e-05, + "loss": 1.105, + "step": 240 + }, + { + "epoch": 0.9551656920077972, + "grad_norm": 1.005858063697815, + "learning_rate": 2.8407065275363756e-05, + "loss": 1.0714, + "step": 245 + }, + { + "epoch": 0.9746588693957114, + "grad_norm": 0.9553005695343018, + "learning_rate": 2.8319053365572135e-05, + "loss": 1.0339, + "step": 250 + }, + { + "epoch": 0.9941520467836257, + "grad_norm": 0.9755716919898987, + "learning_rate": 2.8228818965225325e-05, + "loss": 1.0532, + "step": 255 + }, + { + "epoch": 1.0116959064327484, + "grad_norm": 1.0532433986663818, + "learning_rate": 2.8136377131332043e-05, + "loss": 1.0005, + "step": 260 + }, + { + "epoch": 1.0311890838206628, + "grad_norm": 0.9859170913696289, + "learning_rate": 2.8041743289245503e-05, + "loss": 0.9084, + "step": 265 + }, + { + "epoch": 1.050682261208577, + "grad_norm": 0.9481168389320374, + "learning_rate": 2.7944933230089484e-05, + "loss": 0.9431, + "step": 270 + }, + { + "epoch": 1.0701754385964912, + "grad_norm": 0.9678940773010254, + "learning_rate": 2.784596310812331e-05, + "loss": 0.9262, + "step": 275 + }, + { + "epoch": 1.0896686159844053, + "grad_norm": 0.9351955652236938, + "learning_rate": 2.774484943804629e-05, + "loss": 0.9044, + "step": 280 + }, + { + "epoch": 1.1091617933723197, + "grad_norm": 1.0261086225509644, + "learning_rate": 2.764160909224196e-05, + "loss": 0.9579, + "step": 285 + }, + { + "epoch": 1.128654970760234, + "grad_norm": 0.9796629548072815, + "learning_rate": 2.7536259297962674e-05, + "loss": 0.9283, + "step": 290 + }, + { + "epoch": 1.1481481481481481, + "grad_norm": 0.9690019488334656, + "learning_rate": 2.7428817634454973e-05, + "loss": 0.9017, + "step": 295 + }, + { + "epoch": 1.1676413255360623, + "grad_norm": 1.1676615476608276, + "learning_rate": 2.7319302030026207e-05, + "loss": 0.8919, + "step": 300 + }, + { + "epoch": 1.1871345029239766, + "grad_norm": 1.1467899084091187, + "learning_rate": 2.7207730759052925e-05, + "loss": 0.871, + "step": 305 + }, + { + "epoch": 1.206627680311891, + "grad_norm": 1.0315958261489868, + "learning_rate": 2.7094122438931513e-05, + "loss": 0.8388, + "step": 310 + }, + { + "epoch": 1.226120857699805, + "grad_norm": 1.0231361389160156, + "learning_rate": 2.697849602697159e-05, + "loss": 0.8841, + "step": 315 + }, + { + "epoch": 1.2456140350877192, + "grad_norm": 0.995548665523529, + "learning_rate": 2.6860870817232682e-05, + "loss": 0.8879, + "step": 320 + }, + { + "epoch": 1.2651072124756335, + "grad_norm": 1.102622151374817, + "learning_rate": 2.6741266437304718e-05, + "loss": 0.857, + "step": 325 + }, + { + "epoch": 1.2846003898635479, + "grad_norm": 1.3752007484436035, + "learning_rate": 2.661970284503286e-05, + "loss": 0.8027, + "step": 330 + }, + { + "epoch": 1.304093567251462, + "grad_norm": 1.129936933517456, + "learning_rate": 2.6496200325187223e-05, + "loss": 0.855, + "step": 335 + }, + { + "epoch": 1.323586744639376, + "grad_norm": 1.2070459127426147, + "learning_rate": 2.6370779486078047e-05, + "loss": 0.8544, + "step": 340 + }, + { + "epoch": 1.3430799220272904, + "grad_norm": 1.1400156021118164, + "learning_rate": 2.6243461256116892e-05, + "loss": 0.9026, + "step": 345 + }, + { + "epoch": 1.3625730994152048, + "grad_norm": 0.9957541823387146, + "learning_rate": 2.611426688032439e-05, + "loss": 0.8115, + "step": 350 + }, + { + "epoch": 1.3820662768031189, + "grad_norm": 1.0340908765792847, + "learning_rate": 2.598321791678519e-05, + "loss": 0.8314, + "step": 355 + }, + { + "epoch": 1.401559454191033, + "grad_norm": 1.1327755451202393, + "learning_rate": 2.5850336233050677e-05, + "loss": 0.8319, + "step": 360 + }, + { + "epoch": 1.4210526315789473, + "grad_norm": 1.0419026613235474, + "learning_rate": 2.5715644002489998e-05, + "loss": 0.8115, + "step": 365 + }, + { + "epoch": 1.4405458089668617, + "grad_norm": 1.080306053161621, + "learning_rate": 2.557916370059012e-05, + "loss": 0.7734, + "step": 370 + }, + { + "epoch": 1.4600389863547758, + "grad_norm": 1.1247351169586182, + "learning_rate": 2.544091810120543e-05, + "loss": 0.7968, + "step": 375 + }, + { + "epoch": 1.47953216374269, + "grad_norm": 1.1105760335922241, + "learning_rate": 2.530093027275757e-05, + "loss": 0.7994, + "step": 380 + }, + { + "epoch": 1.4990253411306043, + "grad_norm": 1.3396145105361938, + "learning_rate": 2.5159223574386117e-05, + "loss": 0.7891, + "step": 385 + }, + { + "epoch": 1.5185185185185186, + "grad_norm": 1.2513933181762695, + "learning_rate": 2.501582165205074e-05, + "loss": 0.7825, + "step": 390 + }, + { + "epoch": 1.5380116959064327, + "grad_norm": 1.139088749885559, + "learning_rate": 2.4870748434585514e-05, + "loss": 0.7875, + "step": 395 + }, + { + "epoch": 1.5575048732943468, + "grad_norm": 1.2570606470108032, + "learning_rate": 2.4724028129706012e-05, + "loss": 0.7524, + "step": 400 + }, + { + "epoch": 1.5769980506822612, + "grad_norm": 1.115220069885254, + "learning_rate": 2.4575685219969884e-05, + "loss": 0.7855, + "step": 405 + }, + { + "epoch": 1.5964912280701755, + "grad_norm": 1.2705992460250854, + "learning_rate": 2.442574445869156e-05, + "loss": 0.7338, + "step": 410 + }, + { + "epoch": 1.6159844054580896, + "grad_norm": 1.1325453519821167, + "learning_rate": 2.4274230865811763e-05, + "loss": 0.7528, + "step": 415 + }, + { + "epoch": 1.6354775828460038, + "grad_norm": 1.301644206047058, + "learning_rate": 2.4121169723722566e-05, + "loss": 0.7445, + "step": 420 + }, + { + "epoch": 1.654970760233918, + "grad_norm": 1.4577105045318604, + "learning_rate": 2.396658657304861e-05, + "loss": 0.7335, + "step": 425 + }, + { + "epoch": 1.6744639376218324, + "grad_norm": 1.1624274253845215, + "learning_rate": 2.381050720838528e-05, + "loss": 0.7146, + "step": 430 + }, + { + "epoch": 1.6939571150097466, + "grad_norm": 1.4543778896331787, + "learning_rate": 2.3652957673994448e-05, + "loss": 0.7338, + "step": 435 + }, + { + "epoch": 1.7134502923976607, + "grad_norm": 1.208083987236023, + "learning_rate": 2.3493964259458603e-05, + "loss": 0.7187, + "step": 440 + }, + { + "epoch": 1.732943469785575, + "grad_norm": 1.1645222902297974, + "learning_rate": 2.3333553495294033e-05, + "loss": 0.7325, + "step": 445 + }, + { + "epoch": 1.7524366471734893, + "grad_norm": 1.2736948728561401, + "learning_rate": 2.317175214852377e-05, + "loss": 0.7152, + "step": 450 + }, + { + "epoch": 1.7719298245614035, + "grad_norm": 1.2995411157608032, + "learning_rate": 2.3008587218211127e-05, + "loss": 0.6953, + "step": 455 + }, + { + "epoch": 1.7914230019493176, + "grad_norm": 1.4111621379852295, + "learning_rate": 2.284408593095446e-05, + "loss": 0.7023, + "step": 460 + }, + { + "epoch": 1.810916179337232, + "grad_norm": 1.2578918933868408, + "learning_rate": 2.2678275736344014e-05, + "loss": 0.6771, + "step": 465 + }, + { + "epoch": 1.8304093567251463, + "grad_norm": 1.3310747146606445, + "learning_rate": 2.251118430238151e-05, + "loss": 0.6792, + "step": 470 + }, + { + "epoch": 1.8499025341130604, + "grad_norm": 1.1377298831939697, + "learning_rate": 2.2342839510863323e-05, + "loss": 0.7322, + "step": 475 + }, + { + "epoch": 1.8693957115009745, + "grad_norm": 1.2897000312805176, + "learning_rate": 2.2173269452727965e-05, + "loss": 0.6468, + "step": 480 + }, + { + "epoch": 1.8888888888888888, + "grad_norm": 1.1680046319961548, + "learning_rate": 2.200250242336868e-05, + "loss": 0.6785, + "step": 485 + }, + { + "epoch": 1.9083820662768032, + "grad_norm": 1.2389404773712158, + "learning_rate": 2.183056691791193e-05, + "loss": 0.6521, + "step": 490 + }, + { + "epoch": 1.9278752436647173, + "grad_norm": 1.1207393407821655, + "learning_rate": 2.1657491626462514e-05, + "loss": 0.6607, + "step": 495 + }, + { + "epoch": 1.9473684210526314, + "grad_norm": 1.1037189960479736, + "learning_rate": 2.1483305429316208e-05, + "loss": 0.6033, + "step": 500 + }, + { + "epoch": 1.9668615984405458, + "grad_norm": 1.1787821054458618, + "learning_rate": 2.1308037392140613e-05, + "loss": 0.6254, + "step": 505 + }, + { + "epoch": 1.98635477582846, + "grad_norm": 1.266696572303772, + "learning_rate": 2.113171676112513e-05, + "loss": 0.5962, + "step": 510 + }, + { + "epoch": 2.003898635477583, + "grad_norm": 1.4790624380111694, + "learning_rate": 2.0954372958100746e-05, + "loss": 0.6104, + "step": 515 + }, + { + "epoch": 2.023391812865497, + "grad_norm": 1.6624841690063477, + "learning_rate": 2.0776035575630575e-05, + "loss": 0.5632, + "step": 520 + }, + { + "epoch": 2.042884990253411, + "grad_norm": 1.243015170097351, + "learning_rate": 2.0596734372071855e-05, + "loss": 0.5019, + "step": 525 + }, + { + "epoch": 2.0623781676413255, + "grad_norm": 1.4598456621170044, + "learning_rate": 2.0416499266610316e-05, + "loss": 0.5299, + "step": 530 + }, + { + "epoch": 2.08187134502924, + "grad_norm": 1.2799503803253174, + "learning_rate": 2.023536033426771e-05, + "loss": 0.5432, + "step": 535 + }, + { + "epoch": 2.101364522417154, + "grad_norm": 1.3306210041046143, + "learning_rate": 2.00533478008833e-05, + "loss": 0.5639, + "step": 540 + }, + { + "epoch": 2.120857699805068, + "grad_norm": 1.5138165950775146, + "learning_rate": 1.9870492038070255e-05, + "loss": 0.519, + "step": 545 + }, + { + "epoch": 2.1403508771929824, + "grad_norm": 1.283997893333435, + "learning_rate": 1.9686823558147616e-05, + "loss": 0.5276, + "step": 550 + }, + { + "epoch": 2.159844054580897, + "grad_norm": 1.2089860439300537, + "learning_rate": 1.9502373009048898e-05, + "loss": 0.5384, + "step": 555 + }, + { + "epoch": 2.1793372319688107, + "grad_norm": 1.327548861503601, + "learning_rate": 1.9317171169207973e-05, + "loss": 0.5297, + "step": 560 + }, + { + "epoch": 2.198830409356725, + "grad_norm": 1.3738446235656738, + "learning_rate": 1.913124894242322e-05, + "loss": 0.5324, + "step": 565 + }, + { + "epoch": 2.2183235867446394, + "grad_norm": 1.2523514032363892, + "learning_rate": 1.8944637352700754e-05, + "loss": 0.4629, + "step": 570 + }, + { + "epoch": 2.2378167641325537, + "grad_norm": 1.4739748239517212, + "learning_rate": 1.875736753907758e-05, + "loss": 0.487, + "step": 575 + }, + { + "epoch": 2.257309941520468, + "grad_norm": 1.3324633836746216, + "learning_rate": 1.856947075042557e-05, + "loss": 0.4968, + "step": 580 + }, + { + "epoch": 2.276803118908382, + "grad_norm": 1.3165260553359985, + "learning_rate": 1.8380978340237094e-05, + "loss": 0.5223, + "step": 585 + }, + { + "epoch": 2.2962962962962963, + "grad_norm": 1.3404366970062256, + "learning_rate": 1.8191921761393222e-05, + "loss": 0.4838, + "step": 590 + }, + { + "epoch": 2.3157894736842106, + "grad_norm": 1.3298143148422241, + "learning_rate": 1.8002332560915297e-05, + "loss": 0.4742, + "step": 595 + }, + { + "epoch": 2.3352826510721245, + "grad_norm": 1.2816650867462158, + "learning_rate": 1.7812242374700847e-05, + "loss": 0.4775, + "step": 600 + }, + { + "epoch": 2.354775828460039, + "grad_norm": 1.406812310218811, + "learning_rate": 1.7621682922244637e-05, + "loss": 0.5127, + "step": 605 + }, + { + "epoch": 2.374269005847953, + "grad_norm": 1.3161938190460205, + "learning_rate": 1.7430686001345777e-05, + "loss": 0.4723, + "step": 610 + }, + { + "epoch": 2.3937621832358675, + "grad_norm": 1.4261293411254883, + "learning_rate": 1.7239283482801783e-05, + "loss": 0.5121, + "step": 615 + }, + { + "epoch": 2.413255360623782, + "grad_norm": 1.2458994388580322, + "learning_rate": 1.7047507305090397e-05, + "loss": 0.4821, + "step": 620 + }, + { + "epoch": 2.4327485380116958, + "grad_norm": 1.3558402061462402, + "learning_rate": 1.6855389469040217e-05, + "loss": 0.4423, + "step": 625 + }, + { + "epoch": 2.45224171539961, + "grad_norm": 1.2846639156341553, + "learning_rate": 1.6662962032490814e-05, + "loss": 0.4964, + "step": 630 + }, + { + "epoch": 2.4717348927875245, + "grad_norm": 1.3657177686691284, + "learning_rate": 1.6470257104943414e-05, + "loss": 0.4906, + "step": 635 + }, + { + "epoch": 2.4912280701754383, + "grad_norm": 1.3313744068145752, + "learning_rate": 1.6277306842202902e-05, + "loss": 0.4824, + "step": 640 + }, + { + "epoch": 2.5107212475633527, + "grad_norm": 1.427259922027588, + "learning_rate": 1.6084143441012158e-05, + "loss": 0.5067, + "step": 645 + }, + { + "epoch": 2.530214424951267, + "grad_norm": 1.3357858657836914, + "learning_rate": 1.5890799133679523e-05, + "loss": 0.5128, + "step": 650 + }, + { + "epoch": 2.5497076023391814, + "grad_norm": 1.6301915645599365, + "learning_rate": 1.569730618270031e-05, + "loss": 0.4331, + "step": 655 + }, + { + "epoch": 2.5692007797270957, + "grad_norm": 1.4900562763214111, + "learning_rate": 1.5503696875373353e-05, + "loss": 0.4391, + "step": 660 + }, + { + "epoch": 2.5886939571150096, + "grad_norm": 1.402565836906433, + "learning_rate": 1.5310003518413318e-05, + "loss": 0.4457, + "step": 665 + }, + { + "epoch": 2.608187134502924, + "grad_norm": 1.3605144023895264, + "learning_rate": 1.5116258432559883e-05, + "loss": 0.4546, + "step": 670 + }, + { + "epoch": 2.6276803118908383, + "grad_norm": 1.3472214937210083, + "learning_rate": 1.4922493947184482e-05, + "loss": 0.4791, + "step": 675 + }, + { + "epoch": 2.647173489278752, + "grad_norm": 1.4113649129867554, + "learning_rate": 1.4728742394895666e-05, + "loss": 0.4364, + "step": 680 + }, + { + "epoch": 2.6666666666666665, + "grad_norm": 1.3126051425933838, + "learning_rate": 1.4535036106143895e-05, + "loss": 0.4309, + "step": 685 + }, + { + "epoch": 2.686159844054581, + "grad_norm": 1.4124526977539062, + "learning_rate": 1.4341407403826695e-05, + "loss": 0.4346, + "step": 690 + }, + { + "epoch": 2.705653021442495, + "grad_norm": 1.4493342638015747, + "learning_rate": 1.4147888597895117e-05, + "loss": 0.4175, + "step": 695 + }, + { + "epoch": 2.7251461988304095, + "grad_norm": 1.4753097295761108, + "learning_rate": 1.3954511979962271e-05, + "loss": 0.4275, + "step": 700 + }, + { + "epoch": 2.7446393762183234, + "grad_norm": 1.2626641988754272, + "learning_rate": 1.3761309817915017e-05, + "loss": 0.4182, + "step": 705 + }, + { + "epoch": 2.7641325536062378, + "grad_norm": 1.526242971420288, + "learning_rate": 1.3568314350529534e-05, + "loss": 0.3954, + "step": 710 + }, + { + "epoch": 2.783625730994152, + "grad_norm": 1.5078595876693726, + "learning_rate": 1.3375557782091789e-05, + "loss": 0.4406, + "step": 715 + }, + { + "epoch": 2.803118908382066, + "grad_norm": 1.398832082748413, + "learning_rate": 1.3183072277023766e-05, + "loss": 0.4096, + "step": 720 + }, + { + "epoch": 2.8226120857699804, + "grad_norm": 1.4593217372894287, + "learning_rate": 1.2990889954516313e-05, + "loss": 0.4015, + "step": 725 + }, + { + "epoch": 2.8421052631578947, + "grad_norm": 1.5111234188079834, + "learning_rate": 1.2799042883169576e-05, + "loss": 0.4074, + "step": 730 + }, + { + "epoch": 2.861598440545809, + "grad_norm": 1.4062628746032715, + "learning_rate": 1.2607563075641838e-05, + "loss": 0.4055, + "step": 735 + }, + { + "epoch": 2.8810916179337234, + "grad_norm": 1.3948218822479248, + "learning_rate": 1.2416482483307728e-05, + "loss": 0.4129, + "step": 740 + }, + { + "epoch": 2.9005847953216373, + "grad_norm": 1.408527135848999, + "learning_rate": 1.2225832990926625e-05, + "loss": 0.4073, + "step": 745 + }, + { + "epoch": 2.9200779727095516, + "grad_norm": 1.4006417989730835, + "learning_rate": 1.20356464113222e-05, + "loss": 0.4147, + "step": 750 + }, + { + "epoch": 2.939571150097466, + "grad_norm": 1.3896782398223877, + "learning_rate": 1.184595448007395e-05, + "loss": 0.3975, + "step": 755 + }, + { + "epoch": 2.95906432748538, + "grad_norm": 1.4923348426818848, + "learning_rate": 1.1656788850221617e-05, + "loss": 0.3933, + "step": 760 + }, + { + "epoch": 2.978557504873294, + "grad_norm": 1.3745965957641602, + "learning_rate": 1.1468181086983414e-05, + "loss": 0.395, + "step": 765 + }, + { + "epoch": 2.9980506822612085, + "grad_norm": 1.429971694946289, + "learning_rate": 1.1280162662488834e-05, + "loss": 0.4257, + "step": 770 + } + ], + "logging_steps": 5, + "max_steps": 1280, + "num_input_tokens_seen": 0, + "num_train_epochs": 5, + "save_steps": 20000, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": false + }, + "attributes": {} + } + }, + "total_flos": 1.3359030163514327e+18, + "train_batch_size": 2, + "trial_name": null, + "trial_params": null +} diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/training_args.bin b/not_grounded/0_128_e5_3e-5/checkpoint-771/training_args.bin new file mode 100644 index 0000000000000000000000000000000000000000..4ce5ba5a4e8b2f8e0972e0e67dc1879011fda1a2 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/training_args.bin @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:aa02dca237b948f004cb8d3e65da8c4d5a5b94ef5684cb38197d8ffa2ad8f171 +size 7800 diff --git a/not_grounded/0_128_e5_3e-5/checkpoint-771/zero_to_fp32.py b/not_grounded/0_128_e5_3e-5/checkpoint-771/zero_to_fp32.py new file mode 100644 index 0000000000000000000000000000000000000000..24cc342e78d1a006c782b3a4cd68d9ce786d8fd8 --- /dev/null +++ b/not_grounded/0_128_e5_3e-5/checkpoint-771/zero_to_fp32.py @@ -0,0 +1,604 @@ +#!/usr/bin/env python + +# Copyright (c) Microsoft Corporation. +# SPDX-License-Identifier: Apache-2.0 + +# DeepSpeed Team + +# This script extracts fp32 consolidated weights from a zero 1, 2 and 3 DeepSpeed checkpoints. It gets +# copied into the top level checkpoint dir, so the user can easily do the conversion at any point in +# the future. Once extracted, the weights don't require DeepSpeed and can be used in any +# application. +# +# example: python zero_to_fp32.py . pytorch_model.bin + +import argparse +import torch +import glob +import math +import os +import re +from collections import OrderedDict +from dataclasses import dataclass + +# while this script doesn't use deepspeed to recover data, since the checkpoints are pickled with +# DeepSpeed data structures it has to be available in the current python environment. +from deepspeed.utils import logger +from deepspeed.checkpoint.constants import (DS_VERSION, OPTIMIZER_STATE_DICT, SINGLE_PARTITION_OF_FP32_GROUPS, + FP32_FLAT_GROUPS, ZERO_STAGE, PARTITION_COUNT, PARAM_SHAPES, BUFFER_NAMES, + FROZEN_PARAM_SHAPES, FROZEN_PARAM_FRAGMENTS) + + +@dataclass +class zero_model_state: + buffers: dict() + param_shapes: dict() + shared_params: list + ds_version: int + frozen_param_shapes: dict() + frozen_param_fragments: dict() + + +debug = 0 + +# load to cpu +device = torch.device('cpu') + + +def atoi(text): + return int(text) if text.isdigit() else text + + +def natural_keys(text): + ''' + alist.sort(key=natural_keys) sorts in human order + http://nedbatchelder.com/blog/200712/human_sorting.html + (See Toothy's implementation in the comments) + ''' + return [atoi(c) for c in re.split(r'(\d+)', text)] + + +def get_model_state_file(checkpoint_dir, zero_stage): + if not os.path.isdir(checkpoint_dir): + raise FileNotFoundError(f"Directory '{checkpoint_dir}' doesn't exist") + + # there should be only one file + if zero_stage <= 2: + file = os.path.join(checkpoint_dir, "mp_rank_00_model_states.pt") + elif zero_stage == 3: + file = os.path.join(checkpoint_dir, "zero_pp_rank_0_mp_rank_00_model_states.pt") + + if not os.path.exists(file): + raise FileNotFoundError(f"can't find model states file at '{file}'") + + return file + + +def get_checkpoint_files(checkpoint_dir, glob_pattern): + # XXX: need to test that this simple glob rule works for multi-node setup too + ckpt_files = sorted(glob.glob(os.path.join(checkpoint_dir, glob_pattern)), key=natural_keys) + + if len(ckpt_files) == 0: + raise FileNotFoundError(f"can't find {glob_pattern} files in directory '{checkpoint_dir}'") + + return ckpt_files + + +def get_optim_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_optim_states.pt") + + +def get_model_state_files(checkpoint_dir): + return get_checkpoint_files(checkpoint_dir, "*_model_states.pt") + + +def parse_model_states(files): + zero_model_states = [] + for file in files: + state_dict = torch.load(file, map_location=device) + + if BUFFER_NAMES not in state_dict: + raise ValueError(f"{file} is not a model state checkpoint") + buffer_names = state_dict[BUFFER_NAMES] + if debug: + print("Found buffers:", buffer_names) + + # recover just the buffers while restoring them to fp32 if they were saved in fp16 + buffers = {k: v.float() for k, v in state_dict["module"].items() if k in buffer_names} + param_shapes = state_dict[PARAM_SHAPES] + + # collect parameters that are included in param_shapes + param_names = [] + for s in param_shapes: + for name in s.keys(): + param_names.append(name) + + # update with frozen parameters + frozen_param_shapes = state_dict.get(FROZEN_PARAM_SHAPES, None) + if frozen_param_shapes is not None: + if debug: + print(f"Found frozen_param_shapes: {frozen_param_shapes}") + param_names += list(frozen_param_shapes.keys()) + + # handle shared params + shared_params = [[k, v] for k, v in state_dict["shared_params"].items()] + + ds_version = state_dict.get(DS_VERSION, None) + + frozen_param_fragments = state_dict.get(FROZEN_PARAM_FRAGMENTS, None) + + z_model_state = zero_model_state(buffers=buffers, + param_shapes=param_shapes, + shared_params=shared_params, + ds_version=ds_version, + frozen_param_shapes=frozen_param_shapes, + frozen_param_fragments=frozen_param_fragments) + zero_model_states.append(z_model_state) + + return zero_model_states + + +def parse_optim_states(files, ds_checkpoint_dir): + + total_files = len(files) + state_dicts = [] + for f in files: + state_dict = torch.load(f, map_location=device) + # immediately discard the potentially huge 2 optimizer states as we only care for fp32 master weights + # and also handle the case where it was already removed by another helper script + state_dict["optimizer_state_dict"].pop("optimizer_state_dict", None) + state_dicts.append(state_dict) + + if not ZERO_STAGE in state_dicts[0][OPTIMIZER_STATE_DICT]: + raise ValueError(f"{files[0]} is not a zero checkpoint") + zero_stage = state_dicts[0][OPTIMIZER_STATE_DICT][ZERO_STAGE] + world_size = state_dicts[0][OPTIMIZER_STATE_DICT][PARTITION_COUNT] + + # For ZeRO-2 each param group can have different partition_count as data parallelism for expert + # parameters can be different from data parallelism for non-expert parameters. So we can just + # use the max of the partition_count to get the dp world_size. + + if type(world_size) is list: + world_size = max(world_size) + + if world_size != total_files: + raise ValueError( + f"Expected {world_size} of '*_optim_states.pt' under '{ds_checkpoint_dir}' but found {total_files} files. " + "Possibly due to an overwrite of an old checkpoint, or a checkpoint didn't get saved by one or more processes." + ) + + # the groups are named differently in each stage + if zero_stage <= 2: + fp32_groups_key = SINGLE_PARTITION_OF_FP32_GROUPS + elif zero_stage == 3: + fp32_groups_key = FP32_FLAT_GROUPS + else: + raise ValueError(f"unknown zero stage {zero_stage}") + + if zero_stage <= 2: + fp32_flat_groups = [state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key] for i in range(len(state_dicts))] + elif zero_stage == 3: + # if there is more than one param group, there will be multiple flattened tensors - one + # flattened tensor per group - for simplicity merge them into a single tensor + # + # XXX: could make the script more memory efficient for when there are multiple groups - it + # will require matching the sub-lists of param_shapes for each param group flattened tensor + + fp32_flat_groups = [ + torch.cat(state_dicts[i][OPTIMIZER_STATE_DICT][fp32_groups_key], 0) for i in range(len(state_dicts)) + ] + + return zero_stage, world_size, fp32_flat_groups + + +def _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters): + """ + Returns fp32 state_dict reconstructed from ds checkpoint + + Args: + - ``ds_checkpoint_dir``: path to the deepspeed checkpoint folder (where the optimizer files are) + + """ + print(f"Processing zero checkpoint '{ds_checkpoint_dir}'") + + optim_files = get_optim_files(ds_checkpoint_dir) + zero_stage, world_size, fp32_flat_groups = parse_optim_states(optim_files, ds_checkpoint_dir) + print(f"Detected checkpoint of type zero stage {zero_stage}, world_size: {world_size}") + + model_files = get_model_state_files(ds_checkpoint_dir) + + zero_model_states = parse_model_states(model_files) + print(f'Parsing checkpoint created by deepspeed=={zero_model_states[0].ds_version}') + + if zero_stage <= 2: + return _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + elif zero_stage == 3: + return _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters) + + +def _zero2_merge_frozen_params(state_dict, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + frozen_param_fragments = zero_model_states[0].frozen_param_fragments + + if debug: + num_elem = sum(s.numel() for s in frozen_param_shapes.values()) + print(f'rank 0: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in frozen_param_fragments.values()]) + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + state_dict[name] = frozen_param_fragments[name] + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _has_callable(obj, fn): + attr = getattr(obj, fn, None) + return callable(attr) + + +def _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + + # Reconstruction protocol: + # + # XXX: document this + + if debug: + for i in range(world_size): + for j in range(len(fp32_flat_groups[0])): + print(f"{FP32_FLAT_GROUPS}[{i}][{j}].shape={fp32_flat_groups[i][j].shape}") + + # XXX: memory usage doubles here (zero2) + num_param_groups = len(fp32_flat_groups[0]) + merged_single_partition_of_fp32_groups = [] + for i in range(num_param_groups): + merged_partitions = [sd[i] for sd in fp32_flat_groups] + full_single_fp32_vector = torch.cat(merged_partitions, 0) + merged_single_partition_of_fp32_groups.append(full_single_fp32_vector) + avail_numel = sum( + [full_single_fp32_vector.numel() for full_single_fp32_vector in merged_single_partition_of_fp32_groups]) + + if debug: + wanted_params = sum([len(shapes) for shapes in param_shapes]) + wanted_numel = sum([sum(shape.numel() for shape in shapes.values()) for shapes in param_shapes]) + # not asserting if there is a mismatch due to possible padding + print(f"Have {avail_numel} numels to process.") + print(f"Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + total_numel = 0 + total_params = 0 + for shapes, full_single_fp32_vector in zip(param_shapes, merged_single_partition_of_fp32_groups): + offset = 0 + avail_numel = full_single_fp32_vector.numel() + for name, shape in shapes.items(): + + unpartitioned_numel = shape.numel() if _has_callable(shape, 'numel') else math.prod(shape) + total_numel += unpartitioned_numel + total_params += 1 + + if debug: + print(f"{name} full shape: {shape} unpartitioned numel {unpartitioned_numel} ") + state_dict[name] = full_single_fp32_vector.narrow(0, offset, unpartitioned_numel).view(shape) + offset += unpartitioned_numel + + # Z2 started to align to 2*world_size to improve nccl performance. Therefore both offset and + # avail_numel can differ by anywhere between 0..2*world_size. Due to two unrelated complex + # paddings performed in the code it's almost impossible to predict the exact numbers w/o the + # live optimizer object, so we are checking that the numbers are within the right range + align_to = 2 * world_size + + def zero2_align(x): + return align_to * math.ceil(x / align_to) + + if debug: + print(f"original offset={offset}, avail_numel={avail_numel}") + + offset = zero2_align(offset) + avail_numel = zero2_align(avail_numel) + + if debug: + print(f"aligned offset={offset}, avail_numel={avail_numel}") + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero2_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero2_merge_frozen_params(state_dict, zero_model_states) + + _zero2_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def zero3_partitioned_param_info(unpartitioned_numel, world_size): + remainder = unpartitioned_numel % world_size + padding_numel = (world_size - remainder) if remainder else 0 + partitioned_numel = math.ceil(unpartitioned_numel / world_size) + return partitioned_numel, padding_numel + + +def _zero3_merge_frozen_params(state_dict, world_size, zero_model_states): + if zero_model_states[0].frozen_param_shapes is None or len(zero_model_states[0].frozen_param_shapes) == 0: + return + + if debug: + for i in range(world_size): + num_elem = sum(s.numel() for s in zero_model_states[i].frozen_param_fragments.values()) + print(f'rank {i}: {FROZEN_PARAM_SHAPES}.numel = {num_elem}') + + frozen_param_shapes = zero_model_states[0].frozen_param_shapes + wanted_params = len(frozen_param_shapes) + wanted_numel = sum(s.numel() for s in frozen_param_shapes.values()) + avail_numel = sum([p.numel() for p in zero_model_states[0].frozen_param_fragments.values()]) * world_size + print(f'Frozen params: Have {avail_numel} numels to process.') + print(f'Frozen params: Need {wanted_numel} numels in {wanted_params} params') + + total_params = 0 + total_numel = 0 + for name, shape in zero_model_states[0].frozen_param_shapes.items(): + total_params += 1 + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + + param_frags = tuple(model_state.frozen_param_fragments[name] for model_state in zero_model_states) + state_dict[name] = torch.cat(param_frags, 0).narrow(0, 0, unpartitioned_numel).view(shape) + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Frozen params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + print(f"Reconstructed Frozen fp32 state dict with {total_params} params {total_numel} elements") + + +def _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states): + param_shapes = zero_model_states[0].param_shapes + avail_numel = fp32_flat_groups[0].numel() * world_size + # Reconstruction protocol: For zero3 we need to zip the partitions together at boundary of each + # param, re-consolidating each param, while dealing with padding if any + + # merge list of dicts, preserving order + param_shapes = {k: v for d in param_shapes for k, v in d.items()} + + if debug: + for i in range(world_size): + print(f"{FP32_FLAT_GROUPS}[{i}].shape={fp32_flat_groups[i].shape}") + + wanted_params = len(param_shapes) + wanted_numel = sum(shape.numel() for shape in param_shapes.values()) + # not asserting if there is a mismatch due to possible padding + avail_numel = fp32_flat_groups[0].numel() * world_size + print(f"Trainable params: Have {avail_numel} numels to process.") + print(f"Trainable params: Need {wanted_numel} numels in {wanted_params} params.") + + # params + # XXX: for huge models that can't fit into the host's RAM we will have to recode this to support + # out-of-core computing solution + offset = 0 + total_numel = 0 + total_params = 0 + for name, shape in param_shapes.items(): + + unpartitioned_numel = shape.numel() + total_numel += unpartitioned_numel + total_params += 1 + + partitioned_numel, partitioned_padding_numel = zero3_partitioned_param_info(unpartitioned_numel, world_size) + + if debug: + print( + f"Trainable params: {total_params} {name} full shape: {shape} partition0 numel={partitioned_numel} partitioned_padding_numel={partitioned_padding_numel}" + ) + + # XXX: memory usage doubles here + state_dict[name] = torch.cat( + tuple(fp32_flat_groups[i].narrow(0, offset, partitioned_numel) for i in range(world_size)), + 0).narrow(0, 0, unpartitioned_numel).view(shape) + offset += partitioned_numel + + offset *= world_size + + # Sanity check + if offset != avail_numel: + raise ValueError(f"consumed {offset} numels out of {avail_numel} - something is wrong") + + print(f"Reconstructed Trainable fp32 state dict with {total_params} params {total_numel} elements") + + +def _get_fp32_state_dict_from_zero3_checkpoint(world_size, fp32_flat_groups, zero_model_states, + exclude_frozen_parameters): + state_dict = OrderedDict() + + # buffers + buffers = zero_model_states[0].buffers + state_dict.update(buffers) + if debug: + print(f"added {len(buffers)} buffers") + + if not exclude_frozen_parameters: + _zero3_merge_frozen_params(state_dict, world_size, zero_model_states) + + _zero3_merge_trainable_params(state_dict, world_size, fp32_flat_groups, zero_model_states) + + # recover shared parameters + for pair in zero_model_states[0].shared_params: + if pair[1] in state_dict: + state_dict[pair[0]] = state_dict[pair[1]] + + return state_dict + + +def get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated state_dict that can be loaded with + ``load_state_dict()`` and used for training without DeepSpeed or shared with others, for example + via a model hub. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in 'latest' file. e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + + Returns: + - pytorch ``state_dict`` + + Note: this approach may not work if your application doesn't have sufficient free CPU memory and + you may need to use the offline approach using the ``zero_to_fp32.py`` script that is saved with + the checkpoint. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import get_fp32_state_dict_from_zero_checkpoint + # do the training and checkpoint saving + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir) # already on cpu + model = model.cpu() # move to cpu + model.load_state_dict(state_dict) + # submit to model hub or save the model to share with others + + In this example the ``model`` will no longer be usable in the deepspeed context of the same + application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + If you want it all done for you, use ``load_state_dict_from_zero_checkpoint`` instead. + + """ + if tag is None: + latest_path = os.path.join(checkpoint_dir, 'latest') + if os.path.isfile(latest_path): + with open(latest_path, 'r') as fd: + tag = fd.read().strip() + else: + raise ValueError(f"Unable to find 'latest' file at {latest_path}") + + ds_checkpoint_dir = os.path.join(checkpoint_dir, tag) + + if not os.path.isdir(ds_checkpoint_dir): + raise FileNotFoundError(f"Directory '{ds_checkpoint_dir}' doesn't exist") + + return _get_fp32_state_dict_from_zero_checkpoint(ds_checkpoint_dir, exclude_frozen_parameters) + + +def convert_zero_checkpoint_to_fp32_state_dict(checkpoint_dir, output_file, tag=None, exclude_frozen_parameters=False): + """ + Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` file that can be + loaded with ``torch.load(file)`` + ``load_state_dict()`` and used for training without DeepSpeed. + + Args: + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``output_file``: path to the pytorch fp32 state_dict output file (e.g. path/pytorch_model.bin) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + - ``exclude_frozen_parameters``: exclude frozen parameters + """ + + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag, exclude_frozen_parameters) + print(f"Saving fp32 state dict to {output_file}") + torch.save(state_dict, output_file) + + +def load_state_dict_from_zero_checkpoint(model, checkpoint_dir, tag=None): + """ + 1. Put the provided model to cpu + 2. Convert ZeRO 2 or 3 checkpoint into a single fp32 consolidated ``state_dict`` + 3. Load it into the provided model + + Args: + - ``model``: the model object to update + - ``checkpoint_dir``: path to the desired checkpoint folder. (one that contains the tag-folder, like ``global_step14``) + - ``tag``: checkpoint tag used as a unique identifier for checkpoint. If not provided will attempt to load tag in the file named ``latest`` in the checkpoint folder, e.g., ``global_step14`` + + Returns: + - ``model`: modified model + + Make sure you have plenty of CPU memory available before you call this function. If you don't + have enough use the ``zero_to_fp32.py`` utility to do the conversion. You will find it + conveniently placed for you in the checkpoint folder. + + A typical usage might be :: + + from deepspeed.utils.zero_to_fp32 import load_state_dict_from_zero_checkpoint + model = load_state_dict_from_zero_checkpoint(trainer.model, checkpoint_dir) + # submit to model hub or save the model to share with others + + Note, that once this was run, the ``model`` will no longer be usable in the deepspeed context + of the same application. i.e. you will need to re-initialize the deepspeed engine, since + ``model.load_state_dict(state_dict)`` will remove all the deepspeed magic from it. + + """ + logger.info(f"Extracting fp32 weights") + state_dict = get_fp32_state_dict_from_zero_checkpoint(checkpoint_dir, tag) + + logger.info(f"Overwriting model with fp32 weights") + model = model.cpu() + model.load_state_dict(state_dict, strict=False) + + return model + + +if __name__ == "__main__": + + parser = argparse.ArgumentParser() + parser.add_argument("checkpoint_dir", + type=str, + help="path to the desired checkpoint folder, e.g., path/checkpoint-12") + parser.add_argument( + "output_file", + type=str, + help="path to the pytorch fp32 state_dict output file (e.g. path/checkpoint-12/pytorch_model.bin)") + parser.add_argument("-t", + "--tag", + type=str, + default=None, + help="checkpoint tag used as a unique identifier for checkpoint. e.g., global_step1") + parser.add_argument("--exclude_frozen_parameters", action='store_true', help="exclude frozen parameters") + parser.add_argument("-d", "--debug", action='store_true', help="enable debug") + args = parser.parse_args() + + debug = args.debug + + convert_zero_checkpoint_to_fp32_state_dict(args.checkpoint_dir, + args.output_file, + tag=args.tag, + exclude_frozen_parameters=args.exclude_frozen_parameters)