Upload fine-tuned Llama 3.2 model with LoRA
Browse files- adapter_config.json +1 -1
- adapter_model.safetensors +2 -2
- checkpoint-233/README.md +202 -0
- checkpoint-233/adapter_config.json +34 -0
- checkpoint-233/adapter_model.safetensors +3 -0
- checkpoint-233/optimizer.pt +3 -0
- checkpoint-233/rng_state.pth +3 -0
- checkpoint-233/scheduler.pt +3 -0
- checkpoint-233/trainer_state.json +194 -0
- checkpoint-233/training_args.bin +3 -0
- special_tokens_map.json +4 -10
- tokenizer.json +218 -218
- tokenizer_config.json +5 -6
adapter_config.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
-
"base_model_name_or_path": "
|
| 5 |
"bias": "none",
|
| 6 |
"corda_config": null,
|
| 7 |
"eva_config": null,
|
|
|
|
| 1 |
{
|
| 2 |
"alpha_pattern": {},
|
| 3 |
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "/Users/biomac/converted_llama_3_2_1b",
|
| 5 |
"bias": "none",
|
| 6 |
"corda_config": null,
|
| 7 |
"eva_config": null,
|
adapter_model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:702ad3adb1b9901795aed92f924eedcfefeeb7a315e299dc4518d867ded60656
|
| 3 |
+
size 5778752
|
checkpoint-233/README.md
ADDED
|
@@ -0,0 +1,202 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
base_model: /Users/biomac/converted_llama_3_2_1b
|
| 3 |
+
library_name: peft
|
| 4 |
+
---
|
| 5 |
+
|
| 6 |
+
# Model Card for Model ID
|
| 7 |
+
|
| 8 |
+
<!-- Provide a quick summary of what the model is/does. -->
|
| 9 |
+
|
| 10 |
+
|
| 11 |
+
|
| 12 |
+
## Model Details
|
| 13 |
+
|
| 14 |
+
### Model Description
|
| 15 |
+
|
| 16 |
+
<!-- Provide a longer summary of what this model is. -->
|
| 17 |
+
|
| 18 |
+
|
| 19 |
+
|
| 20 |
+
- **Developed by:** [More Information Needed]
|
| 21 |
+
- **Funded by [optional]:** [More Information Needed]
|
| 22 |
+
- **Shared by [optional]:** [More Information Needed]
|
| 23 |
+
- **Model type:** [More Information Needed]
|
| 24 |
+
- **Language(s) (NLP):** [More Information Needed]
|
| 25 |
+
- **License:** [More Information Needed]
|
| 26 |
+
- **Finetuned from model [optional]:** [More Information Needed]
|
| 27 |
+
|
| 28 |
+
### Model Sources [optional]
|
| 29 |
+
|
| 30 |
+
<!-- Provide the basic links for the model. -->
|
| 31 |
+
|
| 32 |
+
- **Repository:** [More Information Needed]
|
| 33 |
+
- **Paper [optional]:** [More Information Needed]
|
| 34 |
+
- **Demo [optional]:** [More Information Needed]
|
| 35 |
+
|
| 36 |
+
## Uses
|
| 37 |
+
|
| 38 |
+
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 39 |
+
|
| 40 |
+
### Direct Use
|
| 41 |
+
|
| 42 |
+
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
|
| 43 |
+
|
| 44 |
+
[More Information Needed]
|
| 45 |
+
|
| 46 |
+
### Downstream Use [optional]
|
| 47 |
+
|
| 48 |
+
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
|
| 49 |
+
|
| 50 |
+
[More Information Needed]
|
| 51 |
+
|
| 52 |
+
### Out-of-Scope Use
|
| 53 |
+
|
| 54 |
+
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
|
| 55 |
+
|
| 56 |
+
[More Information Needed]
|
| 57 |
+
|
| 58 |
+
## Bias, Risks, and Limitations
|
| 59 |
+
|
| 60 |
+
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
|
| 61 |
+
|
| 62 |
+
[More Information Needed]
|
| 63 |
+
|
| 64 |
+
### Recommendations
|
| 65 |
+
|
| 66 |
+
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
|
| 67 |
+
|
| 68 |
+
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
|
| 69 |
+
|
| 70 |
+
## How to Get Started with the Model
|
| 71 |
+
|
| 72 |
+
Use the code below to get started with the model.
|
| 73 |
+
|
| 74 |
+
[More Information Needed]
|
| 75 |
+
|
| 76 |
+
## Training Details
|
| 77 |
+
|
| 78 |
+
### Training Data
|
| 79 |
+
|
| 80 |
+
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
|
| 81 |
+
|
| 82 |
+
[More Information Needed]
|
| 83 |
+
|
| 84 |
+
### Training Procedure
|
| 85 |
+
|
| 86 |
+
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
|
| 87 |
+
|
| 88 |
+
#### Preprocessing [optional]
|
| 89 |
+
|
| 90 |
+
[More Information Needed]
|
| 91 |
+
|
| 92 |
+
|
| 93 |
+
#### Training Hyperparameters
|
| 94 |
+
|
| 95 |
+
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
|
| 96 |
+
|
| 97 |
+
#### Speeds, Sizes, Times [optional]
|
| 98 |
+
|
| 99 |
+
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
|
| 100 |
+
|
| 101 |
+
[More Information Needed]
|
| 102 |
+
|
| 103 |
+
## Evaluation
|
| 104 |
+
|
| 105 |
+
<!-- This section describes the evaluation protocols and provides the results. -->
|
| 106 |
+
|
| 107 |
+
### Testing Data, Factors & Metrics
|
| 108 |
+
|
| 109 |
+
#### Testing Data
|
| 110 |
+
|
| 111 |
+
<!-- This should link to a Dataset Card if possible. -->
|
| 112 |
+
|
| 113 |
+
[More Information Needed]
|
| 114 |
+
|
| 115 |
+
#### Factors
|
| 116 |
+
|
| 117 |
+
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
|
| 118 |
+
|
| 119 |
+
[More Information Needed]
|
| 120 |
+
|
| 121 |
+
#### Metrics
|
| 122 |
+
|
| 123 |
+
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
| 124 |
+
|
| 125 |
+
[More Information Needed]
|
| 126 |
+
|
| 127 |
+
### Results
|
| 128 |
+
|
| 129 |
+
[More Information Needed]
|
| 130 |
+
|
| 131 |
+
#### Summary
|
| 132 |
+
|
| 133 |
+
|
| 134 |
+
|
| 135 |
+
## Model Examination [optional]
|
| 136 |
+
|
| 137 |
+
<!-- Relevant interpretability work for the model goes here -->
|
| 138 |
+
|
| 139 |
+
[More Information Needed]
|
| 140 |
+
|
| 141 |
+
## Environmental Impact
|
| 142 |
+
|
| 143 |
+
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
| 144 |
+
|
| 145 |
+
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
| 146 |
+
|
| 147 |
+
- **Hardware Type:** [More Information Needed]
|
| 148 |
+
- **Hours used:** [More Information Needed]
|
| 149 |
+
- **Cloud Provider:** [More Information Needed]
|
| 150 |
+
- **Compute Region:** [More Information Needed]
|
| 151 |
+
- **Carbon Emitted:** [More Information Needed]
|
| 152 |
+
|
| 153 |
+
## Technical Specifications [optional]
|
| 154 |
+
|
| 155 |
+
### Model Architecture and Objective
|
| 156 |
+
|
| 157 |
+
[More Information Needed]
|
| 158 |
+
|
| 159 |
+
### Compute Infrastructure
|
| 160 |
+
|
| 161 |
+
[More Information Needed]
|
| 162 |
+
|
| 163 |
+
#### Hardware
|
| 164 |
+
|
| 165 |
+
[More Information Needed]
|
| 166 |
+
|
| 167 |
+
#### Software
|
| 168 |
+
|
| 169 |
+
[More Information Needed]
|
| 170 |
+
|
| 171 |
+
## Citation [optional]
|
| 172 |
+
|
| 173 |
+
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
|
| 174 |
+
|
| 175 |
+
**BibTeX:**
|
| 176 |
+
|
| 177 |
+
[More Information Needed]
|
| 178 |
+
|
| 179 |
+
**APA:**
|
| 180 |
+
|
| 181 |
+
[More Information Needed]
|
| 182 |
+
|
| 183 |
+
## Glossary [optional]
|
| 184 |
+
|
| 185 |
+
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
|
| 186 |
+
|
| 187 |
+
[More Information Needed]
|
| 188 |
+
|
| 189 |
+
## More Information [optional]
|
| 190 |
+
|
| 191 |
+
[More Information Needed]
|
| 192 |
+
|
| 193 |
+
## Model Card Authors [optional]
|
| 194 |
+
|
| 195 |
+
[More Information Needed]
|
| 196 |
+
|
| 197 |
+
## Model Card Contact
|
| 198 |
+
|
| 199 |
+
[More Information Needed]
|
| 200 |
+
### Framework versions
|
| 201 |
+
|
| 202 |
+
- PEFT 0.15.1
|
checkpoint-233/adapter_config.json
ADDED
|
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"alpha_pattern": {},
|
| 3 |
+
"auto_mapping": null,
|
| 4 |
+
"base_model_name_or_path": "/Users/biomac/converted_llama_3_2_1b",
|
| 5 |
+
"bias": "none",
|
| 6 |
+
"corda_config": null,
|
| 7 |
+
"eva_config": null,
|
| 8 |
+
"exclude_modules": null,
|
| 9 |
+
"fan_in_fan_out": false,
|
| 10 |
+
"inference_mode": true,
|
| 11 |
+
"init_lora_weights": true,
|
| 12 |
+
"layer_replication": null,
|
| 13 |
+
"layers_pattern": null,
|
| 14 |
+
"layers_to_transform": null,
|
| 15 |
+
"loftq_config": {},
|
| 16 |
+
"lora_alpha": 16,
|
| 17 |
+
"lora_bias": false,
|
| 18 |
+
"lora_dropout": 0.05,
|
| 19 |
+
"megatron_config": null,
|
| 20 |
+
"megatron_core": "megatron.core",
|
| 21 |
+
"modules_to_save": null,
|
| 22 |
+
"peft_type": "LORA",
|
| 23 |
+
"r": 8,
|
| 24 |
+
"rank_pattern": {},
|
| 25 |
+
"revision": null,
|
| 26 |
+
"target_modules": [
|
| 27 |
+
"q_proj",
|
| 28 |
+
"v_proj"
|
| 29 |
+
],
|
| 30 |
+
"task_type": "CAUSAL_LM",
|
| 31 |
+
"trainable_token_indices": null,
|
| 32 |
+
"use_dora": false,
|
| 33 |
+
"use_rslora": false
|
| 34 |
+
}
|
checkpoint-233/adapter_model.safetensors
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:702ad3adb1b9901795aed92f924eedcfefeeb7a315e299dc4518d867ded60656
|
| 3 |
+
size 5778752
|
checkpoint-233/optimizer.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b42b7ba81ae40e367717b03c57540488948b889a969ca47606b33fc5f71524ce
|
| 3 |
+
size 11606010
|
checkpoint-233/rng_state.pth
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5f2d8e50c2fc8e7f50ca5a741ea4cbc7854d053c3a464a90a36a47cdc38ebe96
|
| 3 |
+
size 13990
|
checkpoint-233/scheduler.pt
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:7a9fc23b13eff1be65c1dc3497088de0b3a3d9eee8e42c9dbfbb436015dab28b
|
| 3 |
+
size 1064
|
checkpoint-233/trainer_state.json
ADDED
|
@@ -0,0 +1,194 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"best_metric": null,
|
| 3 |
+
"best_model_checkpoint": null,
|
| 4 |
+
"epoch": 0.9967914438502674,
|
| 5 |
+
"eval_steps": 500,
|
| 6 |
+
"global_step": 233,
|
| 7 |
+
"is_hyper_param_search": false,
|
| 8 |
+
"is_local_process_zero": true,
|
| 9 |
+
"is_world_process_zero": true,
|
| 10 |
+
"log_history": [
|
| 11 |
+
{
|
| 12 |
+
"epoch": 0.0427807486631016,
|
| 13 |
+
"grad_norm": Infinity,
|
| 14 |
+
"learning_rate": 0.0001666666666666667,
|
| 15 |
+
"loss": 10.5827,
|
| 16 |
+
"step": 10
|
| 17 |
+
},
|
| 18 |
+
{
|
| 19 |
+
"epoch": 0.0855614973262032,
|
| 20 |
+
"grad_norm": Infinity,
|
| 21 |
+
"learning_rate": 0.0001993540527265239,
|
| 22 |
+
"loss": 10.5932,
|
| 23 |
+
"step": 20
|
| 24 |
+
},
|
| 25 |
+
{
|
| 26 |
+
"epoch": 0.12834224598930483,
|
| 27 |
+
"grad_norm": Infinity,
|
| 28 |
+
"learning_rate": 0.00019674418752719833,
|
| 29 |
+
"loss": 10.6018,
|
| 30 |
+
"step": 30
|
| 31 |
+
},
|
| 32 |
+
{
|
| 33 |
+
"epoch": 0.1711229946524064,
|
| 34 |
+
"grad_norm": Infinity,
|
| 35 |
+
"learning_rate": 0.00019218264229806917,
|
| 36 |
+
"loss": 10.5882,
|
| 37 |
+
"step": 40
|
| 38 |
+
},
|
| 39 |
+
{
|
| 40 |
+
"epoch": 0.21390374331550802,
|
| 41 |
+
"grad_norm": Infinity,
|
| 42 |
+
"learning_rate": 0.00018576143989814523,
|
| 43 |
+
"loss": 10.5865,
|
| 44 |
+
"step": 50
|
| 45 |
+
},
|
| 46 |
+
{
|
| 47 |
+
"epoch": 0.25668449197860965,
|
| 48 |
+
"grad_norm": Infinity,
|
| 49 |
+
"learning_rate": 0.00017761011919494488,
|
| 50 |
+
"loss": 10.5991,
|
| 51 |
+
"step": 60
|
| 52 |
+
},
|
| 53 |
+
{
|
| 54 |
+
"epoch": 0.2994652406417112,
|
| 55 |
+
"grad_norm": Infinity,
|
| 56 |
+
"learning_rate": 0.00016789312179720547,
|
| 57 |
+
"loss": 10.5906,
|
| 58 |
+
"step": 70
|
| 59 |
+
},
|
| 60 |
+
{
|
| 61 |
+
"epoch": 0.3422459893048128,
|
| 62 |
+
"grad_norm": Infinity,
|
| 63 |
+
"learning_rate": 0.00015680647467311557,
|
| 64 |
+
"loss": 10.5962,
|
| 65 |
+
"step": 80
|
| 66 |
+
},
|
| 67 |
+
{
|
| 68 |
+
"epoch": 0.3850267379679144,
|
| 69 |
+
"grad_norm": Infinity,
|
| 70 |
+
"learning_rate": 0.00014457383557765386,
|
| 71 |
+
"loss": 10.6006,
|
| 72 |
+
"step": 90
|
| 73 |
+
},
|
| 74 |
+
{
|
| 75 |
+
"epoch": 0.42780748663101603,
|
| 76 |
+
"grad_norm": Infinity,
|
| 77 |
+
"learning_rate": 0.0001314419810670624,
|
| 78 |
+
"loss": 10.59,
|
| 79 |
+
"step": 100
|
| 80 |
+
},
|
| 81 |
+
{
|
| 82 |
+
"epoch": 0.47058823529411764,
|
| 83 |
+
"grad_norm": Infinity,
|
| 84 |
+
"learning_rate": 0.0001176758281235155,
|
| 85 |
+
"loss": 10.5897,
|
| 86 |
+
"step": 110
|
| 87 |
+
},
|
| 88 |
+
{
|
| 89 |
+
"epoch": 0.5133689839572193,
|
| 90 |
+
"grad_norm": Infinity,
|
| 91 |
+
"learning_rate": 0.00010355308982181253,
|
| 92 |
+
"loss": 10.5836,
|
| 93 |
+
"step": 120
|
| 94 |
+
},
|
| 95 |
+
{
|
| 96 |
+
"epoch": 0.5561497326203209,
|
| 97 |
+
"grad_norm": Infinity,
|
| 98 |
+
"learning_rate": 8.935867285261978e-05,
|
| 99 |
+
"loss": 10.5922,
|
| 100 |
+
"step": 130
|
| 101 |
+
},
|
| 102 |
+
{
|
| 103 |
+
"epoch": 0.5989304812834224,
|
| 104 |
+
"grad_norm": Infinity,
|
| 105 |
+
"learning_rate": 7.537892992447273e-05,
|
| 106 |
+
"loss": 10.5835,
|
| 107 |
+
"step": 140
|
| 108 |
+
},
|
| 109 |
+
{
|
| 110 |
+
"epoch": 0.6417112299465241,
|
| 111 |
+
"grad_norm": Infinity,
|
| 112 |
+
"learning_rate": 6.189588299436997e-05,
|
| 113 |
+
"loss": 10.6007,
|
| 114 |
+
"step": 150
|
| 115 |
+
},
|
| 116 |
+
{
|
| 117 |
+
"epoch": 0.6844919786096256,
|
| 118 |
+
"grad_norm": Infinity,
|
| 119 |
+
"learning_rate": 4.9181533865282706e-05,
|
| 120 |
+
"loss": 10.5927,
|
| 121 |
+
"step": 160
|
| 122 |
+
},
|
| 123 |
+
{
|
| 124 |
+
"epoch": 0.7272727272727273,
|
| 125 |
+
"grad_norm": Infinity,
|
| 126 |
+
"learning_rate": 3.7492376926397964e-05,
|
| 127 |
+
"loss": 10.5852,
|
| 128 |
+
"step": 170
|
| 129 |
+
},
|
| 130 |
+
{
|
| 131 |
+
"epoch": 0.7700534759358288,
|
| 132 |
+
"grad_norm": Infinity,
|
| 133 |
+
"learning_rate": 2.7064224733963197e-05,
|
| 134 |
+
"loss": 10.5855,
|
| 135 |
+
"step": 180
|
| 136 |
+
},
|
| 137 |
+
{
|
| 138 |
+
"epoch": 0.8128342245989305,
|
| 139 |
+
"grad_norm": Infinity,
|
| 140 |
+
"learning_rate": 1.8107450819473503e-05,
|
| 141 |
+
"loss": 10.5925,
|
| 142 |
+
"step": 190
|
| 143 |
+
},
|
| 144 |
+
{
|
| 145 |
+
"epoch": 0.8556149732620321,
|
| 146 |
+
"grad_norm": Infinity,
|
| 147 |
+
"learning_rate": 1.0802745694960569e-05,
|
| 148 |
+
"loss": 10.5893,
|
| 149 |
+
"step": 200
|
| 150 |
+
},
|
| 151 |
+
{
|
| 152 |
+
"epoch": 0.8983957219251337,
|
| 153 |
+
"grad_norm": Infinity,
|
| 154 |
+
"learning_rate": 5.29747167209923e-06,
|
| 155 |
+
"loss": 10.5813,
|
| 156 |
+
"step": 210
|
| 157 |
+
},
|
| 158 |
+
{
|
| 159 |
+
"epoch": 0.9411764705882353,
|
| 160 |
+
"grad_norm": Infinity,
|
| 161 |
+
"learning_rate": 1.7026900316098215e-06,
|
| 162 |
+
"loss": 10.6031,
|
| 163 |
+
"step": 220
|
| 164 |
+
},
|
| 165 |
+
{
|
| 166 |
+
"epoch": 0.983957219251337,
|
| 167 |
+
"grad_norm": Infinity,
|
| 168 |
+
"learning_rate": 9.092051569674631e-08,
|
| 169 |
+
"loss": 10.5864,
|
| 170 |
+
"step": 230
|
| 171 |
+
}
|
| 172 |
+
],
|
| 173 |
+
"logging_steps": 10,
|
| 174 |
+
"max_steps": 233,
|
| 175 |
+
"num_input_tokens_seen": 0,
|
| 176 |
+
"num_train_epochs": 1,
|
| 177 |
+
"save_steps": 500,
|
| 178 |
+
"stateful_callbacks": {
|
| 179 |
+
"TrainerControl": {
|
| 180 |
+
"args": {
|
| 181 |
+
"should_epoch_stop": false,
|
| 182 |
+
"should_evaluate": false,
|
| 183 |
+
"should_log": false,
|
| 184 |
+
"should_save": true,
|
| 185 |
+
"should_training_stop": true
|
| 186 |
+
},
|
| 187 |
+
"attributes": {}
|
| 188 |
+
}
|
| 189 |
+
},
|
| 190 |
+
"total_flos": 8838201401671680.0,
|
| 191 |
+
"train_batch_size": 1,
|
| 192 |
+
"trial_name": null,
|
| 193 |
+
"trial_params": null
|
| 194 |
+
}
|
checkpoint-233/training_args.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:26a917f57899e8c56d0800023a8f9e37ddddbb3fff7afcc0e427f9409cb6922b
|
| 3 |
+
size 5368
|
special_tokens_map.json
CHANGED
|
@@ -2,28 +2,22 @@
|
|
| 2 |
"bos_token": {
|
| 3 |
"content": "<s>",
|
| 4 |
"lstrip": false,
|
| 5 |
-
"normalized":
|
| 6 |
"rstrip": false,
|
| 7 |
"single_word": false
|
| 8 |
},
|
| 9 |
"eos_token": {
|
| 10 |
"content": "</s>",
|
| 11 |
"lstrip": false,
|
| 12 |
-
"normalized":
|
| 13 |
-
"rstrip": false,
|
| 14 |
-
"single_word": false
|
| 15 |
-
},
|
| 16 |
-
"pad_token": {
|
| 17 |
-
"content": "</s>",
|
| 18 |
-
"lstrip": false,
|
| 19 |
-
"normalized": false,
|
| 20 |
"rstrip": false,
|
| 21 |
"single_word": false
|
| 22 |
},
|
|
|
|
| 23 |
"unk_token": {
|
| 24 |
"content": "<unk>",
|
| 25 |
"lstrip": false,
|
| 26 |
-
"normalized":
|
| 27 |
"rstrip": false,
|
| 28 |
"single_word": false
|
| 29 |
}
|
|
|
|
| 2 |
"bos_token": {
|
| 3 |
"content": "<s>",
|
| 4 |
"lstrip": false,
|
| 5 |
+
"normalized": true,
|
| 6 |
"rstrip": false,
|
| 7 |
"single_word": false
|
| 8 |
},
|
| 9 |
"eos_token": {
|
| 10 |
"content": "</s>",
|
| 11 |
"lstrip": false,
|
| 12 |
+
"normalized": true,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 13 |
"rstrip": false,
|
| 14 |
"single_word": false
|
| 15 |
},
|
| 16 |
+
"pad_token": "</s>",
|
| 17 |
"unk_token": {
|
| 18 |
"content": "<unk>",
|
| 19 |
"lstrip": false,
|
| 20 |
+
"normalized": true,
|
| 21 |
"rstrip": false,
|
| 22 |
"single_word": false
|
| 23 |
}
|
tokenizer.json
CHANGED
|
@@ -2,15 +2,15 @@
|
|
| 2 |
"version": "1.0",
|
| 3 |
"truncation": {
|
| 4 |
"direction": "Right",
|
| 5 |
-
"max_length":
|
| 6 |
"strategy": "LongestFirst",
|
| 7 |
"stride": 0
|
| 8 |
},
|
| 9 |
"padding": {
|
| 10 |
"strategy": {
|
| 11 |
-
"Fixed":
|
| 12 |
},
|
| 13 |
-
"direction": "
|
| 14 |
"pad_to_multiple_of": null,
|
| 15 |
"pad_id": 2,
|
| 16 |
"pad_type_id": 0,
|
|
@@ -23,7 +23,7 @@
|
|
| 23 |
"single_word": false,
|
| 24 |
"lstrip": false,
|
| 25 |
"rstrip": false,
|
| 26 |
-
"normalized":
|
| 27 |
"special": true
|
| 28 |
},
|
| 29 |
{
|
|
@@ -32,7 +32,7 @@
|
|
| 32 |
"single_word": false,
|
| 33 |
"lstrip": false,
|
| 34 |
"rstrip": false,
|
| 35 |
-
"normalized":
|
| 36 |
"special": true
|
| 37 |
},
|
| 38 |
{
|
|
@@ -41,7 +41,7 @@
|
|
| 41 |
"single_word": false,
|
| 42 |
"lstrip": false,
|
| 43 |
"rstrip": false,
|
| 44 |
-
"normalized":
|
| 45 |
"special": true
|
| 46 |
}
|
| 47 |
],
|
|
@@ -276669,484 +276669,484 @@
|
|
| 276669 |
"ub"
|
| 276670 |
],
|
| 276671 |
[
|
| 276672 |
-
"β",
|
| 276673 |
-
"β"
|
| 276674 |
],
|
| 276675 |
[
|
| 276676 |
"ββ",
|
| 276677 |
-
"ββ"
|
| 276678 |
],
|
| 276679 |
[
|
| 276680 |
-
"ββ
|
| 276681 |
-
"β"
|
| 276682 |
],
|
| 276683 |
[
|
| 276684 |
-
"β",
|
| 276685 |
-
"βββ"
|
| 276686 |
],
|
| 276687 |
[
|
| 276688 |
"ββ",
|
| 276689 |
"ββββββ"
|
| 276690 |
],
|
| 276691 |
[
|
| 276692 |
-
"ββ
|
| 276693 |
-
"ββββ"
|
| 276694 |
],
|
| 276695 |
[
|
| 276696 |
-
"ββ
|
| 276697 |
-
"βββ"
|
| 276698 |
],
|
| 276699 |
[
|
| 276700 |
-
"ββ
|
| 276701 |
-
"ββ"
|
| 276702 |
],
|
| 276703 |
[
|
| 276704 |
-
"ββ
|
| 276705 |
-
"βββββ"
|
| 276706 |
],
|
| 276707 |
[
|
| 276708 |
-
"ββ
|
| 276709 |
-
"β"
|
| 276710 |
],
|
| 276711 |
[
|
| 276712 |
-
"β",
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276713 |
"βββββββ"
|
| 276714 |
],
|
| 276715 |
[
|
| 276716 |
"ββ",
|
| 276717 |
-
"βββ"
|
| 276718 |
],
|
| 276719 |
[
|
| 276720 |
-
"ββ
|
| 276721 |
"β"
|
| 276722 |
],
|
| 276723 |
[
|
| 276724 |
-
"βββ",
|
| 276725 |
"ββ"
|
| 276726 |
],
|
| 276727 |
[
|
| 276728 |
-
"β",
|
| 276729 |
"ββββ"
|
| 276730 |
],
|
| 276731 |
[
|
| 276732 |
-
"ββ",
|
| 276733 |
-
"ββββββββ
|
| 276734 |
],
|
| 276735 |
[
|
| 276736 |
"ββββ",
|
| 276737 |
-
"βββββ
|
| 276738 |
],
|
| 276739 |
[
|
| 276740 |
-
"ββββ
|
| 276741 |
-
"ββββββ
|
| 276742 |
],
|
| 276743 |
[
|
| 276744 |
-
"ββββ
|
| 276745 |
-
"βββββββββββ"
|
| 276746 |
],
|
| 276747 |
[
|
| 276748 |
-
"ββββ
|
| 276749 |
"ββββββββββ"
|
| 276750 |
],
|
| 276751 |
[
|
| 276752 |
-
"ββββ
|
| 276753 |
-
"βββ
|
| 276754 |
],
|
| 276755 |
[
|
| 276756 |
-
"ββββ
|
| 276757 |
-
"βββ"
|
| 276758 |
],
|
| 276759 |
[
|
| 276760 |
-
"ββββ
|
| 276761 |
-
"ββββββ"
|
| 276762 |
],
|
| 276763 |
[
|
| 276764 |
-
"ββββ
|
| 276765 |
-
"ββ"
|
| 276766 |
],
|
| 276767 |
[
|
| 276768 |
-
"βββ",
|
| 276769 |
-
"β
|
| 276770 |
],
|
| 276771 |
[
|
| 276772 |
-
"ββββββββ
|
| 276773 |
-
"ββ
|
| 276774 |
],
|
| 276775 |
[
|
| 276776 |
-
"βββββββ",
|
| 276777 |
-
"ββββ
|
| 276778 |
],
|
| 276779 |
[
|
| 276780 |
-
"ββββββββ
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276781 |
"βββββ"
|
| 276782 |
],
|
| 276783 |
[
|
| 276784 |
-
"ββββββββ
|
| 276785 |
-
"β"
|
| 276786 |
],
|
| 276787 |
[
|
| 276788 |
-
"β",
|
| 276789 |
-
"βββ
|
| 276790 |
],
|
| 276791 |
[
|
| 276792 |
-
"ββ",
|
| 276793 |
-
"ββββ"
|
| 276794 |
],
|
| 276795 |
[
|
| 276796 |
-
"ββββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276797 |
"ββ"
|
| 276798 |
],
|
| 276799 |
[
|
| 276800 |
"βββββ",
|
| 276801 |
-
"β"
|
| 276802 |
],
|
| 276803 |
[
|
| 276804 |
-
"βββ",
|
| 276805 |
-
"βββ"
|
| 276806 |
],
|
| 276807 |
[
|
| 276808 |
-
"β",
|
| 276809 |
"βββββ"
|
| 276810 |
],
|
| 276811 |
[
|
| 276812 |
-
"ββ",
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276813 |
"ββββββββββ"
|
| 276814 |
],
|
| 276815 |
[
|
| 276816 |
-
"ββββ",
|
| 276817 |
-
"βββ
|
| 276818 |
],
|
| 276819 |
[
|
| 276820 |
-
"βββββ
|
| 276821 |
-
"ββββ"
|
| 276822 |
],
|
| 276823 |
[
|
| 276824 |
"βββββ",
|
| 276825 |
"βββββββ"
|
| 276826 |
],
|
| 276827 |
[
|
| 276828 |
-
"βββββ
|
| 276829 |
-
"ββββββ"
|
| 276830 |
],
|
| 276831 |
[
|
| 276832 |
-
"βββββ
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276833 |
"ββ"
|
| 276834 |
],
|
| 276835 |
[
|
| 276836 |
-
"βββ",
|
| 276837 |
-
"ββββ
|
| 276838 |
],
|
| 276839 |
[
|
| 276840 |
-
"ββββββ
|
| 276841 |
-
"βββ"
|
| 276842 |
],
|
| 276843 |
[
|
| 276844 |
-
"ββββββ
|
| 276845 |
"βββββ"
|
| 276846 |
],
|
| 276847 |
[
|
| 276848 |
-
"ββββββ
|
| 276849 |
-
"β"
|
| 276850 |
],
|
| 276851 |
[
|
| 276852 |
-
"β",
|
| 276853 |
-
"ββββββββββ
|
| 276854 |
],
|
| 276855 |
[
|
| 276856 |
-
"ββ",
|
| 276857 |
-
"βββ
|
| 276858 |
],
|
| 276859 |
[
|
| 276860 |
-
"ββββ",
|
| 276861 |
"βββββββββ"
|
| 276862 |
],
|
| 276863 |
-
[
|
| 276864 |
-
"ββββββββ",
|
| 276865 |
-
"βββββ"
|
| 276866 |
-
],
|
| 276867 |
-
[
|
| 276868 |
-
"βββββ",
|
| 276869 |
-
"ββββββββ"
|
| 276870 |
-
],
|
| 276871 |
[
|
| 276872 |
"ββββββ",
|
| 276873 |
"βββββββ"
|
| 276874 |
],
|
| 276875 |
[
|
| 276876 |
-
"ββββββ
|
| 276877 |
"β"
|
| 276878 |
],
|
| 276879 |
[
|
| 276880 |
-
"ββββββββββ",
|
| 276881 |
-
"ββ
|
| 276882 |
],
|
| 276883 |
[
|
| 276884 |
-
"βββ",
|
| 276885 |
-
"ββββ
|
| 276886 |
],
|
| 276887 |
[
|
| 276888 |
-
"βββββββββ",
|
| 276889 |
-
"βββ
|
| 276890 |
],
|
| 276891 |
[
|
| 276892 |
-
"βββββββ",
|
| 276893 |
-
"β
|
| 276894 |
],
|
| 276895 |
[
|
| 276896 |
-
"βββββββββββ",
|
| 276897 |
"ββ"
|
| 276898 |
],
|
| 276899 |
[
|
| 276900 |
-
"β",
|
| 276901 |
-
"βββ
|
| 276902 |
],
|
| 276903 |
[
|
| 276904 |
-
"ββ",
|
| 276905 |
-
"β
|
| 276906 |
],
|
| 276907 |
[
|
| 276908 |
-
"ββββ",
|
| 276909 |
-
"ββ
|
| 276910 |
],
|
| 276911 |
[
|
| 276912 |
-
"ββββββββ",
|
| 276913 |
-
"ββ"
|
| 276914 |
],
|
| 276915 |
[
|
| 276916 |
-
"βββββ",
|
| 276917 |
"βββββ"
|
| 276918 |
],
|
| 276919 |
[
|
| 276920 |
-
"ββββββ",
|
| 276921 |
-
"ββββ"
|
| 276922 |
],
|
| 276923 |
[
|
| 276924 |
-
"βββ",
|
| 276925 |
-
"βββ
|
| 276926 |
],
|
| 276927 |
[
|
| 276928 |
-
"βββββββββ",
|
| 276929 |
"β"
|
| 276930 |
],
|
| 276931 |
[
|
| 276932 |
-
"βββββββ",
|
| 276933 |
-
"ββ
|
| 276934 |
],
|
| 276935 |
[
|
| 276936 |
-
"β",
|
| 276937 |
-
"β
|
| 276938 |
],
|
| 276939 |
[
|
| 276940 |
-
"ββ",
|
| 276941 |
-
"ββ
|
| 276942 |
],
|
| 276943 |
[
|
| 276944 |
-
"βββ
|
| 276945 |
-
"ββββ
|
| 276946 |
],
|
| 276947 |
[
|
| 276948 |
-
"βββ
|
| 276949 |
-
"ββββββ"
|
| 276950 |
],
|
| 276951 |
[
|
| 276952 |
-
"βββ
|
| 276953 |
-
"ββ
|
| 276954 |
],
|
| 276955 |
[
|
| 276956 |
-
"βββ
|
| 276957 |
-
"ββββββ
|
| 276958 |
],
|
| 276959 |
[
|
| 276960 |
-
"βββ
|
| 276961 |
-
"ββ"
|
| 276962 |
],
|
| 276963 |
[
|
| 276964 |
-
"βββ
|
| 276965 |
-
"β"
|
| 276966 |
],
|
| 276967 |
[
|
| 276968 |
-
"βββ
|
| 276969 |
-
"ββββ"
|
| 276970 |
],
|
| 276971 |
[
|
| 276972 |
"βββ",
|
| 276973 |
-
"βββ
|
| 276974 |
],
|
| 276975 |
[
|
| 276976 |
-
"βββ
|
| 276977 |
-
"βββββ"
|
| 276978 |
],
|
| 276979 |
[
|
| 276980 |
-
"βββ
|
| 276981 |
"βββββββ"
|
| 276982 |
],
|
| 276983 |
[
|
| 276984 |
-
"βββ
|
| 276985 |
-
"βββ"
|
| 276986 |
-
],
|
| 276987 |
-
[
|
| 276988 |
-
"β",
|
| 276989 |
-
"βββββββββββββ"
|
| 276990 |
],
|
| 276991 |
[
|
| 276992 |
-
"ββ",
|
| 276993 |
"β"
|
| 276994 |
],
|
| 276995 |
[
|
| 276996 |
-
"β",
|
| 276997 |
"ββ"
|
| 276998 |
],
|
| 276999 |
[
|
| 277000 |
-
"ββ",
|
| 277001 |
-
"ββββ
|
| 277002 |
],
|
| 277003 |
[
|
| 277004 |
-
"ββββ",
|
| 277005 |
"βββββ"
|
| 277006 |
],
|
| 277007 |
[
|
| 277008 |
-
"ββββββββ",
|
| 277009 |
-
"β"
|
| 277010 |
],
|
| 277011 |
[
|
| 277012 |
-
"βββββ",
|
| 277013 |
-
"βββ
|
| 277014 |
],
|
| 277015 |
[
|
| 277016 |
-
"ββββββ",
|
| 277017 |
-
"βββ"
|
| 277018 |
],
|
| 277019 |
[
|
| 277020 |
-
"βββ",
|
| 277021 |
-
"β
|
| 277022 |
],
|
| 277023 |
[
|
| 277024 |
"βββββββ",
|
| 277025 |
"ββ"
|
| 277026 |
],
|
| 277027 |
[
|
| 277028 |
-
"β",
|
| 277029 |
-
"ββββ
|
| 277030 |
],
|
| 277031 |
[
|
| 277032 |
-
"ββ",
|
| 277033 |
-
"βββββ"
|
| 277034 |
],
|
| 277035 |
[
|
| 277036 |
-
"ββββ",
|
| 277037 |
-
"βββ"
|
| 277038 |
],
|
| 277039 |
[
|
| 277040 |
-
"βββββ",
|
| 277041 |
-
"ββ"
|
| 277042 |
],
|
| 277043 |
[
|
| 277044 |
-
"ββββββ",
|
| 277045 |
-
"β"
|
| 277046 |
],
|
| 277047 |
[
|
| 277048 |
-
"βββ",
|
| 277049 |
-
"ββββ"
|
| 277050 |
],
|
| 277051 |
[
|
| 277052 |
-
"β",
|
| 277053 |
-
"ββββββ"
|
| 277054 |
],
|
| 277055 |
[
|
| 277056 |
-
"ββ",
|
| 277057 |
-
"β
|
| 277058 |
],
|
| 277059 |
[
|
| 277060 |
-
"ββββ",
|
| 277061 |
-
"ββ
|
| 277062 |
],
|
| 277063 |
[
|
| 277064 |
-
"ββββββββ",
|
| 277065 |
-
"βββ"
|
| 277066 |
],
|
| 277067 |
[
|
| 277068 |
-
"βββββ",
|
| 277069 |
-
"βββββ
|
| 277070 |
],
|
| 277071 |
[
|
| 277072 |
-
"ββββββ",
|
| 277073 |
-
"βββ
|
| 277074 |
],
|
| 277075 |
[
|
| 277076 |
-
"ββββββββββ",
|
| 277077 |
"β"
|
| 277078 |
],
|
| 277079 |
[
|
| 277080 |
-
"βββ",
|
| 277081 |
-
"β
|
| 277082 |
],
|
| 277083 |
[
|
| 277084 |
-
"β
|
| 277085 |
"ββ"
|
| 277086 |
],
|
| 277087 |
[
|
| 277088 |
-
"β
|
| 277089 |
"ββββ"
|
| 277090 |
],
|
| 277091 |
[
|
| 277092 |
"β",
|
| 277093 |
-
"ββββββββ
|
| 277094 |
-
],
|
| 277095 |
-
[
|
| 277096 |
-
"ββ",
|
| 277097 |
-
"βββββββββββββ"
|
| 277098 |
-
],
|
| 277099 |
-
[
|
| 277100 |
-
"ββββ",
|
| 277101 |
-
"βββββββββββ"
|
| 277102 |
],
|
| 277103 |
[
|
| 277104 |
-
"β
|
| 277105 |
-
"βββββ
|
| 277106 |
],
|
| 277107 |
[
|
| 277108 |
-
"β
|
| 277109 |
-
"ββββββ
|
| 277110 |
],
|
| 277111 |
[
|
| 277112 |
-
"β
|
| 277113 |
-
"βββββββββ"
|
| 277114 |
],
|
| 277115 |
[
|
| 277116 |
-
"
|
| 277117 |
-
"βββ"
|
| 277118 |
],
|
| 277119 |
[
|
| 277120 |
-
"β
|
| 277121 |
-
"ββ"
|
| 277122 |
],
|
| 277123 |
[
|
| 277124 |
-
"β
|
| 277125 |
-
"βββββ"
|
| 277126 |
],
|
| 277127 |
[
|
| 277128 |
-
"β
|
| 277129 |
-
"β"
|
| 277130 |
],
|
| 277131 |
[
|
| 277132 |
-
"β
|
| 277133 |
-
"βββββββββ
|
| 277134 |
],
|
| 277135 |
[
|
| 277136 |
-
"β
|
| 277137 |
-
"ββββββ"
|
| 277138 |
],
|
| 277139 |
[
|
| 277140 |
-
"β
|
| 277141 |
-
"ββββββββ"
|
| 277142 |
],
|
| 277143 |
[
|
| 277144 |
-
"β
|
| 277145 |
-
"ββββ"
|
| 277146 |
],
|
| 277147 |
[
|
| 277148 |
"β",
|
| 277149 |
-
"β
|
| 277150 |
]
|
| 277151 |
]
|
| 277152 |
}
|
|
|
|
| 2 |
"version": "1.0",
|
| 3 |
"truncation": {
|
| 4 |
"direction": "Right",
|
| 5 |
+
"max_length": 256,
|
| 6 |
"strategy": "LongestFirst",
|
| 7 |
"stride": 0
|
| 8 |
},
|
| 9 |
"padding": {
|
| 10 |
"strategy": {
|
| 11 |
+
"Fixed": 256
|
| 12 |
},
|
| 13 |
+
"direction": "Left",
|
| 14 |
"pad_to_multiple_of": null,
|
| 15 |
"pad_id": 2,
|
| 16 |
"pad_type_id": 0,
|
|
|
|
| 23 |
"single_word": false,
|
| 24 |
"lstrip": false,
|
| 25 |
"rstrip": false,
|
| 26 |
+
"normalized": true,
|
| 27 |
"special": true
|
| 28 |
},
|
| 29 |
{
|
|
|
|
| 32 |
"single_word": false,
|
| 33 |
"lstrip": false,
|
| 34 |
"rstrip": false,
|
| 35 |
+
"normalized": true,
|
| 36 |
"special": true
|
| 37 |
},
|
| 38 |
{
|
|
|
|
| 41 |
"single_word": false,
|
| 42 |
"lstrip": false,
|
| 43 |
"rstrip": false,
|
| 44 |
+
"normalized": true,
|
| 45 |
"special": true
|
| 46 |
}
|
| 47 |
],
|
|
|
|
| 276669 |
"ub"
|
| 276670 |
],
|
| 276671 |
[
|
| 276672 |
+
"ββ",
|
| 276673 |
+
"ββ"
|
| 276674 |
],
|
| 276675 |
[
|
| 276676 |
"ββ",
|
| 276677 |
+
"ββββ"
|
| 276678 |
],
|
| 276679 |
[
|
| 276680 |
+
"ββ",
|
| 276681 |
+
"ββββββββ"
|
| 276682 |
],
|
| 276683 |
[
|
| 276684 |
+
"ββ",
|
| 276685 |
+
"βββββ"
|
| 276686 |
],
|
| 276687 |
[
|
| 276688 |
"ββ",
|
| 276689 |
"ββββββ"
|
| 276690 |
],
|
| 276691 |
[
|
| 276692 |
+
"ββ",
|
| 276693 |
+
"ββββββββββββ"
|
| 276694 |
],
|
| 276695 |
[
|
| 276696 |
+
"ββ",
|
| 276697 |
+
"βββββββββββββ"
|
| 276698 |
],
|
| 276699 |
[
|
| 276700 |
+
"ββ",
|
| 276701 |
+
"ββββββββββ"
|
| 276702 |
],
|
| 276703 |
[
|
| 276704 |
+
"ββ",
|
| 276705 |
+
"ββββββββββββββ"
|
| 276706 |
],
|
| 276707 |
[
|
| 276708 |
+
"ββ",
|
| 276709 |
+
"βββ"
|
| 276710 |
],
|
| 276711 |
[
|
| 276712 |
+
"ββ",
|
| 276713 |
+
"βββββββββ"
|
| 276714 |
+
],
|
| 276715 |
+
[
|
| 276716 |
+
"ββ",
|
| 276717 |
"βββββββ"
|
| 276718 |
],
|
| 276719 |
[
|
| 276720 |
"ββ",
|
| 276721 |
+
"βββββββββββ"
|
| 276722 |
],
|
| 276723 |
[
|
| 276724 |
+
"ββ",
|
| 276725 |
"β"
|
| 276726 |
],
|
| 276727 |
[
|
| 276728 |
+
"ββββ",
|
| 276729 |
"ββ"
|
| 276730 |
],
|
| 276731 |
[
|
| 276732 |
+
"ββββ",
|
| 276733 |
"ββββ"
|
| 276734 |
],
|
| 276735 |
[
|
| 276736 |
+
"ββββ",
|
| 276737 |
+
"ββββββββ"
|
| 276738 |
],
|
| 276739 |
[
|
| 276740 |
"ββββ",
|
| 276741 |
+
"βββββ"
|
| 276742 |
],
|
| 276743 |
[
|
| 276744 |
+
"ββββ",
|
| 276745 |
+
"ββββββ"
|
| 276746 |
],
|
| 276747 |
[
|
| 276748 |
+
"ββββ",
|
| 276749 |
+
"ββββββββββββ"
|
| 276750 |
],
|
| 276751 |
[
|
| 276752 |
+
"ββββ",
|
| 276753 |
"ββββββββββ"
|
| 276754 |
],
|
| 276755 |
[
|
| 276756 |
+
"ββββ",
|
| 276757 |
+
"βββ"
|
| 276758 |
],
|
| 276759 |
[
|
| 276760 |
+
"ββββ",
|
| 276761 |
+
"βββββββββ"
|
| 276762 |
],
|
| 276763 |
[
|
| 276764 |
+
"ββββ",
|
| 276765 |
+
"βββββββ"
|
| 276766 |
],
|
| 276767 |
[
|
| 276768 |
+
"ββββ",
|
| 276769 |
+
"βββββββββββ"
|
| 276770 |
],
|
| 276771 |
[
|
| 276772 |
+
"ββββ",
|
| 276773 |
+
"β"
|
| 276774 |
],
|
| 276775 |
[
|
| 276776 |
+
"ββββββββ",
|
| 276777 |
+
"ββ"
|
| 276778 |
],
|
| 276779 |
[
|
| 276780 |
+
"ββββββββ",
|
| 276781 |
+
"ββββ"
|
| 276782 |
],
|
| 276783 |
[
|
| 276784 |
+
"ββββββββ",
|
| 276785 |
+
"ββββββββ"
|
| 276786 |
+
],
|
| 276787 |
+
[
|
| 276788 |
+
"ββββββββ",
|
| 276789 |
"βββββ"
|
| 276790 |
],
|
| 276791 |
[
|
| 276792 |
+
"ββββββββ",
|
| 276793 |
+
"ββββββ"
|
| 276794 |
],
|
| 276795 |
[
|
| 276796 |
+
"ββββββββ",
|
| 276797 |
+
"βββ"
|
| 276798 |
],
|
| 276799 |
[
|
| 276800 |
+
"ββββββββ",
|
| 276801 |
+
"βββββββ"
|
| 276802 |
],
|
| 276803 |
[
|
| 276804 |
+
"ββββββββ",
|
| 276805 |
+
"β"
|
| 276806 |
+
],
|
| 276807 |
+
[
|
| 276808 |
+
"βββββ",
|
| 276809 |
"ββ"
|
| 276810 |
],
|
| 276811 |
[
|
| 276812 |
"βββββ",
|
| 276813 |
+
"ββββ"
|
| 276814 |
],
|
| 276815 |
[
|
| 276816 |
+
"βββββ",
|
| 276817 |
+
"ββββββββ"
|
| 276818 |
],
|
| 276819 |
[
|
| 276820 |
+
"βββββ",
|
| 276821 |
"βββββ"
|
| 276822 |
],
|
| 276823 |
[
|
| 276824 |
+
"βββββ",
|
| 276825 |
+
"ββββββ"
|
| 276826 |
+
],
|
| 276827 |
+
[
|
| 276828 |
+
"βββββ",
|
| 276829 |
"ββββββββββ"
|
| 276830 |
],
|
| 276831 |
[
|
| 276832 |
+
"βββββ",
|
| 276833 |
+
"βββ"
|
| 276834 |
],
|
| 276835 |
[
|
| 276836 |
+
"βββββ",
|
| 276837 |
+
"βββββββββ"
|
| 276838 |
],
|
| 276839 |
[
|
| 276840 |
"βββββ",
|
| 276841 |
"βββββββ"
|
| 276842 |
],
|
| 276843 |
[
|
| 276844 |
+
"βββββ",
|
| 276845 |
+
"βββββββββββ"
|
| 276846 |
],
|
| 276847 |
[
|
| 276848 |
+
"βββββ",
|
| 276849 |
+
"β"
|
| 276850 |
+
],
|
| 276851 |
+
[
|
| 276852 |
+
"ββββββ",
|
| 276853 |
"ββ"
|
| 276854 |
],
|
| 276855 |
[
|
| 276856 |
+
"ββββββ",
|
| 276857 |
+
"ββββ"
|
| 276858 |
],
|
| 276859 |
[
|
| 276860 |
+
"ββββββ",
|
| 276861 |
+
"ββββββββ"
|
| 276862 |
],
|
| 276863 |
[
|
| 276864 |
+
"ββββββ",
|
| 276865 |
"βββββ"
|
| 276866 |
],
|
| 276867 |
[
|
| 276868 |
+
"ββββββ",
|
| 276869 |
+
"ββββββ"
|
| 276870 |
],
|
| 276871 |
[
|
| 276872 |
+
"ββββββ",
|
| 276873 |
+
"ββββββββββ"
|
| 276874 |
],
|
| 276875 |
[
|
| 276876 |
+
"ββββββ",
|
| 276877 |
+
"βββ"
|
| 276878 |
],
|
| 276879 |
[
|
| 276880 |
+
"ββββββ",
|
| 276881 |
"βββββββββ"
|
| 276882 |
],
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276883 |
[
|
| 276884 |
"ββββββ",
|
| 276885 |
"βββββββ"
|
| 276886 |
],
|
| 276887 |
[
|
| 276888 |
+
"ββββββ",
|
| 276889 |
"β"
|
| 276890 |
],
|
| 276891 |
[
|
| 276892 |
+
"ββββββββββββ",
|
| 276893 |
+
"ββ"
|
| 276894 |
],
|
| 276895 |
[
|
| 276896 |
+
"ββββββββββββ",
|
| 276897 |
+
"ββββ"
|
| 276898 |
],
|
| 276899 |
[
|
| 276900 |
+
"ββββββββββββ",
|
| 276901 |
+
"βββ"
|
| 276902 |
],
|
| 276903 |
[
|
| 276904 |
+
"ββββββββββββ",
|
| 276905 |
+
"β"
|
| 276906 |
],
|
| 276907 |
[
|
| 276908 |
+
"βββββββββββββ",
|
| 276909 |
"ββ"
|
| 276910 |
],
|
| 276911 |
[
|
| 276912 |
+
"βββββββββββββ",
|
| 276913 |
+
"βββ"
|
| 276914 |
],
|
| 276915 |
[
|
| 276916 |
+
"βββββββββββββ",
|
| 276917 |
+
"β"
|
| 276918 |
],
|
| 276919 |
[
|
| 276920 |
+
"ββββββββββ",
|
| 276921 |
+
"ββ"
|
| 276922 |
],
|
| 276923 |
[
|
| 276924 |
+
"ββββββββββ",
|
| 276925 |
+
"ββββ"
|
| 276926 |
],
|
| 276927 |
[
|
| 276928 |
+
"ββββββββββ",
|
| 276929 |
"βββββ"
|
| 276930 |
],
|
| 276931 |
[
|
| 276932 |
+
"ββββββββββ",
|
| 276933 |
+
"ββββββ"
|
| 276934 |
],
|
| 276935 |
[
|
| 276936 |
+
"ββββββββββ",
|
| 276937 |
+
"βββ"
|
| 276938 |
],
|
| 276939 |
[
|
| 276940 |
+
"ββββββββββ",
|
| 276941 |
"β"
|
| 276942 |
],
|
| 276943 |
[
|
| 276944 |
+
"ββββββββββββββ",
|
| 276945 |
+
"ββ"
|
| 276946 |
],
|
| 276947 |
[
|
| 276948 |
+
"ββββββββββββββ",
|
| 276949 |
+
"β"
|
| 276950 |
],
|
| 276951 |
[
|
| 276952 |
+
"βββ",
|
| 276953 |
+
"ββ"
|
| 276954 |
],
|
| 276955 |
[
|
| 276956 |
+
"βββ",
|
| 276957 |
+
"ββββ"
|
| 276958 |
],
|
| 276959 |
[
|
| 276960 |
+
"βββ",
|
| 276961 |
+
"ββββββββ"
|
| 276962 |
],
|
| 276963 |
[
|
| 276964 |
+
"βββ",
|
| 276965 |
+
"βββββ"
|
| 276966 |
],
|
| 276967 |
[
|
| 276968 |
+
"βββ",
|
| 276969 |
+
"ββββββ"
|
| 276970 |
],
|
| 276971 |
[
|
| 276972 |
+
"βββ",
|
| 276973 |
+
"ββββββββββββ"
|
| 276974 |
],
|
| 276975 |
[
|
| 276976 |
+
"βββ",
|
| 276977 |
+
"βββββββββββββ"
|
| 276978 |
],
|
| 276979 |
[
|
| 276980 |
+
"βββ",
|
| 276981 |
+
"ββββββββββ"
|
| 276982 |
],
|
| 276983 |
[
|
| 276984 |
"βββ",
|
| 276985 |
+
"βββ"
|
| 276986 |
],
|
| 276987 |
[
|
| 276988 |
+
"βββ",
|
| 276989 |
+
"βββββββββ"
|
| 276990 |
],
|
| 276991 |
[
|
| 276992 |
+
"βββ",
|
| 276993 |
"βββββββ"
|
| 276994 |
],
|
| 276995 |
[
|
| 276996 |
+
"βββ",
|
| 276997 |
+
"βββββββββββ"
|
|
|
|
|
|
|
|
|
|
|
|
|
| 276998 |
],
|
| 276999 |
[
|
| 277000 |
+
"βββ",
|
| 277001 |
"β"
|
| 277002 |
],
|
| 277003 |
[
|
| 277004 |
+
"βββββββββ",
|
| 277005 |
"ββ"
|
| 277006 |
],
|
| 277007 |
[
|
| 277008 |
+
"βββββββββ",
|
| 277009 |
+
"ββββ"
|
| 277010 |
],
|
| 277011 |
[
|
| 277012 |
+
"βββββββββ",
|
| 277013 |
"βββββ"
|
| 277014 |
],
|
| 277015 |
[
|
| 277016 |
+
"βββββββββ",
|
| 277017 |
+
"ββββββ"
|
| 277018 |
],
|
| 277019 |
[
|
| 277020 |
+
"βββββββββ",
|
| 277021 |
+
"βββ"
|
| 277022 |
],
|
| 277023 |
[
|
| 277024 |
+
"βββββββββ",
|
| 277025 |
+
"βββββββ"
|
| 277026 |
],
|
| 277027 |
[
|
| 277028 |
+
"βββββββββ",
|
| 277029 |
+
"β"
|
| 277030 |
],
|
| 277031 |
[
|
| 277032 |
"βββββββ",
|
| 277033 |
"ββ"
|
| 277034 |
],
|
| 277035 |
[
|
| 277036 |
+
"βββββββ",
|
| 277037 |
+
"ββββ"
|
| 277038 |
],
|
| 277039 |
[
|
| 277040 |
+
"βββββββ",
|
| 277041 |
+
"ββββββββ"
|
| 277042 |
],
|
| 277043 |
[
|
| 277044 |
+
"βββββββ",
|
| 277045 |
+
"βββββ"
|
| 277046 |
],
|
| 277047 |
[
|
| 277048 |
+
"βββββββ",
|
| 277049 |
+
"ββββββ"
|
| 277050 |
],
|
| 277051 |
[
|
| 277052 |
+
"βββββββ",
|
| 277053 |
+
"βββ"
|
| 277054 |
],
|
| 277055 |
[
|
| 277056 |
+
"βββββββ",
|
| 277057 |
+
"βββββββββ"
|
| 277058 |
],
|
| 277059 |
[
|
| 277060 |
+
"βββββββ",
|
| 277061 |
+
"βββββββ"
|
| 277062 |
],
|
| 277063 |
[
|
| 277064 |
+
"βββββββ",
|
| 277065 |
+
"β"
|
| 277066 |
],
|
| 277067 |
[
|
| 277068 |
+
"βββββββββββ",
|
| 277069 |
+
"ββ"
|
| 277070 |
],
|
| 277071 |
[
|
| 277072 |
+
"βββββββββββ",
|
| 277073 |
+
"ββββ"
|
| 277074 |
],
|
| 277075 |
[
|
| 277076 |
+
"βββββββββββ",
|
| 277077 |
+
"βββββ"
|
| 277078 |
],
|
| 277079 |
[
|
| 277080 |
+
"βββββββββββ",
|
| 277081 |
+
"βββ"
|
| 277082 |
],
|
| 277083 |
[
|
| 277084 |
+
"βββββββββββ",
|
| 277085 |
"β"
|
| 277086 |
],
|
| 277087 |
[
|
| 277088 |
+
"βββββββββββββββ",
|
| 277089 |
+
"β"
|
| 277090 |
],
|
| 277091 |
[
|
| 277092 |
+
"β",
|
| 277093 |
"ββ"
|
| 277094 |
],
|
| 277095 |
[
|
| 277096 |
+
"β",
|
| 277097 |
"ββββ"
|
| 277098 |
],
|
| 277099 |
[
|
| 277100 |
"β",
|
| 277101 |
+
"ββββββββ"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 277102 |
],
|
| 277103 |
[
|
| 277104 |
+
"β",
|
| 277105 |
+
"βββββ"
|
| 277106 |
],
|
| 277107 |
[
|
| 277108 |
+
"β",
|
| 277109 |
+
"ββββββ"
|
| 277110 |
],
|
| 277111 |
[
|
| 277112 |
+
"β",
|
| 277113 |
+
"ββββββββββββ"
|
| 277114 |
],
|
| 277115 |
[
|
| 277116 |
+
"β",
|
| 277117 |
+
"βββββββββββββ"
|
| 277118 |
],
|
| 277119 |
[
|
| 277120 |
+
"β",
|
| 277121 |
+
"ββββββββββ"
|
| 277122 |
],
|
| 277123 |
[
|
| 277124 |
+
"β",
|
| 277125 |
+
"ββββββββββββββ"
|
| 277126 |
],
|
| 277127 |
[
|
| 277128 |
+
"β",
|
| 277129 |
+
"βββ"
|
| 277130 |
],
|
| 277131 |
[
|
| 277132 |
+
"β",
|
| 277133 |
+
"βββββββββ"
|
| 277134 |
],
|
| 277135 |
[
|
| 277136 |
+
"β",
|
| 277137 |
+
"βββββββ"
|
| 277138 |
],
|
| 277139 |
[
|
| 277140 |
+
"β",
|
| 277141 |
+
"βββββββββββ"
|
| 277142 |
],
|
| 277143 |
[
|
| 277144 |
+
"β",
|
| 277145 |
+
"βββββββββββββββ"
|
| 277146 |
],
|
| 277147 |
[
|
| 277148 |
"β",
|
| 277149 |
+
"β"
|
| 277150 |
]
|
| 277151 |
]
|
| 277152 |
}
|
tokenizer_config.json
CHANGED
|
@@ -6,7 +6,7 @@
|
|
| 6 |
"0": {
|
| 7 |
"content": "<unk>",
|
| 8 |
"lstrip": false,
|
| 9 |
-
"normalized":
|
| 10 |
"rstrip": false,
|
| 11 |
"single_word": false,
|
| 12 |
"special": true
|
|
@@ -14,7 +14,7 @@
|
|
| 14 |
"1": {
|
| 15 |
"content": "<s>",
|
| 16 |
"lstrip": false,
|
| 17 |
-
"normalized":
|
| 18 |
"rstrip": false,
|
| 19 |
"single_word": false,
|
| 20 |
"special": true
|
|
@@ -22,21 +22,20 @@
|
|
| 22 |
"2": {
|
| 23 |
"content": "</s>",
|
| 24 |
"lstrip": false,
|
| 25 |
-
"normalized":
|
| 26 |
"rstrip": false,
|
| 27 |
"single_word": false,
|
| 28 |
"special": true
|
| 29 |
}
|
| 30 |
},
|
| 31 |
"bos_token": "<s>",
|
| 32 |
-
"chat_template": "{% for message in
|
| 33 |
"clean_up_tokenization_spaces": false,
|
| 34 |
"eos_token": "</s>",
|
| 35 |
"extra_special_tokens": {},
|
| 36 |
-
"legacy":
|
| 37 |
"model_max_length": 2048,
|
| 38 |
"pad_token": "</s>",
|
| 39 |
-
"padding_side": "right",
|
| 40 |
"sp_model_kwargs": {},
|
| 41 |
"tokenizer_class": "LlamaTokenizer",
|
| 42 |
"unk_token": "<unk>",
|
|
|
|
| 6 |
"0": {
|
| 7 |
"content": "<unk>",
|
| 8 |
"lstrip": false,
|
| 9 |
+
"normalized": true,
|
| 10 |
"rstrip": false,
|
| 11 |
"single_word": false,
|
| 12 |
"special": true
|
|
|
|
| 14 |
"1": {
|
| 15 |
"content": "<s>",
|
| 16 |
"lstrip": false,
|
| 17 |
+
"normalized": true,
|
| 18 |
"rstrip": false,
|
| 19 |
"single_word": false,
|
| 20 |
"special": true
|
|
|
|
| 22 |
"2": {
|
| 23 |
"content": "</s>",
|
| 24 |
"lstrip": false,
|
| 25 |
+
"normalized": true,
|
| 26 |
"rstrip": false,
|
| 27 |
"single_word": false,
|
| 28 |
"special": true
|
| 29 |
}
|
| 30 |
},
|
| 31 |
"bos_token": "<s>",
|
| 32 |
+
"chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% set system_message = false %}{% endif %}{% for message in loop_messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if loop.index0 == 0 and system_message != false %}{% set content = '<<SYS>>\\n' + system_message + '\\n<</SYS>>\\n\\n' + message['content'] %}{% else %}{% set content = message['content'] %}{% endif %}{% if message['role'] == 'user' %}{{ bos_token + '[INST] ' + content.strip() + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + content.strip() + ' ' + eos_token }}{% endif %}{% endfor %}",
|
| 33 |
"clean_up_tokenization_spaces": false,
|
| 34 |
"eos_token": "</s>",
|
| 35 |
"extra_special_tokens": {},
|
| 36 |
+
"legacy": true,
|
| 37 |
"model_max_length": 2048,
|
| 38 |
"pad_token": "</s>",
|
|
|
|
| 39 |
"sp_model_kwargs": {},
|
| 40 |
"tokenizer_class": "LlamaTokenizer",
|
| 41 |
"unk_token": "<unk>",
|