Sri Santh commited on
Commit
6eb46c5
·
verified ·
1 Parent(s): c125643

Upload 16 files

Browse files
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: mistralai/Mistral-7B-Instruct-v0.1
3
+ library_name: peft
4
+ license: other
5
+ tags:
6
+ - llama-factory
7
+ - lora
8
+ - generated_from_trainer
9
+ model-index:
10
+ - name: mistral_docker
11
+ results: []
12
+ ---
13
+
14
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
15
+ should probably proofread and complete it, then remove this comment. -->
16
+
17
+ # mistral_docker
18
+
19
+ This model is a fine-tuned version of [mistralai/Mistral-7B-Instruct-v0.1](https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1) on the minecraft dataset.
20
+
21
+ ## Model description
22
+
23
+ More information needed
24
+
25
+ ## Intended uses & limitations
26
+
27
+ More information needed
28
+
29
+ ## Training and evaluation data
30
+
31
+ More information needed
32
+
33
+ ## Training procedure
34
+
35
+ ### Training hyperparameters
36
+
37
+ The following hyperparameters were used during training:
38
+ - learning_rate: 5e-05
39
+ - train_batch_size: 2
40
+ - eval_batch_size: 8
41
+ - seed: 42
42
+ - gradient_accumulation_steps: 8
43
+ - total_train_batch_size: 16
44
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
+ - lr_scheduler_type: cosine
46
+ - num_epochs: 3.0
47
+
48
+ ### Training results
49
+
50
+
51
+
52
+ ### Framework versions
53
+
54
+ - PEFT 0.12.0
55
+ - Transformers 4.43.4
56
+ - Pytorch 2.2.1
57
+ - Datasets 2.21.0
58
+ - Tokenizers 0.19.1
adapter_config.json ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
5
+ "bias": "none",
6
+ "fan_in_fan_out": false,
7
+ "inference_mode": true,
8
+ "init_lora_weights": true,
9
+ "layer_replication": null,
10
+ "layers_pattern": null,
11
+ "layers_to_transform": null,
12
+ "loftq_config": {},
13
+ "lora_alpha": 16,
14
+ "lora_dropout": 0,
15
+ "megatron_config": null,
16
+ "megatron_core": "megatron.core",
17
+ "modules_to_save": null,
18
+ "peft_type": "LORA",
19
+ "r": 8,
20
+ "rank_pattern": {},
21
+ "revision": null,
22
+ "target_modules": [
23
+ "gate_proj",
24
+ "v_proj",
25
+ "up_proj",
26
+ "k_proj",
27
+ "q_proj",
28
+ "o_proj",
29
+ "down_proj"
30
+ ],
31
+ "task_type": "CAUSAL_LM",
32
+ "use_dora": false,
33
+ "use_rslora": false
34
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afb161c26ed8b4feebed547a41efca0e1e9a87608e30b343d6e6312fa5c662c6
3
+ size 83945296
all_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "num_input_tokens_seen": 437216,
4
+ "total_flos": 1.8708380691726336e+16,
5
+ "train_loss": 0.044493223137528556,
6
+ "train_runtime": 988.3228,
7
+ "train_samples_per_second": 7.331,
8
+ "train_steps_per_second": 0.458
9
+ }
llamaboard_config.yaml ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ top.booster: auto
2
+ top.checkpoint_path: []
3
+ top.finetuning_type: lora
4
+ top.model_name: Mistral-7B-v0.1-Chat
5
+ top.quantization_bit: none
6
+ top.quantization_method: bitsandbytes
7
+ top.rope_scaling: none
8
+ top.template: mistral
9
+ top.visual_inputs: false
10
+ train.additional_target: ''
11
+ train.badam_mode: layer
12
+ train.badam_switch_interval: 50
13
+ train.badam_switch_mode: ascending
14
+ train.badam_update_ratio: 0.05
15
+ train.batch_size: 2
16
+ train.compute_type: bf16
17
+ train.create_new_adapter: false
18
+ train.cutoff_len: 1024
19
+ train.dataset:
20
+ - minecraft
21
+ train.dataset_dir: data
22
+ train.ds_offload: false
23
+ train.ds_stage: none
24
+ train.freeze_extra_modules: ''
25
+ train.freeze_trainable_layers: 2
26
+ train.freeze_trainable_modules: all
27
+ train.galore_rank: 16
28
+ train.galore_scale: 0.25
29
+ train.galore_target: all
30
+ train.galore_update_interval: 200
31
+ train.gradient_accumulation_steps: 8
32
+ train.learning_rate: 5e-5
33
+ train.logging_steps: 5
34
+ train.lora_alpha: 16
35
+ train.lora_dropout: 0
36
+ train.lora_rank: 8
37
+ train.lora_target: ''
38
+ train.loraplus_lr_ratio: 0
39
+ train.lr_scheduler_type: cosine
40
+ train.mask_history: false
41
+ train.max_grad_norm: '1.0'
42
+ train.max_samples: '100000'
43
+ train.neat_packing: false
44
+ train.neftune_alpha: 0
45
+ train.num_train_epochs: '3.0'
46
+ train.optim: adamw_torch
47
+ train.packing: false
48
+ train.ppo_score_norm: false
49
+ train.ppo_whiten_rewards: false
50
+ train.pref_beta: 0.1
51
+ train.pref_ftx: 0
52
+ train.pref_loss: sigmoid
53
+ train.report_to: false
54
+ train.resize_vocab: false
55
+ train.reward_model: null
56
+ train.save_steps: 100
57
+ train.shift_attn: false
58
+ train.train_on_prompt: false
59
+ train.training_stage: Supervised Fine-Tuning
60
+ train.use_badam: false
61
+ train.use_dora: false
62
+ train.use_galore: false
63
+ train.use_llama_pro: false
64
+ train.use_pissa: false
65
+ train.use_rslora: false
66
+ train.val_size: 0
67
+ train.warmup_steps: 0
running_log.txt ADDED
@@ -0,0 +1,510 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [INFO|parser.py:355] 2024-08-29 19:20:42,528 >> Process rank: 0, device: cuda:0, n_gpu: 1, distributed training: False, compute dtype: torch.bfloat16
2
+
3
+ [INFO|tokenization_utils_base.py:2289] 2024-08-29 19:20:43,824 >> loading file tokenizer.model from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/tokenizer.model
4
+
5
+ [INFO|tokenization_utils_base.py:2289] 2024-08-29 19:20:43,824 >> loading file tokenizer.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/tokenizer.json
6
+
7
+ [INFO|tokenization_utils_base.py:2289] 2024-08-29 19:20:43,824 >> loading file added_tokens.json from cache at None
8
+
9
+ [INFO|tokenization_utils_base.py:2289] 2024-08-29 19:20:43,824 >> loading file special_tokens_map.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/special_tokens_map.json
10
+
11
+ [INFO|tokenization_utils_base.py:2289] 2024-08-29 19:20:43,825 >> loading file tokenizer_config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/tokenizer_config.json
12
+
13
+ [INFO|template.py:373] 2024-08-29 19:20:43,869 >> Add pad token: </s>
14
+
15
+ [INFO|loader.py:52] 2024-08-29 19:20:43,870 >> Loading dataset MattCoddity/dockerNLcommands...
16
+
17
+ [INFO|configuration_utils.py:733] 2024-08-29 19:20:46,934 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
18
+
19
+ [INFO|configuration_utils.py:800] 2024-08-29 19:20:46,935 >> Model config MistralConfig {
20
+ "_name_or_path": "mistralai/Mistral-7B-Instruct-v0.1",
21
+ "architectures": [
22
+ "MistralForCausalLM"
23
+ ],
24
+ "attention_dropout": 0.0,
25
+ "bos_token_id": 1,
26
+ "eos_token_id": 2,
27
+ "head_dim": 128,
28
+ "hidden_act": "silu",
29
+ "hidden_size": 4096,
30
+ "initializer_range": 0.02,
31
+ "intermediate_size": 14336,
32
+ "max_position_embeddings": 32768,
33
+ "model_type": "mistral",
34
+ "num_attention_heads": 32,
35
+ "num_hidden_layers": 32,
36
+ "num_key_value_heads": 8,
37
+ "rms_norm_eps": 1e-05,
38
+ "rope_theta": 10000.0,
39
+ "sliding_window": 4096,
40
+ "tie_word_embeddings": false,
41
+ "torch_dtype": "bfloat16",
42
+ "transformers_version": "4.43.4",
43
+ "use_cache": true,
44
+ "vocab_size": 32000
45
+ }
46
+
47
+
48
+ [INFO|modeling_utils.py:3644] 2024-08-29 19:20:47,387 >> loading weights file model.safetensors from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/model.safetensors.index.json
49
+
50
+ [INFO|modeling_utils.py:1572] 2024-08-29 19:22:01,945 >> Instantiating MistralForCausalLM model under default dtype torch.bfloat16.
51
+
52
+ [INFO|configuration_utils.py:1038] 2024-08-29 19:22:01,946 >> Generate config GenerationConfig {
53
+ "bos_token_id": 1,
54
+ "eos_token_id": 2
55
+ }
56
+
57
+
58
+ [INFO|modeling_utils.py:4473] 2024-08-29 19:22:05,773 >> All model checkpoint weights were used when initializing MistralForCausalLM.
59
+
60
+
61
+ [INFO|modeling_utils.py:4481] 2024-08-29 19:22:05,773 >> All the weights of MistralForCausalLM were initialized from the model checkpoint at mistralai/Mistral-7B-Instruct-v0.1.
62
+ If your task is similar to the task the model of the checkpoint was trained on, you can already use MistralForCausalLM for predictions without further training.
63
+
64
+ [INFO|configuration_utils.py:993] 2024-08-29 19:22:06,005 >> loading configuration file generation_config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/generation_config.json
65
+
66
+ [INFO|configuration_utils.py:1038] 2024-08-29 19:22:06,005 >> Generate config GenerationConfig {
67
+ "bos_token_id": 1,
68
+ "eos_token_id": 2
69
+ }
70
+
71
+
72
+ [INFO|checkpointing.py:103] 2024-08-29 19:22:06,012 >> Gradient checkpointing enabled.
73
+
74
+ [INFO|attention.py:84] 2024-08-29 19:22:06,012 >> Using torch SDPA for faster training and inference.
75
+
76
+ [INFO|adapter.py:302] 2024-08-29 19:22:06,012 >> Upcasting trainable params to float32.
77
+
78
+ [INFO|adapter.py:158] 2024-08-29 19:22:06,012 >> Fine-tuning method: LoRA
79
+
80
+ [INFO|misc.py:51] 2024-08-29 19:22:06,013 >> Found linear modules: gate_proj,v_proj,up_proj,k_proj,q_proj,o_proj,down_proj
81
+
82
+ [INFO|loader.py:196] 2024-08-29 19:22:21,642 >> trainable params: 20,971,520 || all params: 7,262,703,616 || trainable%: 0.2888
83
+
84
+ [INFO|trainer.py:648] 2024-08-29 19:22:21,652 >> Using auto half precision backend
85
+
86
+ [INFO|trainer.py:2134] 2024-08-29 19:22:22,077 >> ***** Running training *****
87
+
88
+ [INFO|trainer.py:2135] 2024-08-29 19:22:22,077 >> Num examples = 2,415
89
+
90
+ [INFO|trainer.py:2136] 2024-08-29 19:22:22,077 >> Num Epochs = 3
91
+
92
+ [INFO|trainer.py:2137] 2024-08-29 19:22:22,077 >> Instantaneous batch size per device = 2
93
+
94
+ [INFO|trainer.py:2140] 2024-08-29 19:22:22,077 >> Total train batch size (w. parallel, distributed & accumulation) = 16
95
+
96
+ [INFO|trainer.py:2141] 2024-08-29 19:22:22,078 >> Gradient Accumulation steps = 8
97
+
98
+ [INFO|trainer.py:2142] 2024-08-29 19:22:22,078 >> Total optimization steps = 453
99
+
100
+ [INFO|trainer.py:2143] 2024-08-29 19:22:22,081 >> Number of trainable parameters = 20,971,520
101
+
102
+ [INFO|callbacks.py:319] 2024-08-29 19:22:33,541 >> {'loss': 0.6997, 'learning_rate': 4.9985e-05, 'epoch': 0.03, 'throughput': 428.77}
103
+
104
+ [INFO|callbacks.py:319] 2024-08-29 19:22:44,569 >> {'loss': 0.3509, 'learning_rate': 4.9940e-05, 'epoch': 0.07, 'throughput': 441.91}
105
+
106
+ [INFO|callbacks.py:319] 2024-08-29 19:22:55,574 >> {'loss': 0.2853, 'learning_rate': 4.9865e-05, 'epoch': 0.10, 'throughput': 443.85}
107
+
108
+ [INFO|callbacks.py:319] 2024-08-29 19:23:06,256 >> {'loss': 0.2456, 'learning_rate': 4.9760e-05, 'epoch': 0.13, 'throughput': 448.8}
109
+
110
+ [INFO|callbacks.py:319] 2024-08-29 19:23:17,193 >> {'loss': 0.2084, 'learning_rate': 4.9625e-05, 'epoch': 0.17, 'throughput': 444.22}
111
+
112
+ [INFO|callbacks.py:319] 2024-08-29 19:23:28,050 >> {'loss': 0.1626, 'learning_rate': 4.9461e-05, 'epoch': 0.20, 'throughput': 444.12}
113
+
114
+ [INFO|callbacks.py:319] 2024-08-29 19:23:38,739 >> {'loss': 0.1405, 'learning_rate': 4.9267e-05, 'epoch': 0.23, 'throughput': 442.72}
115
+
116
+ [INFO|callbacks.py:319] 2024-08-29 19:23:49,936 >> {'loss': 0.1182, 'learning_rate': 4.9044e-05, 'epoch': 0.26, 'throughput': 444.57}
117
+
118
+ [INFO|callbacks.py:319] 2024-08-29 19:24:01,393 >> {'loss': 0.1285, 'learning_rate': 4.8792e-05, 'epoch': 0.30, 'throughput': 438.23}
119
+
120
+ [INFO|callbacks.py:319] 2024-08-29 19:24:12,547 >> {'loss': 0.0906, 'learning_rate': 4.8512e-05, 'epoch': 0.33, 'throughput': 440.48}
121
+
122
+ [INFO|callbacks.py:319] 2024-08-29 19:24:23,416 >> {'loss': 0.0540, 'learning_rate': 4.8203e-05, 'epoch': 0.36, 'throughput': 441.77}
123
+
124
+ [INFO|callbacks.py:319] 2024-08-29 19:24:34,155 >> {'loss': 0.0730, 'learning_rate': 4.7867e-05, 'epoch': 0.40, 'throughput': 442.07}
125
+
126
+ [INFO|callbacks.py:319] 2024-08-29 19:24:44,702 >> {'loss': 0.0524, 'learning_rate': 4.7503e-05, 'epoch': 0.43, 'throughput': 442.81}
127
+
128
+ [INFO|callbacks.py:319] 2024-08-29 19:24:55,348 >> {'loss': 0.0534, 'learning_rate': 4.7112e-05, 'epoch': 0.46, 'throughput': 442.95}
129
+
130
+ [INFO|callbacks.py:319] 2024-08-29 19:25:06,024 >> {'loss': 0.0759, 'learning_rate': 4.6694e-05, 'epoch': 0.50, 'throughput': 442.11}
131
+
132
+ [INFO|callbacks.py:319] 2024-08-29 19:25:16,998 >> {'loss': 0.0766, 'learning_rate': 4.6250e-05, 'epoch': 0.53, 'throughput': 440.9}
133
+
134
+ [INFO|callbacks.py:319] 2024-08-29 19:25:28,034 >> {'loss': 0.0383, 'learning_rate': 4.5781e-05, 'epoch': 0.56, 'throughput': 438.74}
135
+
136
+ [INFO|callbacks.py:319] 2024-08-29 19:25:38,775 >> {'loss': 0.0393, 'learning_rate': 4.5286e-05, 'epoch': 0.60, 'throughput': 439.35}
137
+
138
+ [INFO|callbacks.py:319] 2024-08-29 19:25:49,280 >> {'loss': 0.0589, 'learning_rate': 4.4768e-05, 'epoch': 0.63, 'throughput': 440.4}
139
+
140
+ [INFO|callbacks.py:319] 2024-08-29 19:26:00,073 >> {'loss': 0.0351, 'learning_rate': 4.4225e-05, 'epoch': 0.66, 'throughput': 440.1}
141
+
142
+ [INFO|trainer.py:3503] 2024-08-29 19:26:00,082 >> Saving model checkpoint to saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-100
143
+
144
+ [INFO|configuration_utils.py:733] 2024-08-29 19:26:00,312 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
145
+
146
+ [INFO|configuration_utils.py:800] 2024-08-29 19:26:00,313 >> Model config MistralConfig {
147
+ "architectures": [
148
+ "MistralForCausalLM"
149
+ ],
150
+ "attention_dropout": 0.0,
151
+ "bos_token_id": 1,
152
+ "eos_token_id": 2,
153
+ "head_dim": 128,
154
+ "hidden_act": "silu",
155
+ "hidden_size": 4096,
156
+ "initializer_range": 0.02,
157
+ "intermediate_size": 14336,
158
+ "max_position_embeddings": 32768,
159
+ "model_type": "mistral",
160
+ "num_attention_heads": 32,
161
+ "num_hidden_layers": 32,
162
+ "num_key_value_heads": 8,
163
+ "rms_norm_eps": 1e-05,
164
+ "rope_theta": 10000.0,
165
+ "sliding_window": 4096,
166
+ "tie_word_embeddings": false,
167
+ "torch_dtype": "bfloat16",
168
+ "transformers_version": "4.43.4",
169
+ "use_cache": true,
170
+ "vocab_size": 32000
171
+ }
172
+
173
+
174
+ [INFO|tokenization_utils_base.py:2702] 2024-08-29 19:26:00,409 >> tokenizer config file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-100/tokenizer_config.json
175
+
176
+ [INFO|tokenization_utils_base.py:2711] 2024-08-29 19:26:00,409 >> Special tokens file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-100/special_tokens_map.json
177
+
178
+ [INFO|callbacks.py:319] 2024-08-29 19:26:11,158 >> {'loss': 0.0417, 'learning_rate': 4.3660e-05, 'epoch': 0.70, 'throughput': 439.61}
179
+
180
+ [INFO|callbacks.py:319] 2024-08-29 19:26:22,031 >> {'loss': 0.0437, 'learning_rate': 4.3072e-05, 'epoch': 0.73, 'throughput': 439.17}
181
+
182
+ [INFO|callbacks.py:319] 2024-08-29 19:26:33,043 >> {'loss': 0.0498, 'learning_rate': 4.2462e-05, 'epoch': 0.76, 'throughput': 439.15}
183
+
184
+ [INFO|callbacks.py:319] 2024-08-29 19:26:43,782 >> {'loss': 0.0239, 'learning_rate': 4.1831e-05, 'epoch': 0.79, 'throughput': 438.31}
185
+
186
+ [INFO|callbacks.py:319] 2024-08-29 19:26:54,491 >> {'loss': 0.0228, 'learning_rate': 4.1180e-05, 'epoch': 0.83, 'throughput': 439.58}
187
+
188
+ [INFO|callbacks.py:319] 2024-08-29 19:27:05,490 >> {'loss': 0.0390, 'learning_rate': 4.0510e-05, 'epoch': 0.86, 'throughput': 440.13}
189
+
190
+ [INFO|callbacks.py:319] 2024-08-29 19:27:16,280 >> {'loss': 0.0500, 'learning_rate': 3.9821e-05, 'epoch': 0.89, 'throughput': 440.36}
191
+
192
+ [INFO|callbacks.py:319] 2024-08-29 19:27:26,762 >> {'loss': 0.0200, 'learning_rate': 3.9114e-05, 'epoch': 0.93, 'throughput': 441.12}
193
+
194
+ [INFO|callbacks.py:319] 2024-08-29 19:27:37,562 >> {'loss': 0.0159, 'learning_rate': 3.8390e-05, 'epoch': 0.96, 'throughput': 441.95}
195
+
196
+ [INFO|callbacks.py:319] 2024-08-29 19:27:48,075 >> {'loss': 0.0170, 'learning_rate': 3.7650e-05, 'epoch': 0.99, 'throughput': 442.86}
197
+
198
+ [INFO|callbacks.py:319] 2024-08-29 19:27:58,470 >> {'loss': 0.0298, 'learning_rate': 3.6895e-05, 'epoch': 1.03, 'throughput': 443.06}
199
+
200
+ [INFO|callbacks.py:319] 2024-08-29 19:28:09,021 >> {'loss': 0.0357, 'learning_rate': 3.6125e-05, 'epoch': 1.06, 'throughput': 443.05}
201
+
202
+ [INFO|callbacks.py:319] 2024-08-29 19:28:19,371 >> {'loss': 0.0185, 'learning_rate': 3.5342e-05, 'epoch': 1.09, 'throughput': 444.01}
203
+
204
+ [INFO|callbacks.py:319] 2024-08-29 19:28:30,281 >> {'loss': 0.0339, 'learning_rate': 3.4547e-05, 'epoch': 1.13, 'throughput': 444.2}
205
+
206
+ [INFO|callbacks.py:319] 2024-08-29 19:28:41,235 >> {'loss': 0.0224, 'learning_rate': 3.3740e-05, 'epoch': 1.16, 'throughput': 444.02}
207
+
208
+ [INFO|callbacks.py:319] 2024-08-29 19:28:51,823 >> {'loss': 0.0192, 'learning_rate': 3.2923e-05, 'epoch': 1.19, 'throughput': 444.69}
209
+
210
+ [INFO|callbacks.py:319] 2024-08-29 19:29:02,724 >> {'loss': 0.0275, 'learning_rate': 3.2096e-05, 'epoch': 1.23, 'throughput': 444.69}
211
+
212
+ [INFO|callbacks.py:319] 2024-08-29 19:29:13,570 >> {'loss': 0.0086, 'learning_rate': 3.1261e-05, 'epoch': 1.26, 'throughput': 444.59}
213
+
214
+ [INFO|callbacks.py:319] 2024-08-29 19:29:24,314 >> {'loss': 0.0166, 'learning_rate': 3.0418e-05, 'epoch': 1.29, 'throughput': 445.07}
215
+
216
+ [INFO|callbacks.py:319] 2024-08-29 19:29:35,290 >> {'loss': 0.0167, 'learning_rate': 2.9569e-05, 'epoch': 1.32, 'throughput': 445.5}
217
+
218
+ [INFO|trainer.py:3503] 2024-08-29 19:29:35,291 >> Saving model checkpoint to saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-200
219
+
220
+ [INFO|configuration_utils.py:733] 2024-08-29 19:29:35,516 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
221
+
222
+ [INFO|configuration_utils.py:800] 2024-08-29 19:29:35,516 >> Model config MistralConfig {
223
+ "architectures": [
224
+ "MistralForCausalLM"
225
+ ],
226
+ "attention_dropout": 0.0,
227
+ "bos_token_id": 1,
228
+ "eos_token_id": 2,
229
+ "head_dim": 128,
230
+ "hidden_act": "silu",
231
+ "hidden_size": 4096,
232
+ "initializer_range": 0.02,
233
+ "intermediate_size": 14336,
234
+ "max_position_embeddings": 32768,
235
+ "model_type": "mistral",
236
+ "num_attention_heads": 32,
237
+ "num_hidden_layers": 32,
238
+ "num_key_value_heads": 8,
239
+ "rms_norm_eps": 1e-05,
240
+ "rope_theta": 10000.0,
241
+ "sliding_window": 4096,
242
+ "tie_word_embeddings": false,
243
+ "torch_dtype": "bfloat16",
244
+ "transformers_version": "4.43.4",
245
+ "use_cache": true,
246
+ "vocab_size": 32000
247
+ }
248
+
249
+
250
+ [INFO|tokenization_utils_base.py:2702] 2024-08-29 19:29:35,602 >> tokenizer config file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-200/tokenizer_config.json
251
+
252
+ [INFO|tokenization_utils_base.py:2711] 2024-08-29 19:29:35,602 >> Special tokens file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-200/special_tokens_map.json
253
+
254
+ [INFO|callbacks.py:319] 2024-08-29 19:29:46,672 >> {'loss': 0.0137, 'learning_rate': 2.8714e-05, 'epoch': 1.36, 'throughput': 444.49}
255
+
256
+ [INFO|callbacks.py:319] 2024-08-29 19:29:57,486 >> {'loss': 0.0142, 'learning_rate': 2.7854e-05, 'epoch': 1.39, 'throughput': 444.44}
257
+
258
+ [INFO|callbacks.py:319] 2024-08-29 19:30:08,468 >> {'loss': 0.0136, 'learning_rate': 2.6992e-05, 'epoch': 1.42, 'throughput': 444.3}
259
+
260
+ [INFO|callbacks.py:319] 2024-08-29 19:30:19,208 >> {'loss': 0.0184, 'learning_rate': 2.6127e-05, 'epoch': 1.46, 'throughput': 443.79}
261
+
262
+ [INFO|callbacks.py:319] 2024-08-29 19:30:30,135 >> {'loss': 0.0098, 'learning_rate': 2.5260e-05, 'epoch': 1.49, 'throughput': 443.82}
263
+
264
+ [INFO|callbacks.py:319] 2024-08-29 19:30:40,756 >> {'loss': 0.0120, 'learning_rate': 2.4393e-05, 'epoch': 1.52, 'throughput': 444.28}
265
+
266
+ [INFO|callbacks.py:319] 2024-08-29 19:30:51,882 >> {'loss': 0.0106, 'learning_rate': 2.3527e-05, 'epoch': 1.56, 'throughput': 443.85}
267
+
268
+ [INFO|callbacks.py:319] 2024-08-29 19:31:02,845 >> {'loss': 0.0128, 'learning_rate': 2.2663e-05, 'epoch': 1.59, 'throughput': 443.72}
269
+
270
+ [INFO|callbacks.py:319] 2024-08-29 19:31:13,810 >> {'loss': 0.0139, 'learning_rate': 2.1801e-05, 'epoch': 1.62, 'throughput': 443.33}
271
+
272
+ [INFO|callbacks.py:319] 2024-08-29 19:31:24,525 >> {'loss': 0.0084, 'learning_rate': 2.0944e-05, 'epoch': 1.66, 'throughput': 443.48}
273
+
274
+ [INFO|callbacks.py:319] 2024-08-29 19:31:35,394 >> {'loss': 0.0143, 'learning_rate': 2.0091e-05, 'epoch': 1.69, 'throughput': 443.53}
275
+
276
+ [INFO|callbacks.py:319] 2024-08-29 19:31:46,123 >> {'loss': 0.0110, 'learning_rate': 1.9244e-05, 'epoch': 1.72, 'throughput': 443.94}
277
+
278
+ [INFO|callbacks.py:319] 2024-08-29 19:31:57,253 >> {'loss': 0.0127, 'learning_rate': 1.8404e-05, 'epoch': 1.75, 'throughput': 443.72}
279
+
280
+ [INFO|callbacks.py:319] 2024-08-29 19:32:07,949 >> {'loss': 0.0127, 'learning_rate': 1.7572e-05, 'epoch': 1.79, 'throughput': 443.93}
281
+
282
+ [INFO|callbacks.py:319] 2024-08-29 19:32:18,963 >> {'loss': 0.0122, 'learning_rate': 1.6749e-05, 'epoch': 1.82, 'throughput': 443.72}
283
+
284
+ [INFO|callbacks.py:319] 2024-08-29 19:32:29,997 >> {'loss': 0.0169, 'learning_rate': 1.5936e-05, 'epoch': 1.85, 'throughput': 443.77}
285
+
286
+ [INFO|callbacks.py:319] 2024-08-29 19:32:40,825 >> {'loss': 0.0242, 'learning_rate': 1.5133e-05, 'epoch': 1.89, 'throughput': 444.05}
287
+
288
+ [INFO|callbacks.py:319] 2024-08-29 19:32:51,740 >> {'loss': 0.0079, 'learning_rate': 1.4343e-05, 'epoch': 1.92, 'throughput': 443.8}
289
+
290
+ [INFO|callbacks.py:319] 2024-08-29 19:33:03,290 >> {'loss': 0.0194, 'learning_rate': 1.3565e-05, 'epoch': 1.95, 'throughput': 443.36}
291
+
292
+ [INFO|callbacks.py:319] 2024-08-29 19:33:14,813 >> {'loss': 0.0133, 'learning_rate': 1.2801e-05, 'epoch': 1.99, 'throughput': 443.09}
293
+
294
+ [INFO|trainer.py:3503] 2024-08-29 19:33:14,814 >> Saving model checkpoint to saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-300
295
+
296
+ [INFO|configuration_utils.py:733] 2024-08-29 19:33:15,036 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
297
+
298
+ [INFO|configuration_utils.py:800] 2024-08-29 19:33:15,037 >> Model config MistralConfig {
299
+ "architectures": [
300
+ "MistralForCausalLM"
301
+ ],
302
+ "attention_dropout": 0.0,
303
+ "bos_token_id": 1,
304
+ "eos_token_id": 2,
305
+ "head_dim": 128,
306
+ "hidden_act": "silu",
307
+ "hidden_size": 4096,
308
+ "initializer_range": 0.02,
309
+ "intermediate_size": 14336,
310
+ "max_position_embeddings": 32768,
311
+ "model_type": "mistral",
312
+ "num_attention_heads": 32,
313
+ "num_hidden_layers": 32,
314
+ "num_key_value_heads": 8,
315
+ "rms_norm_eps": 1e-05,
316
+ "rope_theta": 10000.0,
317
+ "sliding_window": 4096,
318
+ "tie_word_embeddings": false,
319
+ "torch_dtype": "bfloat16",
320
+ "transformers_version": "4.43.4",
321
+ "use_cache": true,
322
+ "vocab_size": 32000
323
+ }
324
+
325
+
326
+ [INFO|tokenization_utils_base.py:2702] 2024-08-29 19:33:15,117 >> tokenizer config file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-300/tokenizer_config.json
327
+
328
+ [INFO|tokenization_utils_base.py:2711] 2024-08-29 19:33:15,117 >> Special tokens file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-300/special_tokens_map.json
329
+
330
+ [INFO|callbacks.py:319] 2024-08-29 19:33:26,589 >> {'loss': 0.0049, 'learning_rate': 1.2052e-05, 'epoch': 2.02, 'throughput': 442.42}
331
+
332
+ [INFO|callbacks.py:319] 2024-08-29 19:33:37,297 >> {'loss': 0.0121, 'learning_rate': 1.1319e-05, 'epoch': 2.05, 'throughput': 442.9}
333
+
334
+ [INFO|callbacks.py:319] 2024-08-29 19:33:47,964 >> {'loss': 0.0102, 'learning_rate': 1.0601e-05, 'epoch': 2.09, 'throughput': 442.89}
335
+
336
+ [INFO|callbacks.py:319] 2024-08-29 19:33:59,114 >> {'loss': 0.0072, 'learning_rate': 9.9016e-06, 'epoch': 2.12, 'throughput': 443.13}
337
+
338
+ [INFO|callbacks.py:319] 2024-08-29 19:34:10,116 >> {'loss': 0.0066, 'learning_rate': 9.2199e-06, 'epoch': 2.15, 'throughput': 443.2}
339
+
340
+ [INFO|callbacks.py:319] 2024-08-29 19:34:21,353 >> {'loss': 0.0119, 'learning_rate': 8.5571e-06, 'epoch': 2.19, 'throughput': 442.86}
341
+
342
+ [INFO|callbacks.py:319] 2024-08-29 19:34:32,508 >> {'loss': 0.0057, 'learning_rate': 7.9141e-06, 'epoch': 2.22, 'throughput': 442.91}
343
+
344
+ [INFO|callbacks.py:319] 2024-08-29 19:34:43,266 >> {'loss': 0.0136, 'learning_rate': 7.2917e-06, 'epoch': 2.25, 'throughput': 443.28}
345
+
346
+ [INFO|callbacks.py:319] 2024-08-29 19:34:54,105 >> {'loss': 0.0078, 'learning_rate': 6.6906e-06, 'epoch': 2.28, 'throughput': 443.17}
347
+
348
+ [INFO|callbacks.py:319] 2024-08-29 19:35:05,306 >> {'loss': 0.0141, 'learning_rate': 6.1114e-06, 'epoch': 2.32, 'throughput': 442.62}
349
+
350
+ [INFO|callbacks.py:319] 2024-08-29 19:35:16,627 >> {'loss': 0.0140, 'learning_rate': 5.5550e-06, 'epoch': 2.35, 'throughput': 442.57}
351
+
352
+ [INFO|callbacks.py:319] 2024-08-29 19:35:27,909 >> {'loss': 0.0045, 'learning_rate': 5.0219e-06, 'epoch': 2.38, 'throughput': 441.94}
353
+
354
+ [INFO|callbacks.py:319] 2024-08-29 19:35:38,770 >> {'loss': 0.0062, 'learning_rate': 4.5129e-06, 'epoch': 2.42, 'throughput': 441.94}
355
+
356
+ [INFO|callbacks.py:319] 2024-08-29 19:35:50,161 >> {'loss': 0.0028, 'learning_rate': 4.0285e-06, 'epoch': 2.45, 'throughput': 441.91}
357
+
358
+ [INFO|callbacks.py:319] 2024-08-29 19:36:01,253 >> {'loss': 0.0022, 'learning_rate': 3.5693e-06, 'epoch': 2.48, 'throughput': 441.82}
359
+
360
+ [INFO|callbacks.py:319] 2024-08-29 19:36:11,807 >> {'loss': 0.0061, 'learning_rate': 3.1359e-06, 'epoch': 2.52, 'throughput': 441.87}
361
+
362
+ [INFO|callbacks.py:319] 2024-08-29 19:36:22,359 >> {'loss': 0.0077, 'learning_rate': 2.7288e-06, 'epoch': 2.55, 'throughput': 441.94}
363
+
364
+ [INFO|callbacks.py:319] 2024-08-29 19:36:33,055 >> {'loss': 0.0053, 'learning_rate': 2.3484e-06, 'epoch': 2.58, 'throughput': 441.82}
365
+
366
+ [INFO|callbacks.py:319] 2024-08-29 19:36:43,743 >> {'loss': 0.0021, 'learning_rate': 1.9953e-06, 'epoch': 2.62, 'throughput': 442.19}
367
+
368
+ [INFO|callbacks.py:319] 2024-08-29 19:36:54,467 >> {'loss': 0.0063, 'learning_rate': 1.6698e-06, 'epoch': 2.65, 'throughput': 442.11}
369
+
370
+ [INFO|trainer.py:3503] 2024-08-29 19:36:54,467 >> Saving model checkpoint to saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-400
371
+
372
+ [INFO|configuration_utils.py:733] 2024-08-29 19:36:54,690 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
373
+
374
+ [INFO|configuration_utils.py:800] 2024-08-29 19:36:54,691 >> Model config MistralConfig {
375
+ "architectures": [
376
+ "MistralForCausalLM"
377
+ ],
378
+ "attention_dropout": 0.0,
379
+ "bos_token_id": 1,
380
+ "eos_token_id": 2,
381
+ "head_dim": 128,
382
+ "hidden_act": "silu",
383
+ "hidden_size": 4096,
384
+ "initializer_range": 0.02,
385
+ "intermediate_size": 14336,
386
+ "max_position_embeddings": 32768,
387
+ "model_type": "mistral",
388
+ "num_attention_heads": 32,
389
+ "num_hidden_layers": 32,
390
+ "num_key_value_heads": 8,
391
+ "rms_norm_eps": 1e-05,
392
+ "rope_theta": 10000.0,
393
+ "sliding_window": 4096,
394
+ "tie_word_embeddings": false,
395
+ "torch_dtype": "bfloat16",
396
+ "transformers_version": "4.43.4",
397
+ "use_cache": true,
398
+ "vocab_size": 32000
399
+ }
400
+
401
+
402
+ [INFO|tokenization_utils_base.py:2702] 2024-08-29 19:36:54,768 >> tokenizer config file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-400/tokenizer_config.json
403
+
404
+ [INFO|tokenization_utils_base.py:2711] 2024-08-29 19:36:54,768 >> Special tokens file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-400/special_tokens_map.json
405
+
406
+ [INFO|callbacks.py:319] 2024-08-29 19:37:05,578 >> {'loss': 0.0052, 'learning_rate': 1.3724e-06, 'epoch': 2.68, 'throughput': 442.09}
407
+
408
+ [INFO|callbacks.py:319] 2024-08-29 19:37:16,110 >> {'loss': 0.0051, 'learning_rate': 1.1034e-06, 'epoch': 2.72, 'throughput': 442.47}
409
+
410
+ [INFO|callbacks.py:319] 2024-08-29 19:37:26,696 >> {'loss': 0.0084, 'learning_rate': 8.6311e-07, 'epoch': 2.75, 'throughput': 442.54}
411
+
412
+ [INFO|callbacks.py:319] 2024-08-29 19:37:37,213 >> {'loss': 0.0122, 'learning_rate': 6.5185e-07, 'epoch': 2.78, 'throughput': 442.77}
413
+
414
+ [INFO|callbacks.py:319] 2024-08-29 19:37:48,006 >> {'loss': 0.0060, 'learning_rate': 4.6986e-07, 'epoch': 2.81, 'throughput': 442.71}
415
+
416
+ [INFO|callbacks.py:319] 2024-08-29 19:37:58,673 >> {'loss': 0.0044, 'learning_rate': 3.1736e-07, 'epoch': 2.85, 'throughput': 442.88}
417
+
418
+ [INFO|callbacks.py:319] 2024-08-29 19:38:09,686 >> {'loss': 0.0091, 'learning_rate': 1.9453e-07, 'epoch': 2.88, 'throughput': 442.96}
419
+
420
+ [INFO|callbacks.py:319] 2024-08-29 19:38:20,873 >> {'loss': 0.0050, 'learning_rate': 1.0153e-07, 'epoch': 2.91, 'throughput': 443.07}
421
+
422
+ [INFO|callbacks.py:319] 2024-08-29 19:38:32,015 >> {'loss': 0.0051, 'learning_rate': 3.8466e-08, 'epoch': 2.95, 'throughput': 443.03}
423
+
424
+ [INFO|callbacks.py:319] 2024-08-29 19:38:43,205 >> {'loss': 0.0049, 'learning_rate': 5.4105e-09, 'epoch': 2.98, 'throughput': 442.9}
425
+
426
+ [INFO|trainer.py:3503] 2024-08-29 19:38:49,911 >> Saving model checkpoint to saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-453
427
+
428
+ [INFO|configuration_utils.py:733] 2024-08-29 19:38:50,138 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
429
+
430
+ [INFO|configuration_utils.py:800] 2024-08-29 19:38:50,138 >> Model config MistralConfig {
431
+ "architectures": [
432
+ "MistralForCausalLM"
433
+ ],
434
+ "attention_dropout": 0.0,
435
+ "bos_token_id": 1,
436
+ "eos_token_id": 2,
437
+ "head_dim": 128,
438
+ "hidden_act": "silu",
439
+ "hidden_size": 4096,
440
+ "initializer_range": 0.02,
441
+ "intermediate_size": 14336,
442
+ "max_position_embeddings": 32768,
443
+ "model_type": "mistral",
444
+ "num_attention_heads": 32,
445
+ "num_hidden_layers": 32,
446
+ "num_key_value_heads": 8,
447
+ "rms_norm_eps": 1e-05,
448
+ "rope_theta": 10000.0,
449
+ "sliding_window": 4096,
450
+ "tie_word_embeddings": false,
451
+ "torch_dtype": "bfloat16",
452
+ "transformers_version": "4.43.4",
453
+ "use_cache": true,
454
+ "vocab_size": 32000
455
+ }
456
+
457
+
458
+ [INFO|tokenization_utils_base.py:2702] 2024-08-29 19:38:50,212 >> tokenizer config file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-453/tokenizer_config.json
459
+
460
+ [INFO|tokenization_utils_base.py:2711] 2024-08-29 19:38:50,212 >> Special tokens file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/checkpoint-453/special_tokens_map.json
461
+
462
+ [INFO|trainer.py:2394] 2024-08-29 19:38:50,404 >>
463
+
464
+ Training completed. Do not forget to share your model on huggingface.co/models =)
465
+
466
+
467
+
468
+ [INFO|trainer.py:3503] 2024-08-29 19:38:50,405 >> Saving model checkpoint to saves/Mistral-7B-v0.1-Chat/lora/mistral_docker
469
+
470
+ [INFO|configuration_utils.py:733] 2024-08-29 19:38:50,630 >> loading configuration file config.json from cache at /root/.cache/huggingface/hub/models--mistralai--Mistral-7B-Instruct-v0.1/snapshots/2dcff66eac0c01dc50e4c41eea959968232187fe/config.json
471
+
472
+ [INFO|configuration_utils.py:800] 2024-08-29 19:38:50,631 >> Model config MistralConfig {
473
+ "architectures": [
474
+ "MistralForCausalLM"
475
+ ],
476
+ "attention_dropout": 0.0,
477
+ "bos_token_id": 1,
478
+ "eos_token_id": 2,
479
+ "head_dim": 128,
480
+ "hidden_act": "silu",
481
+ "hidden_size": 4096,
482
+ "initializer_range": 0.02,
483
+ "intermediate_size": 14336,
484
+ "max_position_embeddings": 32768,
485
+ "model_type": "mistral",
486
+ "num_attention_heads": 32,
487
+ "num_hidden_layers": 32,
488
+ "num_key_value_heads": 8,
489
+ "rms_norm_eps": 1e-05,
490
+ "rope_theta": 10000.0,
491
+ "sliding_window": 4096,
492
+ "tie_word_embeddings": false,
493
+ "torch_dtype": "bfloat16",
494
+ "transformers_version": "4.43.4",
495
+ "use_cache": true,
496
+ "vocab_size": 32000
497
+ }
498
+
499
+
500
+ [INFO|tokenization_utils_base.py:2702] 2024-08-29 19:38:50,703 >> tokenizer config file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/tokenizer_config.json
501
+
502
+ [INFO|tokenization_utils_base.py:2711] 2024-08-29 19:38:50,704 >> Special tokens file saved in saves/Mistral-7B-v0.1-Chat/lora/mistral_docker/special_tokens_map.json
503
+
504
+ [WARNING|ploting.py:89] 2024-08-29 19:38:50,814 >> No metric eval_loss to plot.
505
+
506
+ [WARNING|ploting.py:89] 2024-08-29 19:38:50,814 >> No metric eval_accuracy to plot.
507
+
508
+ [INFO|modelcard.py:449] 2024-08-29 19:38:50,816 >> Dropping the following result as it does not have all the necessary fields:
509
+ {'task': {'name': 'Causal Language Modeling', 'type': 'text-generation'}}
510
+
special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "</s>",
17
+ "unk_token": {
18
+ "content": "<unk>",
19
+ "lstrip": false,
20
+ "normalized": false,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "additional_special_tokens": [],
32
+ "bos_token": "<s>",
33
+ "chat_template": "{{ '<s>' }}{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '[INST] ' + content + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ content + '</s>' }}{% endif %}{% endfor %}",
34
+ "clean_up_tokenization_spaces": false,
35
+ "eos_token": "</s>",
36
+ "legacy": false,
37
+ "model_max_length": 1000000000000000019884624838656,
38
+ "pad_token": "</s>",
39
+ "padding_side": "right",
40
+ "sp_model_kwargs": {},
41
+ "spaces_between_special_tokens": false,
42
+ "split_special_tokens": false,
43
+ "tokenizer_class": "LlamaTokenizer",
44
+ "unk_token": "<unk>",
45
+ "use_default_system_prompt": false
46
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "num_input_tokens_seen": 437216,
4
+ "total_flos": 1.8708380691726336e+16,
5
+ "train_loss": 0.044493223137528556,
6
+ "train_runtime": 988.3228,
7
+ "train_samples_per_second": 7.331,
8
+ "train_steps_per_second": 0.458
9
+ }
trainer_log.jsonl ADDED
@@ -0,0 +1,91 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 5, "total_steps": 453, "loss": 0.6997, "learning_rate": 4.998497170031657e-05, "epoch": 0.033112582781456956, "percentage": 1.1, "elapsed_time": "0:00:11", "remaining_time": "0:17:06", "throughput": 428.77, "total_tokens": 4912}
2
+ {"current_steps": 10, "total_steps": 453, "loss": 0.3509, "learning_rate": 4.9939904869249616e-05, "epoch": 0.06622516556291391, "percentage": 2.21, "elapsed_time": "0:00:22", "remaining_time": "0:16:36", "throughput": 441.91, "total_tokens": 9936}
3
+ {"current_steps": 15, "total_steps": 453, "loss": 0.2853, "learning_rate": 4.9864853689026556e-05, "epoch": 0.09933774834437085, "percentage": 3.31, "elapsed_time": "0:00:33", "remaining_time": "0:16:17", "throughput": 443.85, "total_tokens": 14864}
4
+ {"current_steps": 20, "total_steps": 453, "loss": 0.2456, "learning_rate": 4.975990839097764e-05, "epoch": 0.13245033112582782, "percentage": 4.42, "elapsed_time": "0:00:44", "remaining_time": "0:15:56", "throughput": 448.8, "total_tokens": 19824}
5
+ {"current_steps": 25, "total_steps": 453, "loss": 0.2084, "learning_rate": 4.9625195147054034e-05, "epoch": 0.16556291390728478, "percentage": 5.52, "elapsed_time": "0:00:55", "remaining_time": "0:15:43", "throughput": 444.22, "total_tokens": 24480}
6
+ {"current_steps": 30, "total_steps": 453, "loss": 0.1626, "learning_rate": 4.9460875918135804e-05, "epoch": 0.1986754966887417, "percentage": 6.62, "elapsed_time": "0:01:05", "remaining_time": "0:15:30", "throughput": 444.12, "total_tokens": 29296}
7
+ {"current_steps": 35, "total_steps": 453, "loss": 0.1405, "learning_rate": 4.9267148259312224e-05, "epoch": 0.23178807947019867, "percentage": 7.73, "elapsed_time": "0:01:16", "remaining_time": "0:15:15", "throughput": 442.72, "total_tokens": 33936}
8
+ {"current_steps": 40, "total_steps": 453, "loss": 0.1182, "learning_rate": 4.9044245082368415e-05, "epoch": 0.26490066225165565, "percentage": 8.83, "elapsed_time": "0:01:27", "remaining_time": "0:15:07", "throughput": 444.57, "total_tokens": 39056}
9
+ {"current_steps": 45, "total_steps": 453, "loss": 0.1285, "learning_rate": 4.879243437576383e-05, "epoch": 0.2980132450331126, "percentage": 9.93, "elapsed_time": "0:01:39", "remaining_time": "0:15:00", "throughput": 438.23, "total_tokens": 43520}
10
+ {"current_steps": 50, "total_steps": 453, "loss": 0.0906, "learning_rate": 4.8512018882439475e-05, "epoch": 0.33112582781456956, "percentage": 11.04, "elapsed_time": "0:01:50", "remaining_time": "0:14:50", "throughput": 440.48, "total_tokens": 48656}
11
+ {"current_steps": 55, "total_steps": 453, "loss": 0.054, "learning_rate": 4.820333573584091e-05, "epoch": 0.36423841059602646, "percentage": 12.14, "elapsed_time": "0:02:01", "remaining_time": "0:14:37", "throughput": 441.77, "total_tokens": 53600}
12
+ {"current_steps": 60, "total_steps": 453, "loss": 0.073, "learning_rate": 4.786675605459487e-05, "epoch": 0.3973509933774834, "percentage": 13.25, "elapsed_time": "0:02:12", "remaining_time": "0:14:25", "throughput": 442.07, "total_tokens": 58384}
13
+ {"current_steps": 65, "total_steps": 453, "loss": 0.0524, "learning_rate": 4.7502684496326746e-05, "epoch": 0.4304635761589404, "percentage": 14.35, "elapsed_time": "0:02:22", "remaining_time": "0:14:11", "throughput": 442.81, "total_tokens": 63152}
14
+ {"current_steps": 70, "total_steps": 453, "loss": 0.0534, "learning_rate": 4.711155877115523e-05, "epoch": 0.46357615894039733, "percentage": 15.45, "elapsed_time": "0:02:33", "remaining_time": "0:13:58", "throughput": 442.95, "total_tokens": 67888}
15
+ {"current_steps": 75, "total_steps": 453, "loss": 0.0759, "learning_rate": 4.669384911544927e-05, "epoch": 0.4966887417218543, "percentage": 16.56, "elapsed_time": "0:02:43", "remaining_time": "0:13:46", "throughput": 442.11, "total_tokens": 72480}
16
+ {"current_steps": 80, "total_steps": 453, "loss": 0.0766, "learning_rate": 4.625005772647979e-05, "epoch": 0.5298013245033113, "percentage": 17.66, "elapsed_time": "0:02:54", "remaining_time": "0:13:35", "throughput": 440.9, "total_tokens": 77120}
17
+ {"current_steps": 85, "total_steps": 453, "loss": 0.0383, "learning_rate": 4.578071815864602e-05, "epoch": 0.5629139072847682, "percentage": 18.76, "elapsed_time": "0:03:05", "remaining_time": "0:13:25", "throughput": 438.74, "total_tokens": 81584}
18
+ {"current_steps": 90, "total_steps": 453, "loss": 0.0393, "learning_rate": 4.528639468200226e-05, "epoch": 0.5960264900662252, "percentage": 19.87, "elapsed_time": "0:03:16", "remaining_time": "0:13:13", "throughput": 439.35, "total_tokens": 86416}
19
+ {"current_steps": 95, "total_steps": 453, "loss": 0.0589, "learning_rate": 4.476768160385632e-05, "epoch": 0.6291390728476821, "percentage": 20.97, "elapsed_time": "0:03:27", "remaining_time": "0:13:00", "throughput": 440.4, "total_tokens": 91248}
20
+ {"current_steps": 100, "total_steps": 453, "loss": 0.0351, "learning_rate": 4.4225202554255227e-05, "epoch": 0.6622516556291391, "percentage": 22.08, "elapsed_time": "0:03:37", "remaining_time": "0:12:49", "throughput": 440.1, "total_tokens": 95936}
21
+ {"current_steps": 105, "total_steps": 453, "loss": 0.0417, "learning_rate": 4.3659609736217344e-05, "epoch": 0.695364238410596, "percentage": 23.18, "elapsed_time": "0:03:49", "remaining_time": "0:12:39", "throughput": 439.61, "total_tokens": 100704}
22
+ {"current_steps": 110, "total_steps": 453, "loss": 0.0437, "learning_rate": 4.3071583141612135e-05, "epoch": 0.7284768211920529, "percentage": 24.28, "elapsed_time": "0:03:59", "remaining_time": "0:12:28", "throughput": 439.17, "total_tokens": 105376}
23
+ {"current_steps": 115, "total_steps": 453, "loss": 0.0498, "learning_rate": 4.2461829733630435e-05, "epoch": 0.7615894039735099, "percentage": 25.39, "elapsed_time": "0:04:10", "remaining_time": "0:12:17", "throughput": 439.15, "total_tokens": 110208}
24
+ {"current_steps": 120, "total_steps": 453, "loss": 0.0239, "learning_rate": 4.1831082596828106e-05, "epoch": 0.7947019867549668, "percentage": 26.49, "elapsed_time": "0:04:21", "remaining_time": "0:12:06", "throughput": 438.31, "total_tokens": 114704}
25
+ {"current_steps": 125, "total_steps": 453, "loss": 0.0228, "learning_rate": 4.118010005576485e-05, "epoch": 0.8278145695364238, "percentage": 27.59, "elapsed_time": "0:04:32", "remaining_time": "0:11:54", "throughput": 439.58, "total_tokens": 119744}
26
+ {"current_steps": 130, "total_steps": 453, "loss": 0.039, "learning_rate": 4.050966476329793e-05, "epoch": 0.8609271523178808, "percentage": 28.7, "elapsed_time": "0:04:43", "remaining_time": "0:11:44", "throughput": 440.13, "total_tokens": 124736}
27
+ {"current_steps": 135, "total_steps": 453, "loss": 0.05, "learning_rate": 3.9820582759626825e-05, "epoch": 0.8940397350993378, "percentage": 29.8, "elapsed_time": "0:04:54", "remaining_time": "0:11:32", "throughput": 440.36, "total_tokens": 129552}
28
+ {"current_steps": 140, "total_steps": 453, "loss": 0.02, "learning_rate": 3.911368250322014e-05, "epoch": 0.9271523178807947, "percentage": 30.91, "elapsed_time": "0:05:04", "remaining_time": "0:11:21", "throughput": 441.12, "total_tokens": 134400}
29
+ {"current_steps": 145, "total_steps": 453, "loss": 0.0159, "learning_rate": 3.8389813874789856e-05, "epoch": 0.9602649006622517, "percentage": 32.01, "elapsed_time": "0:05:15", "remaining_time": "0:11:10", "throughput": 441.95, "total_tokens": 139424}
30
+ {"current_steps": 150, "total_steps": 453, "loss": 0.017, "learning_rate": 3.764984715551032e-05, "epoch": 0.9933774834437086, "percentage": 33.11, "elapsed_time": "0:05:25", "remaining_time": "0:10:58", "throughput": 442.86, "total_tokens": 144368}
31
+ {"current_steps": 155, "total_steps": 453, "loss": 0.0298, "learning_rate": 3.6894671980710574e-05, "epoch": 1.0264900662251655, "percentage": 34.22, "elapsed_time": "0:05:36", "remaining_time": "0:10:46", "throughput": 443.06, "total_tokens": 149040}
32
+ {"current_steps": 160, "total_steps": 453, "loss": 0.0357, "learning_rate": 3.612519627029787e-05, "epoch": 1.0596026490066226, "percentage": 35.32, "elapsed_time": "0:05:46", "remaining_time": "0:10:35", "throughput": 443.05, "total_tokens": 153712}
33
+ {"current_steps": 165, "total_steps": 453, "loss": 0.0185, "learning_rate": 3.534234513719821e-05, "epoch": 1.0927152317880795, "percentage": 36.42, "elapsed_time": "0:05:57", "remaining_time": "0:10:23", "throughput": 444.01, "total_tokens": 158640}
34
+ {"current_steps": 170, "total_steps": 453, "loss": 0.0339, "learning_rate": 3.4547059775126445e-05, "epoch": 1.1258278145695364, "percentage": 37.53, "elapsed_time": "0:06:08", "remaining_time": "0:10:12", "throughput": 444.2, "total_tokens": 163552}
35
+ {"current_steps": 175, "total_steps": 453, "loss": 0.0224, "learning_rate": 3.3740296327022984e-05, "epoch": 1.1589403973509933, "percentage": 38.63, "elapsed_time": "0:06:19", "remaining_time": "0:10:02", "throughput": 444.02, "total_tokens": 168352}
36
+ {"current_steps": 180, "total_steps": 453, "loss": 0.0192, "learning_rate": 3.292302473551757e-05, "epoch": 1.1920529801324504, "percentage": 39.74, "elapsed_time": "0:06:29", "remaining_time": "0:09:51", "throughput": 444.69, "total_tokens": 173312}
37
+ {"current_steps": 185, "total_steps": 453, "loss": 0.0275, "learning_rate": 3.20962275768022e-05, "epoch": 1.2251655629139073, "percentage": 40.84, "elapsed_time": "0:06:40", "remaining_time": "0:09:40", "throughput": 444.69, "total_tokens": 178160}
38
+ {"current_steps": 190, "total_steps": 453, "loss": 0.0086, "learning_rate": 3.126089887931515e-05, "epoch": 1.2582781456953642, "percentage": 41.94, "elapsed_time": "0:06:51", "remaining_time": "0:09:29", "throughput": 444.59, "total_tokens": 182944}
39
+ {"current_steps": 195, "total_steps": 453, "loss": 0.0166, "learning_rate": 3.0418042928656414e-05, "epoch": 1.2913907284768211, "percentage": 43.05, "elapsed_time": "0:07:02", "remaining_time": "0:09:18", "throughput": 445.07, "total_tokens": 187920}
40
+ {"current_steps": 200, "total_steps": 453, "loss": 0.0167, "learning_rate": 2.9568673060171326e-05, "epoch": 1.3245033112582782, "percentage": 44.15, "elapsed_time": "0:07:13", "remaining_time": "0:09:08", "throughput": 445.5, "total_tokens": 192992}
41
+ {"current_steps": 205, "total_steps": 453, "loss": 0.0137, "learning_rate": 2.8713810440653926e-05, "epoch": 1.3576158940397351, "percentage": 45.25, "elapsed_time": "0:07:24", "remaining_time": "0:08:57", "throughput": 444.49, "total_tokens": 197616}
42
+ {"current_steps": 210, "total_steps": 453, "loss": 0.0142, "learning_rate": 2.7854482840634965e-05, "epoch": 1.390728476821192, "percentage": 46.36, "elapsed_time": "0:07:35", "remaining_time": "0:08:46", "throughput": 444.44, "total_tokens": 202400}
43
+ {"current_steps": 215, "total_steps": 453, "loss": 0.0136, "learning_rate": 2.6991723398730383e-05, "epoch": 1.423841059602649, "percentage": 47.46, "elapsed_time": "0:07:46", "remaining_time": "0:08:36", "throughput": 444.3, "total_tokens": 207216}
44
+ {"current_steps": 220, "total_steps": 453, "loss": 0.0184, "learning_rate": 2.6126569379535985e-05, "epoch": 1.4569536423841059, "percentage": 48.57, "elapsed_time": "0:07:57", "remaining_time": "0:08:25", "throughput": 443.79, "total_tokens": 211744}
45
+ {"current_steps": 225, "total_steps": 453, "loss": 0.0098, "learning_rate": 2.526006092656161e-05, "epoch": 1.490066225165563, "percentage": 49.67, "elapsed_time": "0:08:08", "remaining_time": "0:08:14", "throughput": 443.82, "total_tokens": 216608}
46
+ {"current_steps": 230, "total_steps": 453, "loss": 0.012, "learning_rate": 2.4393239811704e-05, "epoch": 1.5231788079470199, "percentage": 50.77, "elapsed_time": "0:08:18", "remaining_time": "0:08:03", "throughput": 444.28, "total_tokens": 221552}
47
+ {"current_steps": 235, "total_steps": 453, "loss": 0.0106, "learning_rate": 2.3527148182762054e-05, "epoch": 1.5562913907284768, "percentage": 51.88, "elapsed_time": "0:08:29", "remaining_time": "0:07:52", "throughput": 443.85, "total_tokens": 226272}
48
+ {"current_steps": 240, "total_steps": 453, "loss": 0.0128, "learning_rate": 2.2662827310499995e-05, "epoch": 1.589403973509934, "percentage": 52.98, "elapsed_time": "0:08:40", "remaining_time": "0:07:42", "throughput": 443.72, "total_tokens": 231072}
49
+ {"current_steps": 245, "total_steps": 453, "loss": 0.0139, "learning_rate": 2.1801316336765126e-05, "epoch": 1.6225165562913908, "percentage": 54.08, "elapsed_time": "0:08:51", "remaining_time": "0:07:31", "throughput": 443.33, "total_tokens": 235728}
50
+ {"current_steps": 250, "total_steps": 453, "loss": 0.0084, "learning_rate": 2.0943651025164932e-05, "epoch": 1.6556291390728477, "percentage": 55.19, "elapsed_time": "0:09:02", "remaining_time": "0:07:20", "throughput": 443.48, "total_tokens": 240560}
51
+ {"current_steps": 255, "total_steps": 453, "loss": 0.0143, "learning_rate": 2.0090862515805898e-05, "epoch": 1.6887417218543046, "percentage": 56.29, "elapsed_time": "0:09:13", "remaining_time": "0:07:09", "throughput": 443.53, "total_tokens": 245408}
52
+ {"current_steps": 260, "total_steps": 453, "loss": 0.011, "learning_rate": 1.9243976085590824e-05, "epoch": 1.7218543046357615, "percentage": 57.4, "elapsed_time": "0:09:24", "remaining_time": "0:06:58", "throughput": 443.94, "total_tokens": 250400}
53
+ {"current_steps": 265, "total_steps": 453, "loss": 0.0127, "learning_rate": 1.840400991556541e-05, "epoch": 1.7549668874172184, "percentage": 58.5, "elapsed_time": "0:09:35", "remaining_time": "0:06:48", "throughput": 443.72, "total_tokens": 255216}
54
+ {"current_steps": 270, "total_steps": 453, "loss": 0.0127, "learning_rate": 1.7571973866795815e-05, "epoch": 1.7880794701986755, "percentage": 59.6, "elapsed_time": "0:09:45", "remaining_time": "0:06:37", "throughput": 443.93, "total_tokens": 260080}
55
+ {"current_steps": 275, "total_steps": 453, "loss": 0.0122, "learning_rate": 1.6748868266249114e-05, "epoch": 1.8211920529801324, "percentage": 60.71, "elapsed_time": "0:09:56", "remaining_time": "0:06:26", "throughput": 443.72, "total_tokens": 264848}
56
+ {"current_steps": 280, "total_steps": 453, "loss": 0.0169, "learning_rate": 1.5935682704136183e-05, "epoch": 1.8543046357615895, "percentage": 61.81, "elapsed_time": "0:10:07", "remaining_time": "0:06:15", "throughput": 443.77, "total_tokens": 269776}
57
+ {"current_steps": 285, "total_steps": 453, "loss": 0.0242, "learning_rate": 1.5133394844163093e-05, "epoch": 1.8874172185430464, "percentage": 62.91, "elapsed_time": "0:10:18", "remaining_time": "0:06:04", "throughput": 444.05, "total_tokens": 274752}
58
+ {"current_steps": 290, "total_steps": 453, "loss": 0.0079, "learning_rate": 1.4342969248121185e-05, "epoch": 1.9205298013245033, "percentage": 64.02, "elapsed_time": "0:10:29", "remaining_time": "0:05:53", "throughput": 443.8, "total_tokens": 279440}
59
+ {"current_steps": 295, "total_steps": 453, "loss": 0.0194, "learning_rate": 1.3565356216229268e-05, "epoch": 1.9536423841059603, "percentage": 65.12, "elapsed_time": "0:10:41", "remaining_time": "0:05:43", "throughput": 443.36, "total_tokens": 284288}
60
+ {"current_steps": 300, "total_steps": 453, "loss": 0.0133, "learning_rate": 1.2801490644621789e-05, "epoch": 1.9867549668874172, "percentage": 66.23, "elapsed_time": "0:10:52", "remaining_time": "0:05:32", "throughput": 443.09, "total_tokens": 289216}
61
+ {"current_steps": 305, "total_steps": 453, "loss": 0.0049, "learning_rate": 1.2052290901357025e-05, "epoch": 2.019867549668874, "percentage": 67.33, "elapsed_time": "0:11:04", "remaining_time": "0:05:22", "throughput": 442.42, "total_tokens": 293992}
62
+ {"current_steps": 310, "total_steps": 453, "loss": 0.0121, "learning_rate": 1.1318657722296097e-05, "epoch": 2.052980132450331, "percentage": 68.43, "elapsed_time": "0:11:15", "remaining_time": "0:05:11", "throughput": 442.9, "total_tokens": 299048}
63
+ {"current_steps": 315, "total_steps": 453, "loss": 0.0102, "learning_rate": 1.0601473128180855e-05, "epoch": 2.0860927152317883, "percentage": 69.54, "elapsed_time": "0:11:25", "remaining_time": "0:05:00", "throughput": 442.89, "total_tokens": 303768}
64
+ {"current_steps": 320, "total_steps": 453, "loss": 0.0072, "learning_rate": 9.90159936421197e-06, "epoch": 2.119205298013245, "percentage": 70.64, "elapsed_time": "0:11:37", "remaining_time": "0:04:49", "throughput": 443.13, "total_tokens": 308872}
65
+ {"current_steps": 325, "total_steps": 453, "loss": 0.0066, "learning_rate": 9.219877863402682e-06, "epoch": 2.152317880794702, "percentage": 71.74, "elapsed_time": "0:11:48", "remaining_time": "0:04:38", "throughput": 443.2, "total_tokens": 313800}
66
+ {"current_steps": 330, "total_steps": 453, "loss": 0.0119, "learning_rate": 8.55712823495419e-06, "epoch": 2.185430463576159, "percentage": 72.85, "elapsed_time": "0:11:59", "remaining_time": "0:04:28", "throughput": 442.86, "total_tokens": 318536}
67
+ {"current_steps": 335, "total_steps": 453, "loss": 0.0057, "learning_rate": 7.91414727886898e-06, "epoch": 2.218543046357616, "percentage": 73.95, "elapsed_time": "0:12:10", "remaining_time": "0:04:17", "throughput": 442.91, "total_tokens": 323512}
68
+ {"current_steps": 340, "total_steps": 453, "loss": 0.0136, "learning_rate": 7.291708027986988e-06, "epoch": 2.251655629139073, "percentage": 75.06, "elapsed_time": "0:12:21", "remaining_time": "0:04:06", "throughput": 443.28, "total_tokens": 328552}
69
+ {"current_steps": 345, "total_steps": 453, "loss": 0.0078, "learning_rate": 6.690558818595943e-06, "epoch": 2.2847682119205297, "percentage": 76.16, "elapsed_time": "0:12:32", "remaining_time": "0:03:55", "throughput": 443.17, "total_tokens": 333272}
70
+ {"current_steps": 350, "total_steps": 453, "loss": 0.0141, "learning_rate": 6.111422390733715e-06, "epoch": 2.3178807947019866, "percentage": 77.26, "elapsed_time": "0:12:43", "remaining_time": "0:03:44", "throughput": 442.62, "total_tokens": 337816}
71
+ {"current_steps": 355, "total_steps": 453, "loss": 0.014, "learning_rate": 5.55499501926394e-06, "epoch": 2.3509933774834435, "percentage": 78.37, "elapsed_time": "0:12:54", "remaining_time": "0:03:33", "throughput": 442.57, "total_tokens": 342792}
72
+ {"current_steps": 360, "total_steps": 453, "loss": 0.0045, "learning_rate": 5.02194567676986e-06, "epoch": 2.384105960264901, "percentage": 79.47, "elapsed_time": "0:13:05", "remaining_time": "0:03:23", "throughput": 441.94, "total_tokens": 347288}
73
+ {"current_steps": 365, "total_steps": 453, "loss": 0.0062, "learning_rate": 4.51291522927268e-06, "epoch": 2.4172185430463577, "percentage": 80.57, "elapsed_time": "0:13:16", "remaining_time": "0:03:12", "throughput": 441.94, "total_tokens": 352088}
74
+ {"current_steps": 370, "total_steps": 453, "loss": 0.0028, "learning_rate": 4.028515665741439e-06, "epoch": 2.4503311258278146, "percentage": 81.68, "elapsed_time": "0:13:28", "remaining_time": "0:03:01", "throughput": 441.91, "total_tokens": 357096}
75
+ {"current_steps": 375, "total_steps": 453, "loss": 0.0022, "learning_rate": 3.5693293623207086e-06, "epoch": 2.4834437086092715, "percentage": 82.78, "elapsed_time": "0:13:39", "remaining_time": "0:02:50", "throughput": 441.82, "total_tokens": 361928}
76
+ {"current_steps": 380, "total_steps": 453, "loss": 0.0061, "learning_rate": 3.135908382160771e-06, "epoch": 2.5165562913907285, "percentage": 83.89, "elapsed_time": "0:13:49", "remaining_time": "0:02:39", "throughput": 441.87, "total_tokens": 366632}
77
+ {"current_steps": 385, "total_steps": 453, "loss": 0.0077, "learning_rate": 2.728773811691923e-06, "epoch": 2.5496688741721854, "percentage": 84.99, "elapsed_time": "0:14:00", "remaining_time": "0:02:28", "throughput": 441.94, "total_tokens": 371352}
78
+ {"current_steps": 390, "total_steps": 453, "loss": 0.0053, "learning_rate": 2.348415134141102e-06, "epoch": 2.5827814569536423, "percentage": 86.09, "elapsed_time": "0:14:10", "remaining_time": "0:02:17", "throughput": 441.82, "total_tokens": 375976}
79
+ {"current_steps": 395, "total_steps": 453, "loss": 0.0021, "learning_rate": 1.995289641043768e-06, "epoch": 2.6158940397350996, "percentage": 87.2, "elapsed_time": "0:14:21", "remaining_time": "0:02:06", "throughput": 442.19, "total_tokens": 381016}
80
+ {"current_steps": 400, "total_steps": 453, "loss": 0.0063, "learning_rate": 1.6698218824588164e-06, "epoch": 2.6490066225165565, "percentage": 88.3, "elapsed_time": "0:14:32", "remaining_time": "0:01:55", "throughput": 442.11, "total_tokens": 385688}
81
+ {"current_steps": 405, "total_steps": 453, "loss": 0.0052, "learning_rate": 1.3724031565473112e-06, "epoch": 2.6821192052980134, "percentage": 89.4, "elapsed_time": "0:14:43", "remaining_time": "0:01:44", "throughput": 442.09, "total_tokens": 390584}
82
+ {"current_steps": 410, "total_steps": 453, "loss": 0.0051, "learning_rate": 1.1033910391288065e-06, "epoch": 2.7152317880794703, "percentage": 90.51, "elapsed_time": "0:14:54", "remaining_time": "0:01:33", "throughput": 442.47, "total_tokens": 395576}
83
+ {"current_steps": 415, "total_steps": 453, "loss": 0.0084, "learning_rate": 8.631089537808307e-07, "epoch": 2.748344370860927, "percentage": 91.61, "elapsed_time": "0:15:04", "remaining_time": "0:01:22", "throughput": 442.54, "total_tokens": 400328}
84
+ {"current_steps": 420, "total_steps": 453, "loss": 0.0122, "learning_rate": 6.51845782998356e-07, "epoch": 2.781456953642384, "percentage": 92.72, "elapsed_time": "0:15:15", "remaining_time": "0:01:11", "throughput": 442.77, "total_tokens": 405192}
85
+ {"current_steps": 425, "total_steps": 453, "loss": 0.006, "learning_rate": 4.698555208807853e-07, "epoch": 2.814569536423841, "percentage": 93.82, "elapsed_time": "0:15:25", "remaining_time": "0:01:01", "throughput": 442.71, "total_tokens": 409912}
86
+ {"current_steps": 430, "total_steps": 453, "loss": 0.0044, "learning_rate": 3.1735696776400703e-07, "epoch": 2.847682119205298, "percentage": 94.92, "elapsed_time": "0:15:36", "remaining_time": "0:00:50", "throughput": 442.88, "total_tokens": 414792}
87
+ {"current_steps": 435, "total_steps": 453, "loss": 0.0091, "learning_rate": 1.9453346716462317e-07, "epoch": 2.880794701986755, "percentage": 96.03, "elapsed_time": "0:15:47", "remaining_time": "0:00:39", "throughput": 442.96, "total_tokens": 419752}
88
+ {"current_steps": 440, "total_steps": 453, "loss": 0.005, "learning_rate": 1.0153268535264827e-07, "epoch": 2.9139072847682117, "percentage": 97.13, "elapsed_time": "0:15:58", "remaining_time": "0:00:28", "throughput": 443.07, "total_tokens": 424808}
89
+ {"current_steps": 445, "total_steps": 453, "loss": 0.0051, "learning_rate": 3.846643381766879e-08, "epoch": 2.9470198675496686, "percentage": 98.23, "elapsed_time": "0:16:09", "remaining_time": "0:00:17", "throughput": 443.03, "total_tokens": 429704}
90
+ {"current_steps": 450, "total_steps": 453, "loss": 0.0049, "learning_rate": 5.41053484192644e-09, "epoch": 2.980132450331126, "percentage": 99.34, "elapsed_time": "0:16:21", "remaining_time": "0:00:06", "throughput": 442.9, "total_tokens": 434536}
91
+ {"current_steps": 453, "total_steps": 453, "epoch": 3.0, "percentage": 100.0, "elapsed_time": "0:16:28", "remaining_time": "0:00:00", "throughput": 442.38, "total_tokens": 437216}
trainer_state.json ADDED
@@ -0,0 +1,763 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": null,
3
+ "best_model_checkpoint": null,
4
+ "epoch": 3.0,
5
+ "eval_steps": 500,
6
+ "global_step": 453,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.033112582781456956,
13
+ "grad_norm": 2.2606041431427,
14
+ "learning_rate": 4.998497170031657e-05,
15
+ "loss": 0.6997,
16
+ "num_input_tokens_seen": 4912,
17
+ "step": 5
18
+ },
19
+ {
20
+ "epoch": 0.06622516556291391,
21
+ "grad_norm": 2.3543715476989746,
22
+ "learning_rate": 4.9939904869249616e-05,
23
+ "loss": 0.3509,
24
+ "num_input_tokens_seen": 9936,
25
+ "step": 10
26
+ },
27
+ {
28
+ "epoch": 0.09933774834437085,
29
+ "grad_norm": 1.801797866821289,
30
+ "learning_rate": 4.9864853689026556e-05,
31
+ "loss": 0.2853,
32
+ "num_input_tokens_seen": 14864,
33
+ "step": 15
34
+ },
35
+ {
36
+ "epoch": 0.13245033112582782,
37
+ "grad_norm": 1.3575855493545532,
38
+ "learning_rate": 4.975990839097764e-05,
39
+ "loss": 0.2456,
40
+ "num_input_tokens_seen": 19824,
41
+ "step": 20
42
+ },
43
+ {
44
+ "epoch": 0.16556291390728478,
45
+ "grad_norm": 2.5929694175720215,
46
+ "learning_rate": 4.9625195147054034e-05,
47
+ "loss": 0.2084,
48
+ "num_input_tokens_seen": 24480,
49
+ "step": 25
50
+ },
51
+ {
52
+ "epoch": 0.1986754966887417,
53
+ "grad_norm": 0.8744802474975586,
54
+ "learning_rate": 4.9460875918135804e-05,
55
+ "loss": 0.1626,
56
+ "num_input_tokens_seen": 29296,
57
+ "step": 30
58
+ },
59
+ {
60
+ "epoch": 0.23178807947019867,
61
+ "grad_norm": 2.857471466064453,
62
+ "learning_rate": 4.9267148259312224e-05,
63
+ "loss": 0.1405,
64
+ "num_input_tokens_seen": 33936,
65
+ "step": 35
66
+ },
67
+ {
68
+ "epoch": 0.26490066225165565,
69
+ "grad_norm": 1.7763206958770752,
70
+ "learning_rate": 4.9044245082368415e-05,
71
+ "loss": 0.1182,
72
+ "num_input_tokens_seen": 39056,
73
+ "step": 40
74
+ },
75
+ {
76
+ "epoch": 0.2980132450331126,
77
+ "grad_norm": 2.78945255279541,
78
+ "learning_rate": 4.879243437576383e-05,
79
+ "loss": 0.1285,
80
+ "num_input_tokens_seen": 43520,
81
+ "step": 45
82
+ },
83
+ {
84
+ "epoch": 0.33112582781456956,
85
+ "grad_norm": 1.1319329738616943,
86
+ "learning_rate": 4.8512018882439475e-05,
87
+ "loss": 0.0906,
88
+ "num_input_tokens_seen": 48656,
89
+ "step": 50
90
+ },
91
+ {
92
+ "epoch": 0.36423841059602646,
93
+ "grad_norm": 0.8884170055389404,
94
+ "learning_rate": 4.820333573584091e-05,
95
+ "loss": 0.054,
96
+ "num_input_tokens_seen": 53600,
97
+ "step": 55
98
+ },
99
+ {
100
+ "epoch": 0.3973509933774834,
101
+ "grad_norm": 2.0057179927825928,
102
+ "learning_rate": 4.786675605459487e-05,
103
+ "loss": 0.073,
104
+ "num_input_tokens_seen": 58384,
105
+ "step": 60
106
+ },
107
+ {
108
+ "epoch": 0.4304635761589404,
109
+ "grad_norm": 1.3076105117797852,
110
+ "learning_rate": 4.7502684496326746e-05,
111
+ "loss": 0.0524,
112
+ "num_input_tokens_seen": 63152,
113
+ "step": 65
114
+ },
115
+ {
116
+ "epoch": 0.46357615894039733,
117
+ "grad_norm": 1.074171543121338,
118
+ "learning_rate": 4.711155877115523e-05,
119
+ "loss": 0.0534,
120
+ "num_input_tokens_seen": 67888,
121
+ "step": 70
122
+ },
123
+ {
124
+ "epoch": 0.4966887417218543,
125
+ "grad_norm": 1.634122371673584,
126
+ "learning_rate": 4.669384911544927e-05,
127
+ "loss": 0.0759,
128
+ "num_input_tokens_seen": 72480,
129
+ "step": 75
130
+ },
131
+ {
132
+ "epoch": 0.5298013245033113,
133
+ "grad_norm": 1.0059384107589722,
134
+ "learning_rate": 4.625005772647979e-05,
135
+ "loss": 0.0766,
136
+ "num_input_tokens_seen": 77120,
137
+ "step": 80
138
+ },
139
+ {
140
+ "epoch": 0.5629139072847682,
141
+ "grad_norm": 1.3958277702331543,
142
+ "learning_rate": 4.578071815864602e-05,
143
+ "loss": 0.0383,
144
+ "num_input_tokens_seen": 81584,
145
+ "step": 85
146
+ },
147
+ {
148
+ "epoch": 0.5960264900662252,
149
+ "grad_norm": 2.721231698989868,
150
+ "learning_rate": 4.528639468200226e-05,
151
+ "loss": 0.0393,
152
+ "num_input_tokens_seen": 86416,
153
+ "step": 90
154
+ },
155
+ {
156
+ "epoch": 0.6291390728476821,
157
+ "grad_norm": 1.4776098728179932,
158
+ "learning_rate": 4.476768160385632e-05,
159
+ "loss": 0.0589,
160
+ "num_input_tokens_seen": 91248,
161
+ "step": 95
162
+ },
163
+ {
164
+ "epoch": 0.6622516556291391,
165
+ "grad_norm": 1.1105955839157104,
166
+ "learning_rate": 4.4225202554255227e-05,
167
+ "loss": 0.0351,
168
+ "num_input_tokens_seen": 95936,
169
+ "step": 100
170
+ },
171
+ {
172
+ "epoch": 0.695364238410596,
173
+ "grad_norm": 1.1301405429840088,
174
+ "learning_rate": 4.3659609736217344e-05,
175
+ "loss": 0.0417,
176
+ "num_input_tokens_seen": 100704,
177
+ "step": 105
178
+ },
179
+ {
180
+ "epoch": 0.7284768211920529,
181
+ "grad_norm": 1.869552493095398,
182
+ "learning_rate": 4.3071583141612135e-05,
183
+ "loss": 0.0437,
184
+ "num_input_tokens_seen": 105376,
185
+ "step": 110
186
+ },
187
+ {
188
+ "epoch": 0.7615894039735099,
189
+ "grad_norm": 1.0497004985809326,
190
+ "learning_rate": 4.2461829733630435e-05,
191
+ "loss": 0.0498,
192
+ "num_input_tokens_seen": 110208,
193
+ "step": 115
194
+ },
195
+ {
196
+ "epoch": 0.7947019867549668,
197
+ "grad_norm": 0.4084171652793884,
198
+ "learning_rate": 4.1831082596828106e-05,
199
+ "loss": 0.0239,
200
+ "num_input_tokens_seen": 114704,
201
+ "step": 120
202
+ },
203
+ {
204
+ "epoch": 0.8278145695364238,
205
+ "grad_norm": 0.4082948565483093,
206
+ "learning_rate": 4.118010005576485e-05,
207
+ "loss": 0.0228,
208
+ "num_input_tokens_seen": 119744,
209
+ "step": 125
210
+ },
211
+ {
212
+ "epoch": 0.8609271523178808,
213
+ "grad_norm": 0.246117502450943,
214
+ "learning_rate": 4.050966476329793e-05,
215
+ "loss": 0.039,
216
+ "num_input_tokens_seen": 124736,
217
+ "step": 130
218
+ },
219
+ {
220
+ "epoch": 0.8940397350993378,
221
+ "grad_norm": 0.5470453500747681,
222
+ "learning_rate": 3.9820582759626825e-05,
223
+ "loss": 0.05,
224
+ "num_input_tokens_seen": 129552,
225
+ "step": 135
226
+ },
227
+ {
228
+ "epoch": 0.9271523178807947,
229
+ "grad_norm": 0.6097426414489746,
230
+ "learning_rate": 3.911368250322014e-05,
231
+ "loss": 0.02,
232
+ "num_input_tokens_seen": 134400,
233
+ "step": 140
234
+ },
235
+ {
236
+ "epoch": 0.9602649006622517,
237
+ "grad_norm": 0.4274413585662842,
238
+ "learning_rate": 3.8389813874789856e-05,
239
+ "loss": 0.0159,
240
+ "num_input_tokens_seen": 139424,
241
+ "step": 145
242
+ },
243
+ {
244
+ "epoch": 0.9933774834437086,
245
+ "grad_norm": 0.1949557512998581,
246
+ "learning_rate": 3.764984715551032e-05,
247
+ "loss": 0.017,
248
+ "num_input_tokens_seen": 144368,
249
+ "step": 150
250
+ },
251
+ {
252
+ "epoch": 1.0264900662251655,
253
+ "grad_norm": 0.9226788878440857,
254
+ "learning_rate": 3.6894671980710574e-05,
255
+ "loss": 0.0298,
256
+ "num_input_tokens_seen": 149040,
257
+ "step": 155
258
+ },
259
+ {
260
+ "epoch": 1.0596026490066226,
261
+ "grad_norm": 1.5730527639389038,
262
+ "learning_rate": 3.612519627029787e-05,
263
+ "loss": 0.0357,
264
+ "num_input_tokens_seen": 153712,
265
+ "step": 160
266
+ },
267
+ {
268
+ "epoch": 1.0927152317880795,
269
+ "grad_norm": 0.5750863552093506,
270
+ "learning_rate": 3.534234513719821e-05,
271
+ "loss": 0.0185,
272
+ "num_input_tokens_seen": 158640,
273
+ "step": 165
274
+ },
275
+ {
276
+ "epoch": 1.1258278145695364,
277
+ "grad_norm": 0.4966561198234558,
278
+ "learning_rate": 3.4547059775126445e-05,
279
+ "loss": 0.0339,
280
+ "num_input_tokens_seen": 163552,
281
+ "step": 170
282
+ },
283
+ {
284
+ "epoch": 1.1589403973509933,
285
+ "grad_norm": 0.6601276993751526,
286
+ "learning_rate": 3.3740296327022984e-05,
287
+ "loss": 0.0224,
288
+ "num_input_tokens_seen": 168352,
289
+ "step": 175
290
+ },
291
+ {
292
+ "epoch": 1.1920529801324504,
293
+ "grad_norm": 1.2051963806152344,
294
+ "learning_rate": 3.292302473551757e-05,
295
+ "loss": 0.0192,
296
+ "num_input_tokens_seen": 173312,
297
+ "step": 180
298
+ },
299
+ {
300
+ "epoch": 1.2251655629139073,
301
+ "grad_norm": 0.48197436332702637,
302
+ "learning_rate": 3.20962275768022e-05,
303
+ "loss": 0.0275,
304
+ "num_input_tokens_seen": 178160,
305
+ "step": 185
306
+ },
307
+ {
308
+ "epoch": 1.2582781456953642,
309
+ "grad_norm": 0.3365241289138794,
310
+ "learning_rate": 3.126089887931515e-05,
311
+ "loss": 0.0086,
312
+ "num_input_tokens_seen": 182944,
313
+ "step": 190
314
+ },
315
+ {
316
+ "epoch": 1.2913907284768211,
317
+ "grad_norm": 0.5984787940979004,
318
+ "learning_rate": 3.0418042928656414e-05,
319
+ "loss": 0.0166,
320
+ "num_input_tokens_seen": 187920,
321
+ "step": 195
322
+ },
323
+ {
324
+ "epoch": 1.3245033112582782,
325
+ "grad_norm": 0.3351554274559021,
326
+ "learning_rate": 2.9568673060171326e-05,
327
+ "loss": 0.0167,
328
+ "num_input_tokens_seen": 192992,
329
+ "step": 200
330
+ },
331
+ {
332
+ "epoch": 1.3576158940397351,
333
+ "grad_norm": 0.21267688274383545,
334
+ "learning_rate": 2.8713810440653926e-05,
335
+ "loss": 0.0137,
336
+ "num_input_tokens_seen": 197616,
337
+ "step": 205
338
+ },
339
+ {
340
+ "epoch": 1.390728476821192,
341
+ "grad_norm": 0.17382262647151947,
342
+ "learning_rate": 2.7854482840634965e-05,
343
+ "loss": 0.0142,
344
+ "num_input_tokens_seen": 202400,
345
+ "step": 210
346
+ },
347
+ {
348
+ "epoch": 1.423841059602649,
349
+ "grad_norm": 0.552691638469696,
350
+ "learning_rate": 2.6991723398730383e-05,
351
+ "loss": 0.0136,
352
+ "num_input_tokens_seen": 207216,
353
+ "step": 215
354
+ },
355
+ {
356
+ "epoch": 1.4569536423841059,
357
+ "grad_norm": 0.4015306532382965,
358
+ "learning_rate": 2.6126569379535985e-05,
359
+ "loss": 0.0184,
360
+ "num_input_tokens_seen": 211744,
361
+ "step": 220
362
+ },
363
+ {
364
+ "epoch": 1.490066225165563,
365
+ "grad_norm": 0.20611760020256042,
366
+ "learning_rate": 2.526006092656161e-05,
367
+ "loss": 0.0098,
368
+ "num_input_tokens_seen": 216608,
369
+ "step": 225
370
+ },
371
+ {
372
+ "epoch": 1.5231788079470199,
373
+ "grad_norm": 0.3923095762729645,
374
+ "learning_rate": 2.4393239811704e-05,
375
+ "loss": 0.012,
376
+ "num_input_tokens_seen": 221552,
377
+ "step": 230
378
+ },
379
+ {
380
+ "epoch": 1.5562913907284768,
381
+ "grad_norm": 0.2283228635787964,
382
+ "learning_rate": 2.3527148182762054e-05,
383
+ "loss": 0.0106,
384
+ "num_input_tokens_seen": 226272,
385
+ "step": 235
386
+ },
387
+ {
388
+ "epoch": 1.589403973509934,
389
+ "grad_norm": 1.2066081762313843,
390
+ "learning_rate": 2.2662827310499995e-05,
391
+ "loss": 0.0128,
392
+ "num_input_tokens_seen": 231072,
393
+ "step": 240
394
+ },
395
+ {
396
+ "epoch": 1.6225165562913908,
397
+ "grad_norm": 0.39357009530067444,
398
+ "learning_rate": 2.1801316336765126e-05,
399
+ "loss": 0.0139,
400
+ "num_input_tokens_seen": 235728,
401
+ "step": 245
402
+ },
403
+ {
404
+ "epoch": 1.6556291390728477,
405
+ "grad_norm": 0.2383226603269577,
406
+ "learning_rate": 2.0943651025164932e-05,
407
+ "loss": 0.0084,
408
+ "num_input_tokens_seen": 240560,
409
+ "step": 250
410
+ },
411
+ {
412
+ "epoch": 1.6887417218543046,
413
+ "grad_norm": 1.2794650793075562,
414
+ "learning_rate": 2.0090862515805898e-05,
415
+ "loss": 0.0143,
416
+ "num_input_tokens_seen": 245408,
417
+ "step": 255
418
+ },
419
+ {
420
+ "epoch": 1.7218543046357615,
421
+ "grad_norm": 0.5334519147872925,
422
+ "learning_rate": 1.9243976085590824e-05,
423
+ "loss": 0.011,
424
+ "num_input_tokens_seen": 250400,
425
+ "step": 260
426
+ },
427
+ {
428
+ "epoch": 1.7549668874172184,
429
+ "grad_norm": 0.32287150621414185,
430
+ "learning_rate": 1.840400991556541e-05,
431
+ "loss": 0.0127,
432
+ "num_input_tokens_seen": 255216,
433
+ "step": 265
434
+ },
435
+ {
436
+ "epoch": 1.7880794701986755,
437
+ "grad_norm": 0.7233603596687317,
438
+ "learning_rate": 1.7571973866795815e-05,
439
+ "loss": 0.0127,
440
+ "num_input_tokens_seen": 260080,
441
+ "step": 270
442
+ },
443
+ {
444
+ "epoch": 1.8211920529801324,
445
+ "grad_norm": 0.2672172784805298,
446
+ "learning_rate": 1.6748868266249114e-05,
447
+ "loss": 0.0122,
448
+ "num_input_tokens_seen": 264848,
449
+ "step": 275
450
+ },
451
+ {
452
+ "epoch": 1.8543046357615895,
453
+ "grad_norm": 0.32350099086761475,
454
+ "learning_rate": 1.5935682704136183e-05,
455
+ "loss": 0.0169,
456
+ "num_input_tokens_seen": 269776,
457
+ "step": 280
458
+ },
459
+ {
460
+ "epoch": 1.8874172185430464,
461
+ "grad_norm": 0.26331964135169983,
462
+ "learning_rate": 1.5133394844163093e-05,
463
+ "loss": 0.0242,
464
+ "num_input_tokens_seen": 274752,
465
+ "step": 285
466
+ },
467
+ {
468
+ "epoch": 1.9205298013245033,
469
+ "grad_norm": 0.25610876083374023,
470
+ "learning_rate": 1.4342969248121185e-05,
471
+ "loss": 0.0079,
472
+ "num_input_tokens_seen": 279440,
473
+ "step": 290
474
+ },
475
+ {
476
+ "epoch": 1.9536423841059603,
477
+ "grad_norm": 1.2221907377243042,
478
+ "learning_rate": 1.3565356216229268e-05,
479
+ "loss": 0.0194,
480
+ "num_input_tokens_seen": 284288,
481
+ "step": 295
482
+ },
483
+ {
484
+ "epoch": 1.9867549668874172,
485
+ "grad_norm": 0.46571576595306396,
486
+ "learning_rate": 1.2801490644621789e-05,
487
+ "loss": 0.0133,
488
+ "num_input_tokens_seen": 289216,
489
+ "step": 300
490
+ },
491
+ {
492
+ "epoch": 2.019867549668874,
493
+ "grad_norm": 0.06845066696405411,
494
+ "learning_rate": 1.2052290901357025e-05,
495
+ "loss": 0.0049,
496
+ "num_input_tokens_seen": 293992,
497
+ "step": 305
498
+ },
499
+ {
500
+ "epoch": 2.052980132450331,
501
+ "grad_norm": 0.3654076159000397,
502
+ "learning_rate": 1.1318657722296097e-05,
503
+ "loss": 0.0121,
504
+ "num_input_tokens_seen": 299048,
505
+ "step": 310
506
+ },
507
+ {
508
+ "epoch": 2.0860927152317883,
509
+ "grad_norm": 0.1736595779657364,
510
+ "learning_rate": 1.0601473128180855e-05,
511
+ "loss": 0.0102,
512
+ "num_input_tokens_seen": 303768,
513
+ "step": 315
514
+ },
515
+ {
516
+ "epoch": 2.119205298013245,
517
+ "grad_norm": 0.15334352850914001,
518
+ "learning_rate": 9.90159936421197e-06,
519
+ "loss": 0.0072,
520
+ "num_input_tokens_seen": 308872,
521
+ "step": 320
522
+ },
523
+ {
524
+ "epoch": 2.152317880794702,
525
+ "grad_norm": 0.15292225778102875,
526
+ "learning_rate": 9.219877863402682e-06,
527
+ "loss": 0.0066,
528
+ "num_input_tokens_seen": 313800,
529
+ "step": 325
530
+ },
531
+ {
532
+ "epoch": 2.185430463576159,
533
+ "grad_norm": 0.12052378058433533,
534
+ "learning_rate": 8.55712823495419e-06,
535
+ "loss": 0.0119,
536
+ "num_input_tokens_seen": 318536,
537
+ "step": 330
538
+ },
539
+ {
540
+ "epoch": 2.218543046357616,
541
+ "grad_norm": 0.16140712797641754,
542
+ "learning_rate": 7.91414727886898e-06,
543
+ "loss": 0.0057,
544
+ "num_input_tokens_seen": 323512,
545
+ "step": 335
546
+ },
547
+ {
548
+ "epoch": 2.251655629139073,
549
+ "grad_norm": 0.49767032265663147,
550
+ "learning_rate": 7.291708027986988e-06,
551
+ "loss": 0.0136,
552
+ "num_input_tokens_seen": 328552,
553
+ "step": 340
554
+ },
555
+ {
556
+ "epoch": 2.2847682119205297,
557
+ "grad_norm": 0.10487841814756393,
558
+ "learning_rate": 6.690558818595943e-06,
559
+ "loss": 0.0078,
560
+ "num_input_tokens_seen": 333272,
561
+ "step": 345
562
+ },
563
+ {
564
+ "epoch": 2.3178807947019866,
565
+ "grad_norm": 0.2247195988893509,
566
+ "learning_rate": 6.111422390733715e-06,
567
+ "loss": 0.0141,
568
+ "num_input_tokens_seen": 337816,
569
+ "step": 350
570
+ },
571
+ {
572
+ "epoch": 2.3509933774834435,
573
+ "grad_norm": 0.15797987580299377,
574
+ "learning_rate": 5.55499501926394e-06,
575
+ "loss": 0.014,
576
+ "num_input_tokens_seen": 342792,
577
+ "step": 355
578
+ },
579
+ {
580
+ "epoch": 2.384105960264901,
581
+ "grad_norm": 0.33236005902290344,
582
+ "learning_rate": 5.02194567676986e-06,
583
+ "loss": 0.0045,
584
+ "num_input_tokens_seen": 347288,
585
+ "step": 360
586
+ },
587
+ {
588
+ "epoch": 2.4172185430463577,
589
+ "grad_norm": 0.3886609375476837,
590
+ "learning_rate": 4.51291522927268e-06,
591
+ "loss": 0.0062,
592
+ "num_input_tokens_seen": 352088,
593
+ "step": 365
594
+ },
595
+ {
596
+ "epoch": 2.4503311258278146,
597
+ "grad_norm": 0.3580506145954132,
598
+ "learning_rate": 4.028515665741439e-06,
599
+ "loss": 0.0028,
600
+ "num_input_tokens_seen": 357096,
601
+ "step": 370
602
+ },
603
+ {
604
+ "epoch": 2.4834437086092715,
605
+ "grad_norm": 0.045303523540496826,
606
+ "learning_rate": 3.5693293623207086e-06,
607
+ "loss": 0.0022,
608
+ "num_input_tokens_seen": 361928,
609
+ "step": 375
610
+ },
611
+ {
612
+ "epoch": 2.5165562913907285,
613
+ "grad_norm": 0.3307873010635376,
614
+ "learning_rate": 3.135908382160771e-06,
615
+ "loss": 0.0061,
616
+ "num_input_tokens_seen": 366632,
617
+ "step": 380
618
+ },
619
+ {
620
+ "epoch": 2.5496688741721854,
621
+ "grad_norm": 0.31569671630859375,
622
+ "learning_rate": 2.728773811691923e-06,
623
+ "loss": 0.0077,
624
+ "num_input_tokens_seen": 371352,
625
+ "step": 385
626
+ },
627
+ {
628
+ "epoch": 2.5827814569536423,
629
+ "grad_norm": 0.25220000743865967,
630
+ "learning_rate": 2.348415134141102e-06,
631
+ "loss": 0.0053,
632
+ "num_input_tokens_seen": 375976,
633
+ "step": 390
634
+ },
635
+ {
636
+ "epoch": 2.6158940397350996,
637
+ "grad_norm": 0.12679828703403473,
638
+ "learning_rate": 1.995289641043768e-06,
639
+ "loss": 0.0021,
640
+ "num_input_tokens_seen": 381016,
641
+ "step": 395
642
+ },
643
+ {
644
+ "epoch": 2.6490066225165565,
645
+ "grad_norm": 0.2858022153377533,
646
+ "learning_rate": 1.6698218824588164e-06,
647
+ "loss": 0.0063,
648
+ "num_input_tokens_seen": 385688,
649
+ "step": 400
650
+ },
651
+ {
652
+ "epoch": 2.6821192052980134,
653
+ "grad_norm": 0.3448869287967682,
654
+ "learning_rate": 1.3724031565473112e-06,
655
+ "loss": 0.0052,
656
+ "num_input_tokens_seen": 390584,
657
+ "step": 405
658
+ },
659
+ {
660
+ "epoch": 2.7152317880794703,
661
+ "grad_norm": 0.26856377720832825,
662
+ "learning_rate": 1.1033910391288065e-06,
663
+ "loss": 0.0051,
664
+ "num_input_tokens_seen": 395576,
665
+ "step": 410
666
+ },
667
+ {
668
+ "epoch": 2.748344370860927,
669
+ "grad_norm": 0.3648998737335205,
670
+ "learning_rate": 8.631089537808307e-07,
671
+ "loss": 0.0084,
672
+ "num_input_tokens_seen": 400328,
673
+ "step": 415
674
+ },
675
+ {
676
+ "epoch": 2.781456953642384,
677
+ "grad_norm": 1.5257205963134766,
678
+ "learning_rate": 6.51845782998356e-07,
679
+ "loss": 0.0122,
680
+ "num_input_tokens_seen": 405192,
681
+ "step": 420
682
+ },
683
+ {
684
+ "epoch": 2.814569536423841,
685
+ "grad_norm": 0.4184490144252777,
686
+ "learning_rate": 4.698555208807853e-07,
687
+ "loss": 0.006,
688
+ "num_input_tokens_seen": 409912,
689
+ "step": 425
690
+ },
691
+ {
692
+ "epoch": 2.847682119205298,
693
+ "grad_norm": 0.39113691449165344,
694
+ "learning_rate": 3.1735696776400703e-07,
695
+ "loss": 0.0044,
696
+ "num_input_tokens_seen": 414792,
697
+ "step": 430
698
+ },
699
+ {
700
+ "epoch": 2.880794701986755,
701
+ "grad_norm": 0.49995261430740356,
702
+ "learning_rate": 1.9453346716462317e-07,
703
+ "loss": 0.0091,
704
+ "num_input_tokens_seen": 419752,
705
+ "step": 435
706
+ },
707
+ {
708
+ "epoch": 2.9139072847682117,
709
+ "grad_norm": 0.45024681091308594,
710
+ "learning_rate": 1.0153268535264827e-07,
711
+ "loss": 0.005,
712
+ "num_input_tokens_seen": 424808,
713
+ "step": 440
714
+ },
715
+ {
716
+ "epoch": 2.9470198675496686,
717
+ "grad_norm": 0.11354193091392517,
718
+ "learning_rate": 3.846643381766879e-08,
719
+ "loss": 0.0051,
720
+ "num_input_tokens_seen": 429704,
721
+ "step": 445
722
+ },
723
+ {
724
+ "epoch": 2.980132450331126,
725
+ "grad_norm": 0.07164571434259415,
726
+ "learning_rate": 5.41053484192644e-09,
727
+ "loss": 0.0049,
728
+ "num_input_tokens_seen": 434536,
729
+ "step": 450
730
+ },
731
+ {
732
+ "epoch": 3.0,
733
+ "num_input_tokens_seen": 437216,
734
+ "step": 453,
735
+ "total_flos": 1.8708380691726336e+16,
736
+ "train_loss": 0.044493223137528556,
737
+ "train_runtime": 988.3228,
738
+ "train_samples_per_second": 7.331,
739
+ "train_steps_per_second": 0.458
740
+ }
741
+ ],
742
+ "logging_steps": 5,
743
+ "max_steps": 453,
744
+ "num_input_tokens_seen": 437216,
745
+ "num_train_epochs": 3,
746
+ "save_steps": 100,
747
+ "stateful_callbacks": {
748
+ "TrainerControl": {
749
+ "args": {
750
+ "should_epoch_stop": false,
751
+ "should_evaluate": false,
752
+ "should_log": false,
753
+ "should_save": true,
754
+ "should_training_stop": true
755
+ },
756
+ "attributes": {}
757
+ }
758
+ },
759
+ "total_flos": 1.8708380691726336e+16,
760
+ "train_batch_size": 2,
761
+ "trial_name": null,
762
+ "trial_params": null
763
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddac98dc4cdbbb7bedc33599ee666b2ce8e71e9f161cc8201762f26175bbe62e
3
+ size 5368
training_args.yaml ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ bf16: true
2
+ cutoff_len: 1024
3
+ dataset: minecraft
4
+ dataset_dir: data
5
+ ddp_timeout: 180000000
6
+ do_train: true
7
+ finetuning_type: lora
8
+ flash_attn: auto
9
+ gradient_accumulation_steps: 8
10
+ include_num_input_tokens_seen: true
11
+ learning_rate: 5.0e-05
12
+ logging_steps: 5
13
+ lora_alpha: 16
14
+ lora_dropout: 0
15
+ lora_rank: 8
16
+ lora_target: all
17
+ lr_scheduler_type: cosine
18
+ max_grad_norm: 1.0
19
+ max_samples: 100000
20
+ model_name_or_path: mistralai/Mistral-7B-Instruct-v0.1
21
+ num_train_epochs: 3.0
22
+ optim: adamw_torch
23
+ output_dir: saves/Mistral-7B-v0.1-Chat/lora/mistral_docker
24
+ packing: false
25
+ per_device_train_batch_size: 2
26
+ plot_loss: true
27
+ preprocessing_num_workers: 16
28
+ report_to: none
29
+ save_steps: 100
30
+ stage: sft
31
+ template: mistral
32
+ warmup_steps: 0
training_loss.png ADDED