Tippawan commited on
Commit
9d8cc01
·
verified ·
1 Parent(s): 4c0c8de

Model save

Browse files
README.md ADDED
@@ -0,0 +1,62 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: peft
3
+ license: apache-2.0
4
+ base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
5
+ tags:
6
+ - base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0
7
+ - llama-factory
8
+ - lora
9
+ - transformers
10
+ pipeline_tag: text-generation
11
+ model-index:
12
+ - name: tinyllama-codeHtml
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # tinyllama-codeHtml
20
+
21
+ This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-Chat-v1.0](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0) on an unknown dataset.
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 5e-05
41
+ - train_batch_size: 2
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - gradient_accumulation_steps: 4
45
+ - total_train_batch_size: 8
46
+ - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
+ - lr_scheduler_type: cosine
48
+ - lr_scheduler_warmup_ratio: 0.1
49
+ - num_epochs: 3.0
50
+ - mixed_precision_training: Native AMP
51
+
52
+ ### Training results
53
+
54
+
55
+
56
+ ### Framework versions
57
+
58
+ - PEFT 0.17.1
59
+ - Transformers 4.57.1
60
+ - Pytorch 2.8.0+cu126
61
+ - Datasets 4.0.0
62
+ - Tokenizers 0.22.1
adapter_config.json ADDED
@@ -0,0 +1,42 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "alpha_pattern": {},
3
+ "auto_mapping": null,
4
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
5
+ "bias": "none",
6
+ "corda_config": null,
7
+ "eva_config": null,
8
+ "exclude_modules": null,
9
+ "fan_in_fan_out": false,
10
+ "inference_mode": true,
11
+ "init_lora_weights": true,
12
+ "layer_replication": null,
13
+ "layers_pattern": null,
14
+ "layers_to_transform": null,
15
+ "loftq_config": {},
16
+ "lora_alpha": 16,
17
+ "lora_bias": false,
18
+ "lora_dropout": 0.0,
19
+ "megatron_config": null,
20
+ "megatron_core": "megatron.core",
21
+ "modules_to_save": null,
22
+ "peft_type": "LORA",
23
+ "qalora_group_size": 16,
24
+ "r": 8,
25
+ "rank_pattern": {},
26
+ "revision": null,
27
+ "target_modules": [
28
+ "v_proj",
29
+ "up_proj",
30
+ "o_proj",
31
+ "q_proj",
32
+ "k_proj",
33
+ "gate_proj",
34
+ "down_proj"
35
+ ],
36
+ "target_parameters": null,
37
+ "task_type": "CAUSAL_LM",
38
+ "trainable_token_indices": null,
39
+ "use_dora": false,
40
+ "use_qalora": false,
41
+ "use_rslora": false
42
+ }
adapter_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9eed526dfb74b7f26d47ff5921dc19de42e46f31ed18d900149beeb07b417ab1
3
+ size 25271744
all_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "total_flos": 1422043529084928.0,
4
+ "train_loss": 0.12496965477075526,
5
+ "train_runtime": 258.8654,
6
+ "train_samples_per_second": 5.795,
7
+ "train_steps_per_second": 0.73
8
+ }
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ 'System: ' + system_message + '</s>' + '
2
+ ' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'Human: ' + content + '</s>' + '
3
+ Assistant:' }}{% elif message['role'] == 'assistant' %}{{ content + '</s>' + '
4
+ ' }}{% endif %}{% endfor %}
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "</s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": null,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "extra_special_tokens": {},
35
+ "legacy": false,
36
+ "model_max_length": 2048,
37
+ "pad_token": "</s>",
38
+ "padding_side": "right",
39
+ "sp_model_kwargs": {},
40
+ "split_special_tokens": false,
41
+ "tokenizer_class": "LlamaTokenizer",
42
+ "unk_token": "<unk>",
43
+ "use_default_system_prompt": false
44
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.0,
3
+ "total_flos": 1422043529084928.0,
4
+ "train_loss": 0.12496965477075526,
5
+ "train_runtime": 258.8654,
6
+ "train_samples_per_second": 5.795,
7
+ "train_steps_per_second": 0.73
8
+ }
trainer_log.jsonl ADDED
@@ -0,0 +1,39 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {"current_steps": 5, "total_steps": 189, "loss": 1.6436, "lr": 1.0526315789473684e-05, "epoch": 0.08, "percentage": 2.65, "elapsed_time": "0:00:08", "remaining_time": "0:05:10"}
2
+ {"current_steps": 10, "total_steps": 189, "loss": 0.772, "lr": 2.368421052631579e-05, "epoch": 0.16, "percentage": 5.29, "elapsed_time": "0:00:13", "remaining_time": "0:04:00"}
3
+ {"current_steps": 15, "total_steps": 189, "loss": 0.1927, "lr": 3.6842105263157895e-05, "epoch": 0.24, "percentage": 7.94, "elapsed_time": "0:00:19", "remaining_time": "0:03:42"}
4
+ {"current_steps": 20, "total_steps": 189, "loss": 0.08, "lr": 5e-05, "epoch": 0.32, "percentage": 10.58, "elapsed_time": "0:00:24", "remaining_time": "0:03:24"}
5
+ {"current_steps": 25, "total_steps": 189, "loss": 0.0735, "lr": 4.989335440737586e-05, "epoch": 0.4, "percentage": 13.23, "elapsed_time": "0:00:29", "remaining_time": "0:03:14"}
6
+ {"current_steps": 30, "total_steps": 189, "loss": 0.0719, "lr": 4.957432749209755e-05, "epoch": 0.48, "percentage": 15.87, "elapsed_time": "0:00:34", "remaining_time": "0:03:05"}
7
+ {"current_steps": 35, "total_steps": 189, "loss": 0.0702, "lr": 4.9045641079320484e-05, "epoch": 0.56, "percentage": 18.52, "elapsed_time": "0:00:40", "remaining_time": "0:02:59"}
8
+ {"current_steps": 40, "total_steps": 189, "loss": 0.071, "lr": 4.8311805735108894e-05, "epoch": 0.64, "percentage": 21.16, "elapsed_time": "0:00:46", "remaining_time": "0:02:52"}
9
+ {"current_steps": 45, "total_steps": 189, "loss": 0.0677, "lr": 4.7379082283876566e-05, "epoch": 0.72, "percentage": 23.81, "elapsed_time": "0:00:51", "remaining_time": "0:02:44"}
10
+ {"current_steps": 50, "total_steps": 189, "loss": 0.0599, "lr": 4.625542839324036e-05, "epoch": 0.8, "percentage": 26.46, "elapsed_time": "0:00:57", "remaining_time": "0:02:39"}
11
+ {"current_steps": 55, "total_steps": 189, "loss": 0.0685, "lr": 4.4950430682006e-05, "epoch": 0.88, "percentage": 29.1, "elapsed_time": "0:01:02", "remaining_time": "0:02:31"}
12
+ {"current_steps": 60, "total_steps": 189, "loss": 0.0653, "lr": 4.347522293051648e-05, "epoch": 0.96, "percentage": 31.75, "elapsed_time": "0:01:07", "remaining_time": "0:02:25"}
13
+ {"current_steps": 65, "total_steps": 189, "loss": 0.0655, "lr": 4.184239109116393e-05, "epoch": 1.032, "percentage": 34.39, "elapsed_time": "0:01:12", "remaining_time": "0:02:18"}
14
+ {"current_steps": 70, "total_steps": 189, "loss": 0.0629, "lr": 4.0065865909481417e-05, "epoch": 1.112, "percentage": 37.04, "elapsed_time": "0:01:22", "remaining_time": "0:02:19"}
15
+ {"current_steps": 75, "total_steps": 189, "loss": 0.0594, "lr": 3.81608040719339e-05, "epoch": 1.192, "percentage": 39.68, "elapsed_time": "0:01:33", "remaining_time": "0:02:21"}
16
+ {"current_steps": 80, "total_steps": 189, "loss": 0.0594, "lr": 3.6143458894413465e-05, "epoch": 1.272, "percentage": 42.33, "elapsed_time": "0:01:43", "remaining_time": "0:02:20"}
17
+ {"current_steps": 85, "total_steps": 189, "loss": 0.058, "lr": 3.403104165467883e-05, "epoch": 1.3519999999999999, "percentage": 44.97, "elapsed_time": "0:01:52", "remaining_time": "0:02:17"}
18
+ {"current_steps": 90, "total_steps": 189, "loss": 0.0598, "lr": 3.1841574751802076e-05, "epoch": 1.432, "percentage": 47.62, "elapsed_time": "0:01:59", "remaining_time": "0:02:11"}
19
+ {"current_steps": 95, "total_steps": 189, "loss": 0.064, "lr": 2.9593737945414264e-05, "epoch": 1.512, "percentage": 50.26, "elapsed_time": "0:02:04", "remaining_time": "0:02:03"}
20
+ {"current_steps": 100, "total_steps": 189, "loss": 0.0629, "lr": 2.7306708986582553e-05, "epoch": 1.592, "percentage": 52.91, "elapsed_time": "0:02:10", "remaining_time": "0:01:56"}
21
+ {"current_steps": 105, "total_steps": 189, "loss": 0.0609, "lr": 2.5e-05, "epoch": 1.6720000000000002, "percentage": 55.56, "elapsed_time": "0:02:15", "remaining_time": "0:01:48"}
22
+ {"current_steps": 110, "total_steps": 189, "loss": 0.0641, "lr": 2.2693291013417453e-05, "epoch": 1.752, "percentage": 58.2, "elapsed_time": "0:02:20", "remaining_time": "0:01:41"}
23
+ {"current_steps": 115, "total_steps": 189, "loss": 0.0602, "lr": 2.0406262054585738e-05, "epoch": 1.8319999999999999, "percentage": 60.85, "elapsed_time": "0:02:26", "remaining_time": "0:01:34"}
24
+ {"current_steps": 120, "total_steps": 189, "loss": 0.0604, "lr": 1.815842524819793e-05, "epoch": 1.912, "percentage": 63.49, "elapsed_time": "0:02:33", "remaining_time": "0:01:28"}
25
+ {"current_steps": 125, "total_steps": 189, "loss": 0.0557, "lr": 1.5968958345321178e-05, "epoch": 1.992, "percentage": 66.14, "elapsed_time": "0:02:44", "remaining_time": "0:01:24"}
26
+ {"current_steps": 130, "total_steps": 189, "loss": 0.0545, "lr": 1.3856541105586545e-05, "epoch": 2.064, "percentage": 68.78, "elapsed_time": "0:02:54", "remaining_time": "0:01:19"}
27
+ {"current_steps": 135, "total_steps": 189, "loss": 0.0503, "lr": 1.1839195928066102e-05, "epoch": 2.144, "percentage": 71.43, "elapsed_time": "0:03:06", "remaining_time": "0:01:14"}
28
+ {"current_steps": 140, "total_steps": 189, "loss": 0.0543, "lr": 9.934134090518593e-06, "epoch": 2.224, "percentage": 74.07, "elapsed_time": "0:03:16", "remaining_time": "0:01:08"}
29
+ {"current_steps": 145, "total_steps": 189, "loss": 0.0538, "lr": 8.15760890883607e-06, "epoch": 2.304, "percentage": 76.72, "elapsed_time": "0:03:24", "remaining_time": "0:01:01"}
30
+ {"current_steps": 150, "total_steps": 189, "loss": 0.0521, "lr": 6.524777069483526e-06, "epoch": 2.384, "percentage": 79.37, "elapsed_time": "0:03:29", "remaining_time": "0:00:54"}
31
+ {"current_steps": 155, "total_steps": 189, "loss": 0.057, "lr": 5.049569317994013e-06, "epoch": 2.464, "percentage": 82.01, "elapsed_time": "0:03:37", "remaining_time": "0:00:47"}
32
+ {"current_steps": 160, "total_steps": 189, "loss": 0.0545, "lr": 3.7445716067596503e-06, "epoch": 2.544, "percentage": 84.66, "elapsed_time": "0:03:42", "remaining_time": "0:00:40"}
33
+ {"current_steps": 165, "total_steps": 189, "loss": 0.0544, "lr": 2.6209177161234445e-06, "epoch": 2.624, "percentage": 87.3, "elapsed_time": "0:03:50", "remaining_time": "0:00:33"}
34
+ {"current_steps": 170, "total_steps": 189, "loss": 0.0533, "lr": 1.6881942648911076e-06, "epoch": 2.7039999999999997, "percentage": 89.95, "elapsed_time": "0:03:56", "remaining_time": "0:00:26"}
35
+ {"current_steps": 175, "total_steps": 189, "loss": 0.054, "lr": 9.54358920679524e-07, "epoch": 2.784, "percentage": 92.59, "elapsed_time": "0:04:02", "remaining_time": "0:00:19"}
36
+ {"current_steps": 180, "total_steps": 189, "loss": 0.0549, "lr": 4.256725079024554e-07, "epoch": 2.864, "percentage": 95.24, "elapsed_time": "0:04:09", "remaining_time": "0:00:12"}
37
+ {"current_steps": 185, "total_steps": 189, "loss": 0.0532, "lr": 1.0664559262413831e-07, "epoch": 2.944, "percentage": 97.88, "elapsed_time": "0:04:14", "remaining_time": "0:00:05"}
38
+ {"current_steps": 189, "total_steps": 189, "epoch": 3.0, "percentage": 100.0, "elapsed_time": "0:04:18", "remaining_time": "0:00:00"}
39
+ {"current_steps": 189, "total_steps": 189, "epoch": 3.0, "percentage": 100.0, "elapsed_time": "0:00:00", "remaining_time": "0:00:00"}
trainer_state.json ADDED
@@ -0,0 +1,302 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_global_step": null,
3
+ "best_metric": null,
4
+ "best_model_checkpoint": null,
5
+ "epoch": 3.0,
6
+ "eval_steps": 500,
7
+ "global_step": 189,
8
+ "is_hyper_param_search": false,
9
+ "is_local_process_zero": true,
10
+ "is_world_process_zero": true,
11
+ "log_history": [
12
+ {
13
+ "epoch": 0.08,
14
+ "grad_norm": 2.2252185344696045,
15
+ "learning_rate": 1.0526315789473684e-05,
16
+ "loss": 1.6436,
17
+ "step": 5
18
+ },
19
+ {
20
+ "epoch": 0.16,
21
+ "grad_norm": 1.4878095388412476,
22
+ "learning_rate": 2.368421052631579e-05,
23
+ "loss": 0.772,
24
+ "step": 10
25
+ },
26
+ {
27
+ "epoch": 0.24,
28
+ "grad_norm": 0.5770473480224609,
29
+ "learning_rate": 3.6842105263157895e-05,
30
+ "loss": 0.1927,
31
+ "step": 15
32
+ },
33
+ {
34
+ "epoch": 0.32,
35
+ "grad_norm": 0.24875187873840332,
36
+ "learning_rate": 5e-05,
37
+ "loss": 0.08,
38
+ "step": 20
39
+ },
40
+ {
41
+ "epoch": 0.4,
42
+ "grad_norm": 0.2741605043411255,
43
+ "learning_rate": 4.989335440737586e-05,
44
+ "loss": 0.0735,
45
+ "step": 25
46
+ },
47
+ {
48
+ "epoch": 0.48,
49
+ "grad_norm": 0.10829973965883255,
50
+ "learning_rate": 4.957432749209755e-05,
51
+ "loss": 0.0719,
52
+ "step": 30
53
+ },
54
+ {
55
+ "epoch": 0.56,
56
+ "grad_norm": 0.16560488939285278,
57
+ "learning_rate": 4.9045641079320484e-05,
58
+ "loss": 0.0702,
59
+ "step": 35
60
+ },
61
+ {
62
+ "epoch": 0.64,
63
+ "grad_norm": 0.1393565684556961,
64
+ "learning_rate": 4.8311805735108894e-05,
65
+ "loss": 0.071,
66
+ "step": 40
67
+ },
68
+ {
69
+ "epoch": 0.72,
70
+ "grad_norm": 0.1896258294582367,
71
+ "learning_rate": 4.7379082283876566e-05,
72
+ "loss": 0.0677,
73
+ "step": 45
74
+ },
75
+ {
76
+ "epoch": 0.8,
77
+ "grad_norm": 0.1210947260260582,
78
+ "learning_rate": 4.625542839324036e-05,
79
+ "loss": 0.0599,
80
+ "step": 50
81
+ },
82
+ {
83
+ "epoch": 0.88,
84
+ "grad_norm": 0.16256098449230194,
85
+ "learning_rate": 4.4950430682006e-05,
86
+ "loss": 0.0685,
87
+ "step": 55
88
+ },
89
+ {
90
+ "epoch": 0.96,
91
+ "grad_norm": 0.10290851444005966,
92
+ "learning_rate": 4.347522293051648e-05,
93
+ "loss": 0.0653,
94
+ "step": 60
95
+ },
96
+ {
97
+ "epoch": 1.032,
98
+ "grad_norm": 0.10546600073575974,
99
+ "learning_rate": 4.184239109116393e-05,
100
+ "loss": 0.0655,
101
+ "step": 65
102
+ },
103
+ {
104
+ "epoch": 1.112,
105
+ "grad_norm": 0.09690473228693008,
106
+ "learning_rate": 4.0065865909481417e-05,
107
+ "loss": 0.0629,
108
+ "step": 70
109
+ },
110
+ {
111
+ "epoch": 1.192,
112
+ "grad_norm": 0.11338075995445251,
113
+ "learning_rate": 3.81608040719339e-05,
114
+ "loss": 0.0594,
115
+ "step": 75
116
+ },
117
+ {
118
+ "epoch": 1.272,
119
+ "grad_norm": 0.11339253932237625,
120
+ "learning_rate": 3.6143458894413465e-05,
121
+ "loss": 0.0594,
122
+ "step": 80
123
+ },
124
+ {
125
+ "epoch": 1.3519999999999999,
126
+ "grad_norm": 0.18642409145832062,
127
+ "learning_rate": 3.403104165467883e-05,
128
+ "loss": 0.058,
129
+ "step": 85
130
+ },
131
+ {
132
+ "epoch": 1.432,
133
+ "grad_norm": 0.16320252418518066,
134
+ "learning_rate": 3.1841574751802076e-05,
135
+ "loss": 0.0598,
136
+ "step": 90
137
+ },
138
+ {
139
+ "epoch": 1.512,
140
+ "grad_norm": 0.1201169490814209,
141
+ "learning_rate": 2.9593737945414264e-05,
142
+ "loss": 0.064,
143
+ "step": 95
144
+ },
145
+ {
146
+ "epoch": 1.592,
147
+ "grad_norm": 0.18611621856689453,
148
+ "learning_rate": 2.7306708986582553e-05,
149
+ "loss": 0.0629,
150
+ "step": 100
151
+ },
152
+ {
153
+ "epoch": 1.6720000000000002,
154
+ "grad_norm": 0.09128480404615402,
155
+ "learning_rate": 2.5e-05,
156
+ "loss": 0.0609,
157
+ "step": 105
158
+ },
159
+ {
160
+ "epoch": 1.752,
161
+ "grad_norm": 0.1125178411602974,
162
+ "learning_rate": 2.2693291013417453e-05,
163
+ "loss": 0.0641,
164
+ "step": 110
165
+ },
166
+ {
167
+ "epoch": 1.8319999999999999,
168
+ "grad_norm": 0.1417843997478485,
169
+ "learning_rate": 2.0406262054585738e-05,
170
+ "loss": 0.0602,
171
+ "step": 115
172
+ },
173
+ {
174
+ "epoch": 1.912,
175
+ "grad_norm": 0.10574869811534882,
176
+ "learning_rate": 1.815842524819793e-05,
177
+ "loss": 0.0604,
178
+ "step": 120
179
+ },
180
+ {
181
+ "epoch": 1.992,
182
+ "grad_norm": 0.12096602469682693,
183
+ "learning_rate": 1.5968958345321178e-05,
184
+ "loss": 0.0557,
185
+ "step": 125
186
+ },
187
+ {
188
+ "epoch": 2.064,
189
+ "grad_norm": 0.13963642716407776,
190
+ "learning_rate": 1.3856541105586545e-05,
191
+ "loss": 0.0545,
192
+ "step": 130
193
+ },
194
+ {
195
+ "epoch": 2.144,
196
+ "grad_norm": 0.10345666855573654,
197
+ "learning_rate": 1.1839195928066102e-05,
198
+ "loss": 0.0503,
199
+ "step": 135
200
+ },
201
+ {
202
+ "epoch": 2.224,
203
+ "grad_norm": 0.11484814435243607,
204
+ "learning_rate": 9.934134090518593e-06,
205
+ "loss": 0.0543,
206
+ "step": 140
207
+ },
208
+ {
209
+ "epoch": 2.304,
210
+ "grad_norm": 0.1633157581090927,
211
+ "learning_rate": 8.15760890883607e-06,
212
+ "loss": 0.0538,
213
+ "step": 145
214
+ },
215
+ {
216
+ "epoch": 2.384,
217
+ "grad_norm": 0.13634610176086426,
218
+ "learning_rate": 6.524777069483526e-06,
219
+ "loss": 0.0521,
220
+ "step": 150
221
+ },
222
+ {
223
+ "epoch": 2.464,
224
+ "grad_norm": 0.12374045699834824,
225
+ "learning_rate": 5.049569317994013e-06,
226
+ "loss": 0.057,
227
+ "step": 155
228
+ },
229
+ {
230
+ "epoch": 2.544,
231
+ "grad_norm": 0.11331922560930252,
232
+ "learning_rate": 3.7445716067596503e-06,
233
+ "loss": 0.0545,
234
+ "step": 160
235
+ },
236
+ {
237
+ "epoch": 2.624,
238
+ "grad_norm": 0.12591035664081573,
239
+ "learning_rate": 2.6209177161234445e-06,
240
+ "loss": 0.0544,
241
+ "step": 165
242
+ },
243
+ {
244
+ "epoch": 2.7039999999999997,
245
+ "grad_norm": 0.13460087776184082,
246
+ "learning_rate": 1.6881942648911076e-06,
247
+ "loss": 0.0533,
248
+ "step": 170
249
+ },
250
+ {
251
+ "epoch": 2.784,
252
+ "grad_norm": 0.16594184935092926,
253
+ "learning_rate": 9.54358920679524e-07,
254
+ "loss": 0.054,
255
+ "step": 175
256
+ },
257
+ {
258
+ "epoch": 2.864,
259
+ "grad_norm": 0.13098712265491486,
260
+ "learning_rate": 4.256725079024554e-07,
261
+ "loss": 0.0549,
262
+ "step": 180
263
+ },
264
+ {
265
+ "epoch": 2.944,
266
+ "grad_norm": 0.10799074172973633,
267
+ "learning_rate": 1.0664559262413831e-07,
268
+ "loss": 0.0532,
269
+ "step": 185
270
+ },
271
+ {
272
+ "epoch": 3.0,
273
+ "step": 189,
274
+ "total_flos": 1422043529084928.0,
275
+ "train_loss": 0.12496965477075526,
276
+ "train_runtime": 258.8654,
277
+ "train_samples_per_second": 5.795,
278
+ "train_steps_per_second": 0.73
279
+ }
280
+ ],
281
+ "logging_steps": 5,
282
+ "max_steps": 189,
283
+ "num_input_tokens_seen": 0,
284
+ "num_train_epochs": 3,
285
+ "save_steps": 1000,
286
+ "stateful_callbacks": {
287
+ "TrainerControl": {
288
+ "args": {
289
+ "should_epoch_stop": false,
290
+ "should_evaluate": false,
291
+ "should_log": false,
292
+ "should_save": true,
293
+ "should_training_stop": true
294
+ },
295
+ "attributes": {}
296
+ }
297
+ },
298
+ "total_flos": 1422043529084928.0,
299
+ "train_batch_size": 2,
300
+ "trial_name": null,
301
+ "trial_params": null
302
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bac5f09bc422277a87df95233ec3d0e5e5fbc31b9f2198abfe7108c358d5104e
3
+ size 6225