aq1048576 commited on
Commit
30f1c9b
·
verified ·
1 Parent(s): ef16d42

Upload table SFT checkpoint

Browse files
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: meta-llama/Llama-3.1-8B
3
+ library_name: transformers
4
+ model_name: table
5
+ tags:
6
+ - generated_from_trainer
7
+ - trl
8
+ - sft
9
+ licence: license
10
+ ---
11
+
12
+ # Model Card for table
13
+
14
+ This model is a fine-tuned version of [meta-llama/Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B).
15
+ It has been trained using [TRL](https://github.com/huggingface/trl).
16
+
17
+ ## Quick start
18
+
19
+ ```python
20
+ from transformers import pipeline
21
+
22
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
23
+ generator = pipeline("text-generation", model="None", device="cuda")
24
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
25
+ print(output["generated_text"])
26
+ ```
27
+
28
+ ## Training procedure
29
+
30
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/aq1048576-princeton-university/huggingface/runs/xz32vbyx)
31
+
32
+
33
+ This model was trained with SFT.
34
+
35
+ ### Framework versions
36
+
37
+ - TRL: 0.23.0
38
+ - Transformers: 4.57.1
39
+ - Pytorch: 2.8.0
40
+ - Datasets: 3.5.0
41
+ - Tokenizers: 0.22.1
42
+
43
+ ## Citations
44
+
45
+
46
+
47
+ Cite TRL as:
48
+
49
+ ```bibtex
50
+ @misc{vonwerra2022trl,
51
+ title = {{TRL: Transformer Reinforcement Learning}},
52
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
53
+ year = 2020,
54
+ journal = {GitHub repository},
55
+ publisher = {GitHub},
56
+ howpublished = {\url{https://github.com/huggingface/trl}}
57
+ }
58
+ ```
model-00001-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1098dbbb69a063c169d800166d3db23baf269fd4c4823dfeaf6e2aa44262bc44
3
  size 4976698672
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:17390f599a09131d867f5082e79bbfd6bf9739f03fe5201302d31df5ba79c615
3
  size 4976698672
model-00002-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a873d99e2675d4ae38b952373db4655873ffacb59d9f9f0a9734596035fb0b41
3
  size 4999802720
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:868b92a6b2a57794cf7b21af4fc6733d6b2001cca65023dee0482029070f643f
3
  size 4999802720
model-00003-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:860dfed311f046f27da6ff4cd75b51251bb42bbab2f4da8b296a199be3bcf4ac
3
  size 4915916176
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc7fdc9302d0a044f76d9eca53d469df7429d4f4e95dfa21c32f8705cabbe97e
3
  size 4915916176
model-00004-of-00004.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a92d43a37ba8fb61c60b1f045957357bacaf34c3e1436e50c34decbfdaaeaca3
3
  size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61f5926292a1202347dc6501f27f48facefcc6bb7e1efbdf47a924b8db713795
3
  size 1168138808
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2f90a0ee1b41702c7b233b02234294a53bc0684a08d3bcd8c8ff702e9a12f64
3
- size 17210019
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b9e4e7fb171f92fd137b777cc2714bf87d11576700a1dcd7a399e7bbe39537b
3
+ size 17209920
tokenizer_config.json CHANGED
@@ -2059,5 +2059,6 @@
2059
  ],
2060
  "model_max_length": 4096,
2061
  "pad_token": "<|end_of_text|>",
2062
- "tokenizer_class": "PreTrainedTokenizerFast"
 
2063
  }
 
2059
  ],
2060
  "model_max_length": 4096,
2061
  "pad_token": "<|end_of_text|>",
2062
+ "tokenizer_class": "PreTrainedTokenizerFast",
2063
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{ '<|system|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'user' %}{{ '<|user|>\n' + message['content'] + '\n' }}{% elif message['role'] == 'assistant' %}{% if not loop.last %}{{ '<|assistant|>\n' + message['content'] + eos_token + '\n' }}{% else %}{{ '<|assistant|>\n' + message['content'] + eos_token }}{% endif %}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<|assistant|>\n' }}{% endif %}{% endfor %}"
2064
  }
trainer_state.json CHANGED
@@ -4,30 +4,26 @@
4
  "best_model_checkpoint": null,
5
  "epoch": 1.0,
6
  "eval_steps": 500,
7
- "global_step": 20,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
 
13
  "epoch": 1.0,
14
- "grad_norm": 1.3046875,
15
- "learning_rate": 3.4096741493194197e-07,
16
- "loss": 1.1363,
17
- "step": 20
18
- },
19
- {
20
- "epoch": 1.0,
21
- "step": 20,
22
- "total_flos": 3.434522481451336e+17,
23
- "train_loss": 1.136316967010498,
24
- "train_runtime": 261.6908,
25
- "train_samples_per_second": 18.915,
26
- "train_steps_per_second": 0.076
27
  }
28
  ],
29
- "logging_steps": 20,
30
- "max_steps": 20,
31
  "num_input_tokens_seen": 0,
32
  "num_train_epochs": 1,
33
  "save_steps": 500,
@@ -43,7 +39,7 @@
43
  "attributes": {}
44
  }
45
  },
46
- "total_flos": 3.434522481451336e+17,
47
  "train_batch_size": 4,
48
  "trial_name": null,
49
  "trial_params": null
 
4
  "best_model_checkpoint": null,
5
  "epoch": 1.0,
6
  "eval_steps": 500,
7
+ "global_step": 52,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
11
  "log_history": [
12
  {
13
+ "entropy": 0.7864327458146121,
14
  "epoch": 1.0,
15
+ "mean_token_accuracy": 0.8194242536877315,
16
+ "num_tokens": 2959320.0,
17
+ "step": 52,
18
+ "total_flos": 2.0714871222750413e+17,
19
+ "train_loss": 0.8364706039428711,
20
+ "train_runtime": 316.5604,
21
+ "train_samples_per_second": 15.637,
22
+ "train_steps_per_second": 0.164
 
 
 
 
 
23
  }
24
  ],
25
+ "logging_steps": 500,
26
+ "max_steps": 52,
27
  "num_input_tokens_seen": 0,
28
  "num_train_epochs": 1,
29
  "save_steps": 500,
 
39
  "attributes": {}
40
  }
41
  },
42
+ "total_flos": 2.0714871222750413e+17,
43
  "train_batch_size": 4,
44
  "trial_name": null,
45
  "trial_params": null
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b88a9b23344ae5f92849fdcdce40ebff35bbe01c31a6da2f2d6b19bbeab64a25
3
- size 5841
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ffddec3fcb612943270afb048c817d5e689eabd08f240219e334957235badf64
3
+ size 6289