25b3nk commited on
Commit
0c8ece3
·
verified ·
1 Parent(s): ff714d4

End of training

Browse files
README.md CHANGED
@@ -29,18 +29,18 @@ print(output["generated_text"])
29
 
30
  ## Training procedure
31
 
32
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/25b3nk-eagle-eye-networks/huggingface/runs/b75ck0tz)
33
 
34
 
35
  This model was trained with SFT.
36
 
37
  ### Framework versions
38
 
39
- - TRL: 0.15.2
40
- - Transformers: 4.48.3
41
- - Pytorch: 2.5.1+cu124
42
- - Datasets: 3.3.2
43
- - Tokenizers: 0.21.0
44
 
45
  ## Citations
46
 
@@ -51,7 +51,7 @@ Cite TRL as:
51
  ```bibtex
52
  @misc{vonwerra2022trl,
53
  title = {{TRL: Transformer Reinforcement Learning}},
54
- author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
55
  year = 2020,
56
  journal = {GitHub repository},
57
  publisher = {GitHub},
 
29
 
30
  ## Training procedure
31
 
32
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/25b3nk-eagle-eye-networks/huggingface/runs/t1moa5z7)
33
 
34
 
35
  This model was trained with SFT.
36
 
37
  ### Framework versions
38
 
39
+ - TRL: 0.19.1
40
+ - Transformers: 4.53.3
41
+ - Pytorch: 2.6.0+cu124
42
+ - Datasets: 4.0.0
43
+ - Tokenizers: 0.21.2
44
 
45
  ## Citations
46
 
 
51
  ```bibtex
52
  @misc{vonwerra2022trl,
53
  title = {{TRL: Transformer Reinforcement Learning}},
54
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec},
55
  year = 2020,
56
  journal = {GitHub repository},
57
  publisher = {GitHub},
chat_template.jinja ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {% for message in messages %}{{'<|im_start|>' + message['role'] + '
2
+ ' + message['content'] + '<|im_end|>' + '
3
+ '}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant
4
+ ' }}{% endif %}
config.json CHANGED
@@ -1,5 +1,4 @@
1
  {
2
- "_name_or_path": "HuggingFaceTB/SmolLM2-135M",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -27,7 +26,7 @@
27
  "rope_theta": 100000,
28
  "tie_word_embeddings": true,
29
  "torch_dtype": "float32",
30
- "transformers_version": "4.48.3",
31
  "use_cache": true,
32
  "vocab_size": 49152
33
  }
 
1
  {
 
2
  "architectures": [
3
  "LlamaForCausalLM"
4
  ],
 
26
  "rope_theta": 100000,
27
  "tie_word_embeddings": true,
28
  "torch_dtype": "float32",
29
+ "transformers_version": "4.53.3",
30
  "use_cache": true,
31
  "vocab_size": 49152
32
  }
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 2,
6
- "transformers_version": "4.48.3"
7
  }
 
3
  "bos_token_id": 1,
4
  "eos_token_id": 2,
5
  "pad_token_id": 2,
6
+ "transformers_version": "4.53.3"
7
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:933a6c4a30c3f03ac8e9d1e398e01313cde51cc6ab5e82b6b940d562d0a1b7c0
3
  size 538090408
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:364c6ac4e942db610f94519fb4bffbde7c832226eca31e9d3968f3a911504afc
3
  size 538090408
runs/Jul28_11-45-57_1b06f6560f86/events.out.tfevents.1753703201.1b06f6560f86.507.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d973a3eb75065376d52f76aa628f1dbfddadbb8e7966bcd6be15dda545b3ca9f
3
+ size 46615
tokenizer_config.json CHANGED
@@ -143,7 +143,6 @@
143
  "<|im_end|>"
144
  ],
145
  "bos_token": "<|im_start|>",
146
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
147
  "clean_up_tokenization_spaces": false,
148
  "eos_token": "<|im_end|>",
149
  "extra_special_tokens": {},
 
143
  "<|im_end|>"
144
  ],
145
  "bos_token": "<|im_start|>",
 
146
  "clean_up_tokenization_spaces": false,
147
  "eos_token": "<|im_end|>",
148
  "extra_special_tokens": {},
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ac95dbe45e27b5b81d5400cf786b1e95cf88ff3541d557189477e1c61a4a55be
3
- size 5560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1440ef4dfe0a4313750b44bd6a0f1a9408b7189e3dfe6077f5560f06582e4f69
3
+ size 5752