NguyenDuyPhuc commited on
Commit
372dd10
·
verified ·
1 Parent(s): defb23c

End of training

Browse files
Files changed (35) hide show
  1. .gitattributes +1 -0
  2. README.md +59 -0
  3. added_tokens.json +24 -0
  4. config.json +29 -0
  5. generation_config.json +11 -0
  6. merges.txt +0 -0
  7. model.safetensors +3 -0
  8. runs/Apr01_14-39-31_caicc/events.out.tfevents.1743532773.caicc +3 -0
  9. runs/Apr01_14-44-03_caicc/events.out.tfevents.1743533070.caicc +3 -0
  10. runs/Apr01_14-48-52_caicc/events.out.tfevents.1743533353.caicc +3 -0
  11. runs/Apr01_15-00-03_caicc/events.out.tfevents.1743534245.caicc +3 -0
  12. runs/Apr01_15-18-41_caicc/events.out.tfevents.1743535315.caicc +3 -0
  13. runs/Apr01_15-23-03_caicc/events.out.tfevents.1743535383.caicc +3 -0
  14. runs/Apr01_15-27-21_caicc/events.out.tfevents.1743535836.caicc +3 -0
  15. runs/Apr01_15-32-07_caicc/events.out.tfevents.1743535928.caicc +3 -0
  16. runs/Apr01_15-58-49_caicc/events.out.tfevents.1743537529.caicc +3 -0
  17. runs/Apr01_15-59-16_caicc/events.out.tfevents.1743537556.caicc +3 -0
  18. runs/Apr01_16-07-22_caicc/events.out.tfevents.1743538089.caicc +3 -0
  19. runs/Apr01_16-08-37_caicc/events.out.tfevents.1743538152.caicc +3 -0
  20. runs/Apr01_16-09-43_caicc/events.out.tfevents.1743538222.caicc +3 -0
  21. runs/Apr01_16-14-20_caicc/events.out.tfevents.1743538502.caicc +3 -0
  22. runs/Apr01_16-16-21_caicc/events.out.tfevents.1743538584.caicc +3 -0
  23. runs/Mar31_00-09-41_caicc/events.out.tfevents.1743394189.caicc +3 -0
  24. runs/Mar31_00-14-18_caicc/events.out.tfevents.1743394460.caicc +3 -0
  25. runs/Mar31_00-14-51_caicc/events.out.tfevents.1743394492.caicc +3 -0
  26. runs/Mar31_00-19-52_caicc/events.out.tfevents.1743394794.caicc +3 -0
  27. runs/Mar31_00-25-08_caicc/events.out.tfevents.1743395109.caicc +3 -0
  28. runs/Mar31_00-30-22_caicc/events.out.tfevents.1743395426.caicc +3 -0
  29. runs/Mar31_00-32-20_caicc/events.out.tfevents.1743395544.caicc +3 -0
  30. runs/Mar31_00-42-27_caicc/events.out.tfevents.1743396680.caicc +3 -0
  31. special_tokens_map.json +21 -0
  32. tokenizer.json +3 -0
  33. tokenizer_config.json +197 -0
  34. training_args.bin +3 -0
  35. vocab.json +0 -0
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: Qwen/Qwen2.5-0.5B-Instruct
3
+ library_name: transformers
4
+ model_name: DistressAI
5
+ tags:
6
+ - generated_from_trainer
7
+ - DistressAI
8
+ - trl
9
+ - sft
10
+ licence: license
11
+ ---
12
+
13
+ # Model Card for DistressAI
14
+
15
+ This model is a fine-tuned version of [Qwen/Qwen2.5-0.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-0.5B-Instruct).
16
+ It has been trained using [TRL](https://github.com/huggingface/trl).
17
+
18
+ ## Quick start
19
+
20
+ ```python
21
+ from transformers import pipeline
22
+
23
+ question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?"
24
+ generator = pipeline("text-generation", model="NguyenDuyPhuc/DistressAI", device="cuda")
25
+ output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0]
26
+ print(output["generated_text"])
27
+ ```
28
+
29
+ ## Training procedure
30
+
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/duyphuclenguyen-university-of-south-florida/huggingface/runs/dwg97uxc)
32
+
33
+
34
+ This model was trained with SFT.
35
+
36
+ ### Framework versions
37
+
38
+ - TRL: 0.15.2
39
+ - Transformers: 4.50.3
40
+ - Pytorch: 2.6.0
41
+ - Datasets: 3.3.2
42
+ - Tokenizers: 0.21.0
43
+
44
+ ## Citations
45
+
46
+
47
+
48
+ Cite TRL as:
49
+
50
+ ```bibtex
51
+ @misc{vonwerra2022trl,
52
+ title = {{TRL: Transformer Reinforcement Learning}},
53
+ author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
54
+ year = 2020,
55
+ journal = {GitHub repository},
56
+ publisher = {GitHub},
57
+ howpublished = {\url{https://github.com/huggingface/trl}}
58
+ }
59
+ ```
added_tokens.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
+ }
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "Qwen2ForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 151644,
7
+ "eos_token_id": 151645,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 896,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 4864,
12
+ "max_position_embeddings": 32768,
13
+ "max_window_layers": 21,
14
+ "model_type": "qwen2",
15
+ "num_attention_heads": 14,
16
+ "num_hidden_layers": 24,
17
+ "num_key_value_heads": 2,
18
+ "pad_token_id": 151645,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 1000000.0,
22
+ "sliding_window": 32768,
23
+ "tie_word_embeddings": true,
24
+ "torch_dtype": "float32",
25
+ "transformers_version": "4.50.3",
26
+ "use_cache": true,
27
+ "use_sliding_window": false,
28
+ "vocab_size": 151665
29
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151644,
3
+ "do_sample": true,
4
+ "eos_token_id": 151645,
5
+ "pad_token_id": 151645,
6
+ "repetition_penalty": 1.1,
7
+ "temperature": 0.7,
8
+ "top_k": 20,
9
+ "top_p": 0.8,
10
+ "transformers_version": "4.50.3"
11
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc6fa80b3498c949b8da388f7e6401fae464c851f6adcd1f8172c8418d2a1305
3
+ size 1975192208
runs/Apr01_14-39-31_caicc/events.out.tfevents.1743532773.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3d50be1fd564f6e74974601dcc311fb3705d5d9a4e20a4e389acefbfb2972112
3
+ size 4547
runs/Apr01_14-44-03_caicc/events.out.tfevents.1743533070.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cd1330d07b7e426d29c804b5c384d03f4187941886f6dd7f8c39b59d43806fb2
3
+ size 4547
runs/Apr01_14-48-52_caicc/events.out.tfevents.1743533353.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3915c2c44e0546e592c78a72322900453ecc33cf8fecc0d4e4e744df1a3e3d41
3
+ size 33175
runs/Apr01_15-00-03_caicc/events.out.tfevents.1743534245.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:055b6679317a19e3316d14747665fd910f9a64ceecffa748bb355cb1499aae3d
3
+ size 33175
runs/Apr01_15-18-41_caicc/events.out.tfevents.1743535315.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f448c3246d44100edeb245d81e2defbe405a5d668badecd56cd8f75408cab77
3
+ size 4547
runs/Apr01_15-23-03_caicc/events.out.tfevents.1743535383.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a45ef359ec407f19f2265a705dfc8038b8f00feca4752d84dff24b9ea992bff
3
+ size 33175
runs/Apr01_15-27-21_caicc/events.out.tfevents.1743535836.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8c120d60e035dea4aa9366bb8e39226ef0f086353f7df9a1cd189c988bb08d3
3
+ size 4548
runs/Apr01_15-32-07_caicc/events.out.tfevents.1743535928.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df3026f96fe05126e19406978369e6ce389cb116863b3e9308bf951f4997b8ef
3
+ size 143175
runs/Apr01_15-58-49_caicc/events.out.tfevents.1743537529.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7466ac44498304b832e5955b2a3cf514369f8496e13b7ee4e3d5f4984bc9afed
3
+ size 7785
runs/Apr01_15-59-16_caicc/events.out.tfevents.1743537556.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3224bce17bc4eda79b38d8f2cbc2ceaf0b1d9d42c73aede5ce6308fdb93506a0
3
+ size 9695
runs/Apr01_16-07-22_caicc/events.out.tfevents.1743538089.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:991e7ba39e8d73e28088e96f4e76ba253df2df21a33e168bce92087c2170257f
3
+ size 4557
runs/Apr01_16-08-37_caicc/events.out.tfevents.1743538152.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:797c540d1579b7ce764fe4a5f672465dc1d9bc07f436b186df325334872c8fbe
3
+ size 5391
runs/Apr01_16-09-43_caicc/events.out.tfevents.1743538222.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e30498a7edc36bdd137e38718b49d20ef49275b305018b9b1600942b33d5930
3
+ size 5391
runs/Apr01_16-14-20_caicc/events.out.tfevents.1743538502.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:820db177aeca2baff9e2357ef05e744efc83df3b16abfce8c55d837b3f59b8ae
3
+ size 7817
runs/Apr01_16-16-21_caicc/events.out.tfevents.1743538584.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b61c3981bec486feda3c695f8c0eacc8c0914266fd966bd41cf287593c6f13f
3
+ size 33181
runs/Mar31_00-09-41_caicc/events.out.tfevents.1743394189.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed34140f62139114aa99a95c5ab90a89d208f181fb3733c354b12019bb5e6c89
3
+ size 4534
runs/Mar31_00-14-18_caicc/events.out.tfevents.1743394460.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4697dec1d69b2ace049776a55fdcd353f6424553c4638849958f035a4a1cf72
3
+ size 4534
runs/Mar31_00-14-51_caicc/events.out.tfevents.1743394492.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1f6beff40ec6dd74ae94eb96e22bcef58813f1a93c8bcf11dc749ac1cbb721f
3
+ size 5569
runs/Mar31_00-19-52_caicc/events.out.tfevents.1743394794.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c09a713a751a930fdf5a4050d13dba369ffe4bf580ccec831884f8065a8b2e2
3
+ size 5565
runs/Mar31_00-25-08_caicc/events.out.tfevents.1743395109.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:948410e26a748a8ea4cb10aaa4530191d9fea61848b8cd881fd24964d8e2627b
3
+ size 5653
runs/Mar31_00-30-22_caicc/events.out.tfevents.1743395426.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc506f2ca01457de3ea97da182ba5e99ef865e56394d04f1d3c9a32d66b8acfc
3
+ size 5653
runs/Mar31_00-32-20_caicc/events.out.tfevents.1743395544.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4928ad28600078543c8082b6eb857e8bf7814d3ac4940846281210670eb3cd9
3
+ size 5923
runs/Mar31_00-42-27_caicc/events.out.tfevents.1743396680.caicc ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83e9a4638f88e10ce86a1bee5de497ed3d50ff58576bf61ceddac2233e82893a
3
+ size 5536
special_tokens_map.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ {
4
+ "content": "<|im_start|>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false
9
+ },
10
+ {
11
+ "content": "<|im_end|>",
12
+ "lstrip": false,
13
+ "normalized": false,
14
+ "rstrip": false,
15
+ "single_word": false
16
+ }
17
+ ],
18
+ "bos_token": "<|im_start|>",
19
+ "eos_token": "<|im_end|>",
20
+ "pad_token": "<|im_end|>"
21
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json ADDED
@@ -0,0 +1,197 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
+ "added_tokens_decoder": {
5
+ "151643": {
6
+ "content": "<|endoftext|>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "151644": {
14
+ "content": "<|im_start|>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "151645": {
22
+ "content": "<|im_end|>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "151646": {
30
+ "content": "<|object_ref_start|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "151647": {
38
+ "content": "<|object_ref_end|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "151648": {
46
+ "content": "<|box_start|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "151649": {
54
+ "content": "<|box_end|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "151650": {
62
+ "content": "<|quad_start|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "151651": {
70
+ "content": "<|quad_end|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "151652": {
78
+ "content": "<|vision_start|>",
79
+ "lstrip": false,
80
+ "normalized": false,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ },
85
+ "151653": {
86
+ "content": "<|vision_end|>",
87
+ "lstrip": false,
88
+ "normalized": false,
89
+ "rstrip": false,
90
+ "single_word": false,
91
+ "special": true
92
+ },
93
+ "151654": {
94
+ "content": "<|vision_pad|>",
95
+ "lstrip": false,
96
+ "normalized": false,
97
+ "rstrip": false,
98
+ "single_word": false,
99
+ "special": true
100
+ },
101
+ "151655": {
102
+ "content": "<|image_pad|>",
103
+ "lstrip": false,
104
+ "normalized": false,
105
+ "rstrip": false,
106
+ "single_word": false,
107
+ "special": true
108
+ },
109
+ "151656": {
110
+ "content": "<|video_pad|>",
111
+ "lstrip": false,
112
+ "normalized": false,
113
+ "rstrip": false,
114
+ "single_word": false,
115
+ "special": true
116
+ },
117
+ "151657": {
118
+ "content": "<tool_call>",
119
+ "lstrip": false,
120
+ "normalized": false,
121
+ "rstrip": false,
122
+ "single_word": false,
123
+ "special": false
124
+ },
125
+ "151658": {
126
+ "content": "</tool_call>",
127
+ "lstrip": false,
128
+ "normalized": false,
129
+ "rstrip": false,
130
+ "single_word": false,
131
+ "special": false
132
+ },
133
+ "151659": {
134
+ "content": "<|fim_prefix|>",
135
+ "lstrip": false,
136
+ "normalized": false,
137
+ "rstrip": false,
138
+ "single_word": false,
139
+ "special": false
140
+ },
141
+ "151660": {
142
+ "content": "<|fim_middle|>",
143
+ "lstrip": false,
144
+ "normalized": false,
145
+ "rstrip": false,
146
+ "single_word": false,
147
+ "special": false
148
+ },
149
+ "151661": {
150
+ "content": "<|fim_suffix|>",
151
+ "lstrip": false,
152
+ "normalized": false,
153
+ "rstrip": false,
154
+ "single_word": false,
155
+ "special": false
156
+ },
157
+ "151662": {
158
+ "content": "<|fim_pad|>",
159
+ "lstrip": false,
160
+ "normalized": false,
161
+ "rstrip": false,
162
+ "single_word": false,
163
+ "special": false
164
+ },
165
+ "151663": {
166
+ "content": "<|repo_name|>",
167
+ "lstrip": false,
168
+ "normalized": false,
169
+ "rstrip": false,
170
+ "single_word": false,
171
+ "special": false
172
+ },
173
+ "151664": {
174
+ "content": "<|file_sep|>",
175
+ "lstrip": false,
176
+ "normalized": false,
177
+ "rstrip": false,
178
+ "single_word": false,
179
+ "special": false
180
+ }
181
+ },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>"
185
+ ],
186
+ "bos_token": "<|im_start|>",
187
+ "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
188
+ "clean_up_tokenization_spaces": false,
189
+ "eos_token": "<|im_end|>",
190
+ "errors": "replace",
191
+ "extra_special_tokens": {},
192
+ "model_max_length": 131072,
193
+ "pad_token": "<|im_end|>",
194
+ "split_special_tokens": false,
195
+ "tokenizer_class": "Qwen2Tokenizer",
196
+ "unk_token": null
197
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b4c027ac7217e8400a148d6df73f0df0c9b9148e95c742d940ecb19b730c2771
3
+ size 5624
vocab.json ADDED
The diff for this file is too large to render. See raw diff