Delta-Vector commited on
Commit
d03afad
·
verified ·
1 Parent(s): 8b9ff7b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +90 -52
README.md CHANGED
@@ -1,20 +1,92 @@
1
  ---
2
- library_name: transformers
3
- base_model: Dans-DiscountModels/Mistral-NeMo-Minitron-8B-Base-ChatML
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4
  tags:
5
- - generated_from_trainer
6
- model-index:
7
- - name: workspace/data/8b-nemo-fft-out
8
- results: []
 
9
  ---
10
 
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
 
14
- [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  <details><summary>See axolotl config</summary>
16
 
17
- axolotl version: `0.4.1`
18
  ```yaml
19
  base_model: Dans-DiscountModels/Mistral-NeMo-Minitron-8B-Base-ChatML
20
  model_type: AutoModelForCausalLM
@@ -33,7 +105,7 @@ load_in_4bit: false
33
  strict: false
34
 
35
  datasets:
36
- - path: anthracite-core/c2_logs_16k_llama_v1.1
37
  type: sharegpt
38
  conversation: chatml
39
  - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
@@ -116,52 +188,18 @@ fsdp:
116
  fsdp_config:
117
  special_tokens:
118
  pad_token: <pad>
119
- ```
120
-
121
- </details><br>
122
-
123
- # workspace/data/8b-nemo-fft-out
124
-
125
- This model is a fine-tuned version of [Dans-DiscountModels/Mistral-NeMo-Minitron-8B-Base-ChatML](https://huggingface.co/Dans-DiscountModels/Mistral-NeMo-Minitron-8B-Base-ChatML) on the None dataset.
126
-
127
- ## Model description
128
 
129
- More information needed
130
 
131
- ## Intended uses & limitations
132
-
133
- More information needed
134
-
135
- ## Training and evaluation data
136
-
137
- More information needed
138
-
139
- ## Training procedure
140
-
141
- ### Training hyperparameters
142
 
143
- The following hyperparameters were used during training:
144
- - learning_rate: 1e-05
145
- - train_batch_size: 2
146
- - eval_batch_size: 2
147
- - seed: 42
148
- - distributed_type: multi-GPU
149
- - num_devices: 10
150
- - gradient_accumulation_steps: 2
151
- - total_train_batch_size: 40
152
- - total_eval_batch_size: 20
153
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
154
- - lr_scheduler_type: cosine
155
- - lr_scheduler_warmup_steps: 10
156
- - num_epochs: 4
157
 
158
- ### Training results
159
 
 
160
 
161
 
162
- ### Framework versions
 
163
 
164
- - Transformers 4.45.0.dev0
165
- - Pytorch 2.4.0+cu121
166
- - Datasets 2.21.0
167
- - Tokenizers 0.19.1
 
1
  ---
2
+ License: agpl-3.0
3
+ Language:
4
+ - En
5
+ Pipeline_tag: text-generation
6
+ Base_model: nvidia/Mistral-NeMo-Minitron-8B-Base
7
+ Tags:
8
+ - Chat
9
+ license: agpl-3.0
10
+ datasets:
11
+ - anthracite-org/kalo-opus-instruct-22k-no-refusal
12
+ - Epiculous/SynthRP-Gens-v1.1-Filtered-n-Cleaned
13
+ - lodrick-the-lafted/kalo-opus-instruct-3k-filtered
14
+ - anthracite-org/nopm_claude_writing_fixed
15
+ - Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned
16
+ - anthracite-org/kalo_opus_misc_240827
17
+ - anthracite-org/kalo_misc_part2
18
  tags:
19
+ - chat
20
+ language:
21
+ - en
22
+ base_model:
23
+ - nvidia/Mistral-NeMo-Minitron-8B-Base
24
  ---
25
 
 
 
26
 
27
+ This is the fully cooked, 4 epoch version of [Tor-8B](), this is an experimental version, despite being trained for 4 epochs, the model feels fresh and new and is not overfit, This model aims to have generally good prose and writing while not falling into claude-isms, it follows the *actions* "dialogue" format heavily.
28
+
29
+
30
+ # Quants
31
+
32
+ GGUF: https://huggingface.co/Delta-Vector/Darkens-8B-GGUF
33
+
34
+ EXL2: https://huggingface.co/Delta-Vector/Darkens-8B-EXL2
35
+
36
+
37
+ ## Prompting
38
+ Model has been Instruct tuned with the ChatML formatting. A typical input would look like this:
39
+
40
+ ```py
41
+ """<|im_start|>system
42
+ system prompt<|im_end|>
43
+ <|im_start|>user
44
+ Hi there!<|im_end|>
45
+ <|im_start|>assistant
46
+ Nice to meet you!<|im_end|>
47
+ <|im_start|>user
48
+ Can I ask a question?<|im_end|>
49
+ <|im_start|>assistant
50
+ """
51
+ ```
52
+ ## System Prompting
53
+
54
+ I would highly recommend using Sao10k's Euryale System prompt, But the "Roleplay Simple" system prompt provided within SillyTavern will work aswell.
55
+
56
+ ```
57
+ Currently, your role is {{char}}, described in detail below. As {{char}}, continue the narrative exchange with {{user}}.
58
+
59
+ <Guidelines>
60
+ • Maintain the character persona but allow it to evolve with the story.
61
+ • Be creative and proactive. Drive the story forward, introducing plotlines and events when relevant.
62
+ • All types of outputs are encouraged; respond accordingly to the narrative.
63
+ • Include dialogues, actions, and thoughts in each response.
64
+ • Utilize all five senses to describe scenarios within {{char}}'s dialogue.
65
+ • Use emotional symbols such as "!" and "~" in appropriate contexts.
66
+ • Incorporate onomatopoeia when suitable.
67
+ • Allow time for {{user}} to respond with their own input, respecting their agency.
68
+ • Act as secondary characters and NPCs as needed, and remove them when appropriate.
69
+ • When prompted for an Out of Character [OOC:] reply, answer neutrally and in plaintext, not as {{char}}.
70
+ </Guidelines>
71
+
72
+ <Forbidden>
73
+ • Using excessive literary embellishments and purple prose unless dictated by {{char}}'s persona.
74
+ • Writing for, speaking, thinking, acting, or replying as {{user}} in your response.
75
+ • Repetitive and monotonous outputs.
76
+ • Positivity bias in your replies.
77
+ • Being overly extreme or NSFW when the narrative context is inappropriate.
78
+ </Forbidden>
79
+
80
+ Follow the instructions in <Guidelines></Guidelines>, avoiding the items listed in <Forbidden></Forbidden>.
81
+
82
+ ```
83
+
84
+
85
+ ## Axolotl config
86
+
87
  <details><summary>See axolotl config</summary>
88
 
89
+ Axolotl version: `0.4.1`
90
  ```yaml
91
  base_model: Dans-DiscountModels/Mistral-NeMo-Minitron-8B-Base-ChatML
92
  model_type: AutoModelForCausalLM
 
105
  strict: false
106
 
107
  datasets:
108
+ - path: PRIVATE CLAUDE LOG FILTER
109
  type: sharegpt
110
  conversation: chatml
111
  - path: anthracite-org/kalo-opus-instruct-22k-no-refusal
 
188
  fsdp_config:
189
  special_tokens:
190
  pad_token: <pad>
 
 
 
 
 
 
 
 
 
191
 
 
192
 
193
+ ```
 
 
 
 
 
 
 
 
 
 
194
 
195
+ </details><br>
 
 
 
 
 
 
 
 
 
 
 
 
 
196
 
197
+ ## Credits
198
 
199
+ Thank you to [Lucy Knada](https://huggingface.co/lucyknada), [Kalomaze](https://huggingface.co/kalomaze), [Kubernetes Bad](https://huggingface.co/kubernetes-bad) and the rest of [Anthracite](https://huggingface.co/anthracite-org) (But not Alpin.)
200
 
201
 
202
+ ## Training
203
+ The training was done for 4 epochs. I used 10 x [A40s](https://www.nvidia.com/en-us/data-center/a40/) GPUs graciously provided by [Kalomaze](https://huggingface.co/kalomaze) for the full-parameter fine-tuning of the model.
204
 
205
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)