Text Generation
Transformers
Safetensors
mistral
conversational
text-generation-inference
zerofata commited on
Commit
4fd6637
·
verified ·
1 Parent(s): de59b9d

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +495 -0
README.md ADDED
@@ -0,0 +1,495 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: apache-2.0
4
+ datasets:
5
+ - zerofata/Roleplay-Anime-Characters
6
+ - zerofata/Instruct-Anime-CreativeWriting
7
+ - zerofata/Summaries-Anime-FandomPages
8
+ base_model:
9
+ - mistralai/Mistral-Small-3.2-24B-Instruct-2506
10
+ ---
11
+ <!DOCTYPE html>
12
+ <style>
13
+ body {
14
+ font-family: sans-serif;
15
+ color: #e8f4f8;
16
+ line-height: 1.6;
17
+ margin: 0;
18
+ padding: 0;
19
+ background-color: #0a1628;
20
+ }
21
+
22
+ .lemonade-text {
23
+ color: #4fc3f7;
24
+ position: relative;
25
+ z-index: 2;
26
+ margin-left: 0.2em;
27
+ text-shadow: 0 0 15px #4fc3f7;
28
+ }
29
+
30
+ /* Section styling */
31
+ .section-container {
32
+ background-color: rgba(10, 22, 40, 0.8);
33
+ margin-bottom: 30px;
34
+ position: relative;
35
+ overflow: hidden;
36
+ border-bottom: 1px solid #4fc3f7;
37
+ box-shadow: 0 4px 15px rgba(79, 195, 247, 0.1);
38
+ }
39
+
40
+ .section-header {
41
+ display: flex;
42
+ align-items: center;
43
+ background-color: rgba(79, 195, 247, 0.06);
44
+ padding: 10px 20px;
45
+ }
46
+
47
+ .section-indicator {
48
+ width: 8px;
49
+ height: 20px;
50
+ background-color: #4fc3f7;
51
+ margin-right: 15px;
52
+ box-shadow: 0 0 8px rgba(79, 195, 247, 0.4);
53
+ }
54
+
55
+ .section-title {
56
+ font-family: 'Georgia', 'Times New Roman', serif;
57
+ color: #e8f4f8;
58
+ font-size: 1.4rem;
59
+ margin: 0;
60
+ letter-spacing: 1px;
61
+ font-weight: 400;
62
+ text-transform: capitalize;
63
+ }
64
+
65
+ .section-content {
66
+ padding: 20px;
67
+ font-family: sans-serif;
68
+ color: #e8f4f8;
69
+ line-height: 1.6;
70
+ }
71
+
72
+ /* Title styling */
73
+ .title-container {
74
+ background-color: #051017;
75
+ position: relative;
76
+ overflow: hidden;
77
+ margin-bottom: 40px;
78
+ border-left: 3px solid #4fc3f7;
79
+ box-shadow: 0 6px 20px rgba(79, 195, 247, 0.15);
80
+ }
81
+
82
+ .title-wrapper {
83
+ position: relative;
84
+ z-index: 2;
85
+ padding: 25px 20px 30px 30px;
86
+ font-family: 'Georgia', 'Times New Roman', serif;
87
+ }
88
+
89
+ .title-main {
90
+ color: #e8f4f8;
91
+ font-size: 2.5rem;
92
+ font-weight: 700;
93
+ margin: 0;
94
+ letter-spacing: 2px;
95
+ display: inline-block;
96
+ position: relative;
97
+ text-transform: uppercase;
98
+ }
99
+
100
+ .title-prefix {
101
+ position: relative;
102
+ z-index: 2;
103
+ }
104
+
105
+ .title-subtitle {
106
+ padding-left: 15px;
107
+ margin-top: 5px;
108
+ margin-left: 5px;
109
+ }
110
+
111
+ .subtitle-text {
112
+ color: #b39ddb;
113
+ font-size: 1.2rem;
114
+ font-family: 'Georgia', 'Times New Roman', serif;
115
+ font-weight: 300;
116
+ letter-spacing: 3px;
117
+ text-transform: uppercase;
118
+ display: inline-block;
119
+ }
120
+
121
+ .glitchy-overlay {
122
+ position: absolute;
123
+ top: 0;
124
+ left: 0;
125
+ width: 100%;
126
+ height: 100%;
127
+ background-image: repeating-linear-gradient(0deg, rgba(0,0,0,0) 0, rgba(79, 195, 247, 0.08) 1px, rgba(0,0,0,0) 2px);
128
+ z-index: 1;
129
+ }
130
+
131
+ /* Data box styling */
132
+ .data-box {
133
+ background-color: rgba(5, 16, 23, 0.6);
134
+ padding: 15px;
135
+ border-left: 2px solid #4fc3f7;
136
+ margin-bottom: 20px;
137
+ box-shadow: 0 2px 10px rgba(79, 195, 247, 0.1);
138
+ }
139
+
140
+ .data-row {
141
+ display: flex;
142
+ margin-bottom: 8px;
143
+ }
144
+
145
+ .data-arrow {
146
+ color: #4fc3f7;
147
+ width: 20px;
148
+ display: inline-block;
149
+ }
150
+
151
+ .data-label {
152
+ color: #b39ddb;
153
+ width: 80px;
154
+ display: inline-block;
155
+ }
156
+
157
+ /* Subheading styling */
158
+ .subheading {
159
+ color: #b39ddb;
160
+ font-size: 1.1rem;
161
+ margin-top: 20px;
162
+ margin-bottom: 15px;
163
+ font-weight: 400;
164
+ border-bottom: 1px dashed rgba(179, 157, 219, 0.4);
165
+ display: inline-block;
166
+ text-transform: uppercase;
167
+ letter-spacing: 1px;
168
+ font-family: 'Georgia', 'Times New Roman', serif;
169
+ }
170
+
171
+ /* Links */
172
+ a {
173
+ color: #b39ddb;
174
+ text-decoration: none;
175
+ }
176
+
177
+ a:hover {
178
+ text-decoration: underline;
179
+ color: #f8bbd9;
180
+ }
181
+
182
+ /* Container */
183
+ .container {
184
+ max-width: 1200px;
185
+ margin: 20px auto;
186
+ padding: 40px 20px;
187
+ background-color: #051017;
188
+ background-image:
189
+ radial-gradient(circle at 20% 80%, rgba(79, 195, 247, 0.03) 0%, transparent 50%),
190
+ radial-gradient(circle at 80% 20%, rgba(179, 157, 219, 0.03) 0%, transparent 50%),
191
+ radial-gradient(circle at 40% 40%, rgba(248, 187, 217, 0.02) 0%, transparent 50%);
192
+ min-height: calc(100vh - 40px);
193
+ border: 1px solid #4fc3f7;
194
+ border-radius: 8px;
195
+ box-shadow: 0 8px 32px rgba(79, 195, 247, 0.15);
196
+ }
197
+
198
+ /* Dropdown styling */
199
+ .dropdown-container {
200
+ margin-top: 20px;
201
+ }
202
+
203
+ .dropdown-summary {
204
+ cursor: pointer;
205
+ padding: 10px 0;
206
+ border-bottom: 1px dashed rgba(179, 157, 219, 0.4);
207
+ color: #b39ddb;
208
+ font-size: 1.1rem;
209
+ font-weight: 400;
210
+ text-transform: uppercase;
211
+ letter-spacing: 1px;
212
+ font-family: 'Georgia', 'Times New Roman', serif;
213
+ list-style: none;
214
+ display: flex;
215
+ align-items: center;
216
+ }
217
+
218
+ .dropdown-summary::-webkit-details-marker {
219
+ display: none;
220
+ }
221
+
222
+ .dropdown-arrow {
223
+ color: #4fc3f7;
224
+ margin-right: 10px;
225
+ transition: transform 0.3s ease;
226
+ }
227
+
228
+ .dropdown-container[open] .dropdown-arrow {
229
+ transform: rotate(90deg);
230
+ }
231
+
232
+ .dropdown-content {
233
+ margin-top: 15px;
234
+ padding: 15px;
235
+ background-color: rgba(5, 16, 23, 0.6);
236
+ border-left: 2px solid #4fc3f7;
237
+ box-shadow: 0 2px 10px rgba(79, 195, 247, 0.1);
238
+ }
239
+
240
+ .config-title {
241
+ color: #4fc3f7;
242
+ font-size: 1rem;
243
+ margin-bottom: 10px;
244
+ font-family: 'Georgia', 'Times New Roman', serif;
245
+ text-transform: uppercase;
246
+ letter-spacing: 1px;
247
+ }
248
+
249
+ </style>
250
+ <html lang="en">
251
+ <head>
252
+ <meta charset="UTF-8">
253
+ <meta name="viewport" content="width=device-width, initial-scale=1.0">
254
+ <title>Apocalypse Lemonade</title>
255
+ <link href="https://fonts.googleapis.com/css2?family=Orbitron:wght@400;500;600;700&family=JetBrains+Mono:wght@100;300;400;700&display=swap" rel="stylesheet">
256
+ </head>
257
+ <body>
258
+
259
+ <div class="container">
260
+ <div class="title-container">
261
+ <!-- Glitchy overlay -->
262
+ <div class="glitchy-overlay"></div>
263
+ <!-- Main title -->
264
+ <div class="title-wrapper">
265
+ <h1 class="title-main">
266
+ <span class="title-prefix">APOCALYPSE</span>
267
+ <span class="lemonade-text">LEMONADE</span> <!-- Static text with glow -->
268
+ </h1>
269
+ <div class="title-subtitle">
270
+ <span class="subtitle-text">Mistral Small 3.2 24B</span>
271
+ </div>
272
+ </div>
273
+ </div>
274
+
275
+
276
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65b19c6c638328850e12d38c/0Ka08CdFUIJtYctBeBATo.png)
277
+
278
+
279
+ <div class="section-container">
280
+ <div class="section-header">
281
+ <div class="section-indicator"></div>
282
+ <h2 class="section-title">Overview</h2>
283
+ </div>
284
+ <div class="section-content">
285
+ <p>Experimental release.</p>
286
+ <p>First Mistral model and first small model I've tuned in general.</p>
287
+ <p>This is an uncensored creative model intended to excel at character driven RP / ERP.</p>
288
+ <p>This model is designed to provide longer, narrative heavy responses where characters are portrayed accurately and proactively.</p>
289
+ <p>Mistral Small 3.2 surprised me. Normally I find small models annoying inconsistent, but this has been relatively creative and stable, all things considered. Possibly due to this being a QLora finetune, the model has similar issues to default MS3.2 (repetition and system prompt adherance), but I think the outputs are nice.</p>
290
+ </div>
291
+ </div>
292
+
293
+ <div class="section-container">
294
+ <div class="section-header">
295
+ <div class="section-indicator"></div>
296
+ <h2 class="section-title">SillyTavern Settings</h2>
297
+ </div>
298
+ <div class="section-content">
299
+ <h3 class="subheading">Recommended Roleplay Format</h3>
300
+ <div class="data-box">
301
+ <div class="data-row">
302
+ <span class="data-arrow">></span>
303
+ <span class="data-label">Actions:</span>
304
+ <span>In plaintext</span>
305
+ </div>
306
+ <div class="data-row">
307
+ <span class="data-arrow">></span>
308
+ <span class="data-label">Dialogue:</span>
309
+ <span>"In quotes"</span>
310
+ </div>
311
+ <div class="data-row">
312
+ <span class="data-arrow">></span>
313
+ <span class="data-label">Thoughts:</span>
314
+ <span>*In asterisks*</span>
315
+ </div>
316
+ </div>
317
+ <h3 class="subheading">Recommended Samplers</h3>
318
+ <div class="data-box">
319
+ <div class="data-row">
320
+ <span class="data-arrow">></span>
321
+ <span class="data-label">Temp:</span>
322
+ <span>0.8</span>
323
+ </div>
324
+ <div class="data-row">
325
+ <span class="data-arrow">></span>
326
+ <span class="data-label">MinP:</span>
327
+ <span>0.04 - 0.05</span>
328
+ </div>
329
+ <div class="data-row">
330
+ <span class="data-arrow">></span>
331
+ <span class="data-label">TopP:</span>
332
+ <span>0.95 - 1.0</span>
333
+ </div>
334
+ <div class="data-row">
335
+ <span class="data-arrow">></span>
336
+ <span class="data-label">Dry:</span>
337
+ <span>0.8, 1.75, 4</span>
338
+ </div>
339
+ </div>
340
+ <h3 class="subheading">Instruct</h3>
341
+ <div class="data-box">
342
+ <p style="margin: 0;">Mistral v11 (or v7) Tekken</p>
343
+ </div>
344
+ </div>
345
+ </div>
346
+
347
+ <div class="section-container">
348
+ <div class="section-header">
349
+ <div class="section-indicator"></div>
350
+ <h2 class="section-title">Quantizations</h2>
351
+ </div>
352
+ <div class="section-content">
353
+ <div style="margin-bottom: 20px;">
354
+ <h3 class="subheading">GGUF</h3>
355
+ <div class="data-box">
356
+ <div class="data-row">
357
+ <span style="color: #4fc3f7; display: inline-block; margin-right: 10px;">> </span><a href="">iMatrix (TBC)</a>
358
+ </div>
359
+ </div>
360
+ </div>
361
+ <div>
362
+ <h3 class="subheading">EXL3</h3>
363
+ <div class="data-box">
364
+ <div class="data-row">
365
+ <span style="color: #4fc3f7; display: inline-block; margin-right: 10px;">> </span><a href="">4bpw</a>
366
+ </div>
367
+ <div class="data-row">
368
+ <span style="color: #4fc3f7; display: inline-block; margin-right: 10px;">> </span><a href="">4.5bpw</a>
369
+ </div>
370
+ <div class="data-row">
371
+ <span style="color: #4fc3f7; display: inline-block; margin-right: 10px;">> </span><a href="">4.65bpw</a>
372
+ </div>
373
+ <div class="data-row">
374
+ <span style="color: #4fc3f7; display: inline-block; margin-right: 10px;">> </span><a href="">6bpw</a>
375
+ </div>
376
+ </div>
377
+ </div>
378
+ </div>
379
+ </div>
380
+
381
+ <div class="section-container">
382
+ <div class="section-header">
383
+ <div class="section-indicator"></div>
384
+ <h2 class="section-title">Training Process</h2>
385
+ </div>
386
+ <div class="section-content">
387
+ <p>Training process: Pretrain > SFT > DPO > DPO 2</p>
388
+ <p>Did a small pretrain on some light novels and Frieren wiki data as a test. Hasn't seemed to hurt the model and model has shown some small improvements in the lore of series that were included.</p>
389
+ <p>The model then went through the standard SFT using a dataset of approx 3.6 million tokens, 700 RP conversations, 1000 creative writing / instruct samples and about 100 summaries. The bulk of this data has been made public.</p>
390
+ <p>Finally DPO was used to make the model a little more consistent. The first stage of DPO focused on instruction following and the second tried to burn out some Mistral-isms.</p>
391
+ <div class="dropdown-container">
392
+ <details>
393
+ <summary class="dropdown-summary">
394
+ <span class="dropdown-arrow">></span>
395
+ Axolotl configs
396
+ </summary>
397
+ <div class="dropdown-content">
398
+ <p>Not optimized for cost / performance efficiency, YMMV.</p>
399
+ <div class="config-title">SFT 1*H100</div>
400
+ <pre style="background-color: rgba(5, 16, 23, 0.4); border-radius: 6px; padding: 15px; overflow-x: auto; border: 1px solid #4fc3f7; position: relative;"><code style="font-family: 'Fira Code', 'Courier New', Courier, monospace; font-size: 0.9em; color: #e8f4f8;">&#35; ====================
401
+ &#35; MODEL CONFIGURATION
402
+ &#35; ====================
403
+ base_model: ./MS3-2-Pretrain/merged
404
+ model_type: AutoModelForCausalLM
405
+ tokenizer_type: AutoTokenizer
406
+ <br>
407
+ &#35; ====================
408
+ &#35; DATASET CONFIGURATION
409
+ &#35; ====================
410
+ datasets:
411
+ - path: ./dataset.jsonl
412
+ type: chat_template
413
+ split: train
414
+ chat_template_strategy: tokenizer
415
+ field_messages: messages
416
+ message_property_mappings:
417
+ role: role
418
+ content: content
419
+ roles:
420
+ user: ["user"]
421
+ assistant: ["assistant"]
422
+ system: ["system"]
423
+
424
+ dataset_prepared_path:
425
+ train_on_inputs: false &#35; Only train on assistant responses
426
+
427
+ &#35; ====================
428
+ &#35; QLORA CONFIGURATION
429
+ &#35; ====================
430
+ adapter: qlora
431
+ load_in_4bit: true
432
+ lora_r: 128
433
+ lora_alpha: 128
434
+ lora_dropout: 0.1
435
+ lora_target_linear: true
436
+ &#35; lora_modules_to_save: &#35; Uncomment only if you added NEW tokens
437
+
438
+ &#35; ====================
439
+ &#35; TRAINING PARAMETERS
440
+ &#35; ====================
441
+ num_epochs: 3
442
+ micro_batch_size: 4
443
+ gradient_accumulation_steps: 2
444
+ learning_rate: 1e-5
445
+ optimizer: paged_adamw_8bit
446
+ lr_scheduler: rex
447
+ warmup_ratio: 0.05
448
+ weight_decay: 0.01
449
+ max_grad_norm: 1.0
450
+
451
+ &#35; ====================
452
+ &#35; SEQUENCE &amp; PACKING
453
+ &#35; ====================
454
+ sequence_len: 8192
455
+ sample_packing: true
456
+ eval_sample_packing: false
457
+ pad_to_sequence_len: true
458
+
459
+ &#35; ====================
460
+ &#35; HARDWARE OPTIMIZATIONS
461
+ &#35; ====================
462
+ bf16: auto
463
+ flash_attention: true
464
+ gradient_checkpointing: true
465
+
466
+ &#35; ====================
467
+ &#35; EVALUATION &amp; CHECKPOINTING
468
+ &#35; ====================
469
+ save_strategy: steps
470
+ save_steps: 5
471
+ save_total_limit: 5 &#35; Keep best + last few checkpoints
472
+ load_best_model_at_end: true
473
+ greater_is_better: false
474
+
475
+ &#35; ====================
476
+ &#35; LOGGING &amp; OUTPUT
477
+ &#35; ====================
478
+ output_dir: ./MS3-2-SFT-2
479
+ logging_steps: 2
480
+ save_safetensors: true
481
+
482
+ &#35; ====================
483
+ &#35; WANDB TRACKING
484
+ &#35; ====================
485
+ wandb_project: MS3-2-SFT
486
+ &#35; wandb_entity: your_entity
487
+ wandb_name: run_name</code></pre>
488
+ </div>
489
+ </details>
490
+ </div>
491
+ </div>
492
+ </div>
493
+ </div>
494
+ </body>
495
+ </html>