AnonSOB commited on
Commit
0714121
·
verified ·
1 Parent(s): 671fda8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -1
README.md CHANGED
@@ -9,16 +9,31 @@ base_model:
9
  tags:
10
  - meme
11
  - crysis
 
12
  ---
13
 
14
  This model is a fine-tuned version of [google/gemma-3-1b-it-qat-q4_0-unquantized](https://huggingface.co/google/gemma-3-1b-it-qat-q4_0-unquantized) on [AnonSOB/Crysis](https://huggingface.co/datasets/AnonSOB/Crysis).
15
 
 
 
16
  The following hyperparameters were used during training:
17
  - learning_rate: 1e-5
18
  - train_batch_size: 5
19
  - seed: 0
20
  - num_epochs: 6000
21
 
 
 
22
  Trained with mlx-lm.
 
 
 
23
  Thank you Alex Ziskind for original TinyLlama Crysis ❤️
24
- This model is unquantized. I dont know why I trained it on Gemma 3 QAT because it acts weird when quantized.
 
 
 
 
 
 
 
 
9
  tags:
10
  - meme
11
  - crysis
12
+ library_name: mlx
13
  ---
14
 
15
  This model is a fine-tuned version of [google/gemma-3-1b-it-qat-q4_0-unquantized](https://huggingface.co/google/gemma-3-1b-it-qat-q4_0-unquantized) on [AnonSOB/Crysis](https://huggingface.co/datasets/AnonSOB/Crysis).
16
 
17
+ ### Training hyperparameters
18
+
19
  The following hyperparameters were used during training:
20
  - learning_rate: 1e-5
21
  - train_batch_size: 5
22
  - seed: 0
23
  - num_epochs: 6000
24
 
25
+ ### MLX.
26
+
27
  Trained with mlx-lm.
28
+
29
+ ### Thank you
30
+
31
  Thank you Alex Ziskind for original TinyLlama Crysis ❤️
32
+
33
+ This model is unquantized. I dont know why I trained it on Gemma 3 QAT because it acts weird when quantized.
34
+
35
+ ## Suggested parameters
36
+
37
+ - Temperature: `0.8`
38
+ - Max Tokens: `32`
39
+ - Max seq len: `64`