nRuaif commited on
Commit
416e9d2
·
1 Parent(s): 6c3a385

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +5 -8
README.md CHANGED
@@ -4,7 +4,7 @@ base_model: mistralai/Mistral-7B-v0.1
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
- - name: aesir-rpg-mistral-out
8
  results: []
9
  ---
10
 
@@ -12,7 +12,7 @@ model-index:
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
15
- # aesir-rpg-mistral-out
16
 
17
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
18
  It achieves the following results on the evaluation set:
@@ -20,22 +20,19 @@ It achieves the following results on the evaluation set:
20
 
21
  ## Model description
22
 
23
- More information needed
24
 
25
  ## Intended uses & limitations
26
 
27
- More information needed
28
 
29
- ## Training and evaluation data
30
 
31
- More information needed
32
 
33
- ## Training procedure
34
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - learning_rate: 0.0001
39
  - train_batch_size: 4
40
  - eval_batch_size: 4
41
  - seed: 42
 
4
  tags:
5
  - generated_from_trainer
6
  model-index:
7
+ - name: Kimiko-Mistral-7B
8
  results: []
9
  ---
10
 
 
12
  should probably proofread and complete it, then remove this comment. -->
13
 
14
  [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
15
+ # Kimiko-Mistral-7B
16
 
17
  This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
18
  It achieves the following results on the evaluation set:
 
20
 
21
  ## Model description
22
 
23
+ Same dataset as Kimiko-v2 but on new model. THIS IS NOT TRAIN ON V3 DATASET
24
 
25
  ## Intended uses & limitations
26
 
27
+ As a finetuning experiment on new 7B model. You can use this for roleplay or as an assistant
28
 
 
29
 
 
30
 
 
31
 
32
  ### Training hyperparameters
33
 
34
  The following hyperparameters were used during training:
35
+ - learning_rate: 0.00005
36
  - train_batch_size: 4
37
  - eval_batch_size: 4
38
  - seed: 42