turtle170 commited on
Commit
8bd2e83
·
verified ·
1 Parent(s): f09a2f5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +35 -10
README.md CHANGED
@@ -6,33 +6,58 @@ tags:
6
  - base_model:adapter:microsoft/Phi-3-mini-4k-instruct
7
  - lora
8
  - transformers
 
 
 
 
 
 
9
  ---
10
 
11
  # Model Card for Model ID
12
 
13
- <!-- Provide a quick summary of what the model is/does. -->
 
14
 
15
 
16
 
17
  ## Model Details
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
18
 
19
  ### Model Description
20
 
21
- <!-- Provide a longer summary of what this model is. -->
22
 
23
 
24
 
25
- - **Developed by:** [More Information Needed]
26
- - **Funded by [optional]:** [More Information Needed]
27
- - **Shared by [optional]:** [More Information Needed]
28
- - **Model type:** [More Information Needed]
29
- - **Language(s) (NLP):** [More Information Needed]
30
- - **License:** [More Information Needed]
31
- - **Finetuned from model [optional]:** [More Information Needed]
32
 
33
  ### Model Sources [optional]
34
 
35
- <!-- Provide the basic links for the model. -->
36
 
37
  - **Repository:** [More Information Needed]
38
  - **Paper [optional]:** [More Information Needed]
 
6
  - base_model:adapter:microsoft/Phi-3-mini-4k-instruct
7
  - lora
8
  - transformers
9
+ license: apache-2.0
10
+ datasets:
11
+ - teknium/OpenHermes-2.5
12
+ - Magpie-Align/Magpie-Phi3-Pro-300K-Filtered
13
+ language:
14
+ - en
15
  ---
16
 
17
  # Model Card for Model ID
18
 
19
+ Phi-3-Mini-OpenHermes-Magpie-V1 is a general purpose model trained on both the teknium/OpenHermes-2.5 dataset and the Magpie-Align/Phi3-Pro-300K-Filtered dataset
20
+ and designed to provide speed, efficiency, and intelligence.
21
 
22
 
23
 
24
  ## Model Details
25
+ OpenHermes dataset:
26
+ 1 Epoch
27
+ 8 Batch Size
28
+ 1 Gradient Accumulation
29
+ 5e-5 LR
30
+ 16 LoRa r
31
+ 32 LoRa Alpha
32
+ 300 Warmup steps
33
+ 500 Eval steps
34
+ Trained only on Attention layers.
35
+
36
+ Magpie dataset:
37
+ 1 Epoch
38
+ 16 Batch Size
39
+ 1 Gradient Accumulation
40
+ 1e-4 LR
41
+ 16 LoRa r
42
+ 32 LoRa Alpha
43
+ 150 Warmup steps
44
+ 500 Eval steps
45
+ Trained with Gate, Up, and Down layers.
46
 
47
  ### Model Description
48
 
49
+ This model excels at creating bullet point formatting, while still mantaining
50
 
51
 
52
 
53
+ - **Developed by:** Turtle170 (anonymous
54
+ - **Language(s) (NLP):** English
55
+ - **License:** apache-2.0
56
+ - **Finetuned from model :** Phi-3-Mini-4k-Instruct with turtle170/Phi-3-Mini-OpenHermes-V1 adapters
 
 
 
57
 
58
  ### Model Sources [optional]
59
 
60
+ <!-- Provide the basic
61
 
62
  - **Repository:** [More Information Needed]
63
  - **Paper [optional]:** [More Information Needed]