ItsMeDevRoland commited on
Commit
0be289c
·
verified ·
1 Parent(s): a1550b7

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +64 -8
README.md CHANGED
@@ -1,22 +1,78 @@
1
  ---
2
- base_model: unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
3
  tags:
 
 
4
  - text-generation-inference
5
  - transformers
6
  - unsloth
7
  - llama
8
  - trl
9
- license: apache-2.0
 
 
 
10
  language:
11
  - en
 
 
 
 
 
12
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
13
 
14
- # Uploaded model
15
 
16
- - **Developed by:** N-Bot-Int
17
- - **License:** apache-2.0
18
- - **Finetuned from model :** unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
 
19
 
20
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
 
 
 
 
 
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
1
  ---
2
+ license: mit
3
  tags:
4
+ - unsloth
5
+ - Uncensored
6
  - text-generation-inference
7
  - transformers
8
  - unsloth
9
  - llama
10
  - trl
11
+ datasets:
12
+ - openerotica/mixed-rp
13
+ - kingbri/PIPPA-shareGPT
14
+ - flammenai/character-roleplay-DPO
15
  language:
16
  - en
17
+ base_model:
18
+ - N-Bot-Int/OpenRP3B-Llama3.2
19
+ new_version: N-Bot-Int/OpenElla3-Llama3.2B
20
+ pipeline_tag: text-generation
21
+ library_name: peft
22
  ---
23
+ <a href="https://ibb.co/GvDjFcVp"><img src="https://i.ibb.co/pvTpnJ3w/image.webp" alt="image" border="0"></a>
24
+
25
+ # Llama3.2 - OpenElla3B
26
+ - OpenElla Model **B**, is a Llama3.2 **3B** Parameter Model,
27
+ That is fine-tuned for Roleplaying purposes, even if it only have a limited Parameters.
28
+ This is achieved through Series of Dataset Finetuning, using 3 Dataset with different
29
+ Weight, Aiming to Counter Llama3.2's Generalist Approach and focusing On Specializing with
30
+ Roleplaying and Acting.
31
+
32
+ - OpenElla3A Excells in Outputting **RAW** and **UNCENSORED** Output And Acknowledges OpenElla3A's weakness
33
+ for Following Prompt, Due to this, the model is re-finetuned, which **solves the issue with OpenElla3A's
34
+ Disobidience**, This allows the Model to Engage in Uncensored response and with appropriate responses, rivaling
35
+ its older models
36
+
37
+ - OpenElla3B contains more Fine-tuned Dataset so please Report any issues found through our email
38
+ <link src="mailto:nexus.networkinteractives@gmail.com">nexus.networkinteractives@gmail.com</link>,
39
+ about any overfitting, or improvements for the future Model **C**,
40
+ Once again feel free to Modify the LORA to your likings, However please consider Adding this Page
41
+ for credits and if you'll increase its **NSFW**, then please handle it with care and ethical considerations
42
+
43
+ - OpenElla3B is
44
+ - **Developed by:** N-Bot-Int
45
+ - **License:** apache-2.0
46
+ - **Parent Model from model:** unsloth/llama-3.2-3b-instruct-unsloth-bnb-4bit
47
+ - **Sequential Trained from Model:** N-Bot-Int/OpenElla3-Llama3.2A
48
+ - **Dataset Combined Using:** Mosher-R1(Propietary Software)
49
+
50
+ - OpenElla3B Is **NOT YET RANKED WITH ANY METRICS**
51
+ - Feel free to support by Emailing me: <link src="mailto:nexus.networkinteractives@gmail.com">nexus.networkinteractives@gmail.com</link>
52
+
53
+ - # Notice
54
+ - **For a Good Experience, Please use**
55
+ - Low temperature 1.5, min_p = 0.1 and max_new_tokens = 128
56
 
 
57
 
58
+ - # Detail card:
59
+ - Parameter
60
+ - 3 Billion Parameters
61
+ - (Please visit your GPU Vendor if you can Run 3B models)
62
 
63
+ - Training
64
+ - 500 steps
65
+ - Mixed-RP Startup Dataset
66
+ - 200 steps
67
+ - PIPPA-ShareGPT for bypassing Llama's Censorship Model
68
+ - 150 steps(Re-fining)
69
+ - PIPPA-ShareGPT to further increase weight of uncensorsed prompt
70
+ - 500 steps(Lower LR)
71
+ - Character-roleplay-DO to further encourage the Model to respond appropriately with the RP scenario
72
 
73
+ - Finetuning tool:
74
+ - Unsloth AI
75
+ - This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
76
+ [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
77
+ - Fine-tuned Using:
78
+ - Google Colab