RichardErkhov commited on
Commit
8185666
·
verified ·
1 Parent(s): e7de0e9

uploaded readme

Browse files
Files changed (1) hide show
  1. README.md +58 -0
README.md ADDED
@@ -0,0 +1,58 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Quantization made by Richard Erkhov.
2
+
3
+ [Github](https://github.com/RichardErkhov)
4
+
5
+ [Discord](https://discord.gg/pvy7H8DZMG)
6
+
7
+ [Request more models](https://github.com/RichardErkhov/quant_request)
8
+
9
+
10
+ minirewrite - bnb 4bits
11
+ - Model creator: https://huggingface.co/kevin009/
12
+ - Original model: https://huggingface.co/kevin009/minirewrite/
13
+
14
+
15
+
16
+
17
+ Original model description:
18
+ ---
19
+ language:
20
+ - en
21
+ license: apache-2.0
22
+ tags:
23
+ - text-generation-inference
24
+ - transformers
25
+ - mistral
26
+ - trl
27
+ ---
28
+
29
+ # Model Card: Minimalist Assistant
30
+
31
+ ## Model Details
32
+ - **Base Model**: Mistral Instruct v2
33
+ - **Tokenizer**: based on Mistral Instruction following
34
+
35
+ ## Intended Use
36
+ - As Editor Assistant for revision and paraphrasing
37
+ - Avoids technical jargon in favor of clear and accessible language
38
+
39
+ ## Training Data
40
+ - **Initial Training**: 14,000 conversations in minimalist style and more accessible language
41
+ - Dataset: kevin009/system-defined-sft-llama3-14k
42
+ - **Further Training**: 8,000 revision conversations to enhance rewriting and paraphrasing tasks.
43
+
44
+ ## Performance and Limitations
45
+ - **Limitations**:
46
+ - May produce shorter outputs compared to original version.
47
+ - Potential biases
48
+
49
+ ## Ethical Considerations
50
+ - Designed for daily use, potential biases from training data should be considered
51
+ - The model does not have implemented safety measures to prevent generation of potentially harmful or offensive content
52
+
53
+ ## Additional Information
54
+ - Fine-tuned to address limitations in writing tasks observed in other models
55
+ - Personalized for everyday use cases
56
+ - Motivation for development was to create a model better suited for writing tasks, as existing models were found lacking in this area
57
+ - SFT fine-tuned model
58
+