Shoukaku07 commited on
Commit
d260543
·
verified ·
1 Parent(s): 6079965

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +40 -3
README.md CHANGED
@@ -1,3 +1,40 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: mit
3
+ datasets:
4
+ - wikimedia/wikipedia
5
+ - roneneldan/TinyStories
6
+ - ajibawa-2023/Children-Stories-Collection
7
+ - stas/c4-en-10k
8
+ pipeline_tag: text-generation
9
+ ---
10
+
11
+ # Serayuki-2B
12
+
13
+ **Model Developer**: Shoukaku07
14
+ <br>
15
+ **Model Type**: Causal Language Model
16
+
17
+ ## Example Usage
18
+
19
+ Using Hugging Face Transformers:
20
+
21
+ ```python
22
+ from transformers import AutoModelForCausalLM, AutoTokenizer
23
+
24
+ model = AutoModelForCausalLM.from_pretrained("SeraphyneLab/Serayuki-2B")
25
+ tokenizer = AutoTokenizer.from_pretrained("SeraphyneLab/Serayuki-2B")
26
+
27
+ input_text = "Once upon a time"
28
+ inputs = tokenizer(input_text, return_tensors="pt")
29
+ outputs = model.generate(**inputs, max_new_tokens=128)
30
+
31
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
32
+ ```
33
+
34
+ ## License
35
+
36
+ This model is licensed under the [MIT License](https://opensource.org/licenses/MIT).
37
+
38
+ ## Tokenizer Notice
39
+
40
+ This model was trained from scratch; however, it uses the tokenizer from Meta’s LLaMA 3.2 3B Instruct model. As such, the tokenizer is subject to Meta’s [LLaMA 3 license](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct/blob/main/LICENSE.txt). Please review their terms before using this model or tokenizer in commercial applications.