prelington commited on
Commit
6dd4f47
·
verified ·
1 Parent(s): 57be3b5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +24 -1
README.md CHANGED
@@ -10,4 +10,27 @@ The project aims to create a flexible LLM capable of natural conversation, reaso
10
  - **Pipeline:** `text-generation`
11
  - **Library:** `transformers`
12
  - **License:** Apache 2.0
13
- - **Training:** AutoTrain (no dataset
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  - **Pipeline:** `text-generation`
11
  - **Library:** `transformers`
12
  - **License:** Apache 2.0
13
+ - **Training:** AutoTrain (no dataset yet — pretrained mode)
14
+
15
+ ---
16
+
17
+ ## 🚀 Goals
18
+ - Serve as a foundation for text generation and conversational tasks
19
+ - Support English and optionally other languages
20
+ - Enable later fine-tuning with domain-specific data
21
+
22
+ ---
23
+
24
+ ## 📘 Usage
25
+
26
+ ```python
27
+ from transformers import AutoTokenizer, AutoModelForCausalLM
28
+
29
+ tokenizer = AutoTokenizer.from_pretrained("prelington/Mineral-1B")
30
+ model = AutoModelForCausalLM.from_pretrained("prelington/Mineral-1B")
31
+
32
+ prompt = "Hello! What is Mineral-1B?"
33
+ inputs = tokenizer(prompt, return_tensors="pt")
34
+ outputs = model.generate(**inputs, max_new_tokens=100)
35
+
36
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))