legolasyiu commited on
Commit
07746fb
·
verified ·
1 Parent(s): 8a5395e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +55 -0
README.md CHANGED
@@ -11,6 +11,61 @@ language:
11
  - en
12
  ---
13
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  # Uploaded model
15
 
16
  - **Developed by:** EpistemeAI
 
11
  - en
12
  ---
13
 
14
+ This is a reasoning and reflect instruction-tuned generative model in 3B size (text in/text out).
15
+
16
+ **Model Architecture:**
17
+ Llama 3.2 is an auto-regressive language model that uses an optimized transformer architecture. The tuned versions use supervised fine-tuning (SFT) and reinforcement learning with human feedback (RLHF) with GRPO fine tuning using unsloth, to align with human preferences for helpfulness and safety.
18
+ Fine tune with Numina math dataset.
19
+
20
+
21
+ ### Use with transformers
22
+
23
+ Starting with `transformers >= 4.43.0` onward, you can run conversational inference using the Transformers `pipeline` abstraction or by leveraging the Auto classes with the `generate()` function.
24
+
25
+ Make sure to update your transformers installation via `pip install --upgrade transformers`.
26
+
27
+ ```python
28
+ import torch
29
+ from transformers import pipeline
30
+
31
+ model_id = "EpistemeAI/ReasoningCore-3B-Instruct-r01-Reflect-ThinkMath"
32
+ pipe = pipeline(
33
+ "text-generation",
34
+ model=model_id,
35
+ torch_dtype=torch.bfloat16,
36
+ device_map="auto",
37
+ )
38
+ messages = [
39
+ {"role": "system", "content": "You are a powerful assistant Respond in the following format:
40
+ <reasoning>
41
+ ...
42
+ </reasoning>
43
+ <reflecting>
44
+ ...
45
+ </reflecting>
46
+ <answer>
47
+ ...
48
+ </answer>"},
49
+ {"role": "user", "content": "Which is bigger? 9.11 or 9.9?"},
50
+ ]
51
+ outputs = pipe(
52
+ messages,
53
+ max_new_tokens=256,
54
+ )
55
+ print(outputs[0]["generated_text"][-1])
56
+ ```
57
+
58
+ ## Using [SuperTransformer](https://github.com/tomtyiu/SuperTransformer-SHF)
59
+ ```python
60
+ import SuperTransformer
61
+ # Load SuperTransformer Class, (1) Loads Huggingface model, (2) System Prompt (3) Text/prompt (4)Max tokens
62
+ SuperTransformers = SuperTransformers("EpistemeAI/ReasoningCore-3B-Instruct-r01-Reflect-ThinkMath","You are a highly knowledgeable assistant with expertise in mathematics. <reasoning>...</reasoning><reflecting>...</reflecting><answer>...</answer>","What is the area of a circle, radius=16, reason step by step", 2026)
63
+ # 8-bit quantization
64
+ SuperTransformers.HuggingFaceTransformer8bit()
65
+ # or 4-bit quantization
66
+ SuperTransformers.HuggingFaceTransformer4bit()
67
+ ```
68
+
69
  # Uploaded model
70
 
71
  - **Developed by:** EpistemeAI