RessAI commited on
Commit
6b70a03
·
verified ·
1 Parent(s): 3d54a0a

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +83 -0
README.md ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - HuggingFaceFW/fineweb-edu
5
+ language:
6
+ - en
7
+ library_name: transformers
8
+ tags:
9
+ - pytorch
10
+ - causal-lm
11
+ - text-generation
12
+ - onner
13
+ ---
14
+ # 🚀 RessAI-Ultra 2B
15
+
16
+ **RessAI-Ultra 2B** is a custom 2.56 Billion parameter language model built on the highly optimized `onner` architecture. Designed for deep reasoning and long-context understanding, it features a 128k context window and a "Deep & Dense" layer structure.
17
+
18
+ <div align="center">
19
+ <img src="https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/transformers_logo_name.png" width="300"/>
20
+ </div>
21
+
22
+ ## 🔍 Model Details
23
+
24
+ - **Model Name:** RessAI-Ultra 2B
25
+ - **Organization:** RessAI
26
+ - **Architecture:** `RessAiForCausalLM` (Custom Llama-based structure)
27
+ - **Model Type:** `onner`
28
+ - **Parameters:** ~2.56 Billion
29
+ - **Context Window:** 131,072 tokens (128k)
30
+ - **Training Precision:** Bfloat16
31
+ - **License:** Apache 2.0
32
+
33
+ ## 🧠 Technical Specifications
34
+
35
+ RessAI-Ultra utilizes a custom configuration designed for efficiency and long-range dependencies:
36
+
37
+ | Hyperparameter | Value | Description |
38
+ | :--- | :--- | :--- |
39
+ | **Hidden Size** | 2560 | Custom embedding dimension |
40
+ | **Layers** | 32 | Deep network structure |
41
+ | **Attention Heads** | 32 | Standard query heads |
42
+ | **KV Heads** | 4 | Grouped Query Attention (GQA) 8:1 Ratio |
43
+ | **Intermediate Size** | 7168 | Wide MLP for high capacity |
44
+ | **RoPE Theta** | 2,000,000 | Enhanced for long context stability |
45
+ | **Vocab Size** | 128,256 | Llama-3 Tokenizer compatibility |
46
+
47
+ ## 💻 Usage
48
+
49
+ Because this model uses a custom architecture type (`onner`) and configuration (`RessAiConfig`), you can load it using the standard `transformers` library.
50
+
51
+ ### Python Code
52
+
53
+ ```python
54
+ from transformers import AutoModelForCausalLM, AutoTokenizer
55
+ import torch
56
+
57
+ model_id = "RessAI/RessAI-Ultra"
58
+
59
+ # Load Tokenizer
60
+ tokenizer = AutoTokenizer.from_pretrained(model_id)
61
+
62
+ # Load Model
63
+ # Note: Ensure you have the latest transformers version
64
+ model = AutoModelForCausalLM.from_pretrained(
65
+ model_id,
66
+ torch_dtype=torch.bfloat16,
67
+ device_map="auto",
68
+ trust_remote_code=True # Required for custom config/arch if code is present
69
+ )
70
+
71
+ # Inference
72
+ prompt = "The future of artificial intelligence is"
73
+ inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
74
+
75
+ outputs = model.generate(
76
+ **inputs,
77
+ max_new_tokens=100,
78
+ temperature=0.7,
79
+ top_p=0.9,
80
+ do_sample=True
81
+ )
82
+
83
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))