ruv commited on
Commit
2093b08
·
verified ·
1 Parent(s): 2855f1e

Enhanced model card with badges, tutorials, and documentation

Browse files
Files changed (1) hide show
  1. README.md +69 -22
README.md CHANGED
@@ -12,43 +12,90 @@ tags:
12
  pipeline_tag: text-generation
13
  ---
14
 
15
- # RuvLTRA Medium (1.1B)
16
 
17
- **Balanced RuvLTRA model for general-purpose tasks**
18
 
19
- ## Model Description
 
 
20
 
21
- RuvLTRA Medium provides a balance between capability and resource usage, suitable for general-purpose text generation and coding tasks.
22
 
23
- - **SONA Integration**: Self-Optimizing Neural Architecture support
24
- - **Extended Context**: 8192 token context window
25
- - **Q4_K_M Quantization**: Efficient 4-bit quantization
26
 
27
- ## Model Details
 
 
 
 
 
 
28
 
29
  | Property | Value |
30
  |----------|-------|
31
- | Parameters | 1.1B |
32
- | Quantization | Q4_K_M |
33
- | Context Length | 8192 tokens |
34
- | File Size | ~669 MB |
35
- | Format | GGUF |
 
36
 
37
- ## Hardware Requirements
38
 
39
- - **Minimum RAM**: 2 GB
40
- - **Recommended RAM**: 4 GB
41
- - **Supports**: Apple Neural Engine, Metal, CUDA, CPU
42
 
43
- ## Usage
 
 
 
 
44
 
 
 
 
 
 
 
 
 
 
 
45
  ```rust
46
  use ruvllm::hub::ModelDownloader;
47
 
48
- let downloader = ModelDownloader::new();
49
- let path = downloader.download("ruv/ruvltra-medium", None).await?;
 
50
  ```
51
 
52
- ## License
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
 
54
- Apache 2.0
 
12
  pipeline_tag: text-generation
13
  ---
14
 
15
+ <div align="center">
16
 
17
+ # RuvLTRA Medium
18
 
19
+ [![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
20
+ [![HuggingFace](https://img.shields.io/badge/🤗%20Hugging%20Face-Model-yellow)](https://huggingface.co/ruv/ruvltra-medium)
21
+ [![GGUF](https://img.shields.io/badge/Format-GGUF-green)](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md)
22
 
23
+ **⚖️ Balanced Model for General-Purpose Tasks**
24
 
25
+ </div>
 
 
26
 
27
+ ---
28
+
29
+ ## Overview
30
+
31
+ RuvLTRA Medium provides the sweet spot between capability and resource usage. Ideal for desktop applications, development workstations, and moderate-scale deployments.
32
+
33
+ ## Model Card
34
 
35
  | Property | Value |
36
  |----------|-------|
37
+ | **Parameters** | 1.1 Billion |
38
+ | **Quantization** | Q4_K_M |
39
+ | **Context** | 8,192 tokens |
40
+ | **Size** | ~669 MB |
41
+ | **Min RAM** | 2 GB |
42
+ | **Recommended RAM** | 4 GB |
43
 
44
+ ## 🚀 Quick Start
45
 
46
+ ```bash
47
+ # Download
48
+ wget https://huggingface.co/ruv/ruvltra-medium/resolve/main/ruvltra-1.1b-q4_k_m.gguf
49
 
50
+ # Run inference
51
+ ./llama-cli -m ruvltra-1.1b-q4_k_m.gguf \
52
+ -p "Explain quantum computing in simple terms:" \
53
+ -n 512 -c 8192
54
+ ```
55
 
56
+ ## 💡 Use Cases
57
+
58
+ - **Development**: Code assistance and generation
59
+ - **Writing**: Content creation and editing
60
+ - **Analysis**: Document summarization
61
+ - **Chat**: Conversational AI applications
62
+
63
+ ## 🔧 Integration
64
+
65
+ ### Rust
66
  ```rust
67
  use ruvllm::hub::ModelDownloader;
68
 
69
+ let path = ModelDownloader::new()
70
+ .download("ruv/ruvltra-medium", None)
71
+ .await?;
72
  ```
73
 
74
+ ### Python
75
+ ```python
76
+ from llama_cpp import Llama
77
+ from huggingface_hub import hf_hub_download
78
+
79
+ model_path = hf_hub_download("ruv/ruvltra-medium", "ruvltra-1.1b-q4_k_m.gguf")
80
+ llm = Llama(model_path=model_path, n_ctx=8192)
81
+ ```
82
+
83
+ ### OpenAI-Compatible Server
84
+
85
+ ```bash
86
+ python -m llama_cpp.server \
87
+ --model ruvltra-1.1b-q4_k_m.gguf \
88
+ --host 0.0.0.0 --port 8000
89
+ ```
90
+
91
+ ## Performance
92
+
93
+ | Platform | Tokens/sec |
94
+ |----------|------------|
95
+ | M2 Pro (Metal) | 65 tok/s |
96
+ | RTX 4080 (CUDA) | 95 tok/s |
97
+ | i9-13900K (CPU) | 25 tok/s |
98
+
99
+ ---
100
 
101
+ **License**: Apache 2.0 | **GitHub**: [ruvnet/ruvector](https://github.com/ruvnet/ruvector)