File size: 2,275 Bytes
2855f1e
 
 
 
 
 
 
 
 
 
 
 
 
 
2093b08
2855f1e
2093b08
2855f1e
2093b08
 
 
2855f1e
2093b08
2855f1e
2093b08
2855f1e
2093b08
 
 
 
 
 
 
2855f1e
 
 
2093b08
 
 
 
 
 
2855f1e
2093b08
2855f1e
2093b08
 
 
2855f1e
2093b08
 
 
 
 
2855f1e
2093b08
 
 
 
 
 
 
 
 
 
2855f1e
 
 
2093b08
 
 
2855f1e
 
2093b08
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2855f1e
2093b08
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
---
language:
- en
license: apache-2.0
library_name: gguf
tags:
- ruvltra
- sona
- adaptive-learning
- gguf
- quantized
pipeline_tag: text-generation
---

<div align="center">

# RuvLTRA Medium

[![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
[![HuggingFace](https://img.shields.io/badge/🤗%20Hugging%20Face-Model-yellow)](https://huggingface.co/ruv/ruvltra-medium)
[![GGUF](https://img.shields.io/badge/Format-GGUF-green)](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md)

**⚖️ Balanced Model for General-Purpose Tasks**

</div>

---

## Overview

RuvLTRA Medium provides the sweet spot between capability and resource usage. Ideal for desktop applications, development workstations, and moderate-scale deployments.

## Model Card

| Property | Value |
|----------|-------|
| **Parameters** | 1.1 Billion |
| **Quantization** | Q4_K_M |
| **Context** | 8,192 tokens |
| **Size** | ~669 MB |
| **Min RAM** | 2 GB |
| **Recommended RAM** | 4 GB |

## 🚀 Quick Start

```bash
# Download
wget https://huggingface.co/ruv/ruvltra-medium/resolve/main/ruvltra-1.1b-q4_k_m.gguf

# Run inference
./llama-cli -m ruvltra-1.1b-q4_k_m.gguf \
  -p "Explain quantum computing in simple terms:" \
  -n 512 -c 8192
```

## 💡 Use Cases

- **Development**: Code assistance and generation
- **Writing**: Content creation and editing
- **Analysis**: Document summarization
- **Chat**: Conversational AI applications

## 🔧 Integration

### Rust
```rust
use ruvllm::hub::ModelDownloader;

let path = ModelDownloader::new()
    .download("ruv/ruvltra-medium", None)
    .await?;
```

### Python
```python
from llama_cpp import Llama
from huggingface_hub import hf_hub_download

model_path = hf_hub_download("ruv/ruvltra-medium", "ruvltra-1.1b-q4_k_m.gguf")
llm = Llama(model_path=model_path, n_ctx=8192)
```

### OpenAI-Compatible Server

```bash
python -m llama_cpp.server \
  --model ruvltra-1.1b-q4_k_m.gguf \
  --host 0.0.0.0 --port 8000
```

## Performance

| Platform | Tokens/sec |
|----------|------------|
| M2 Pro (Metal) | 65 tok/s |
| RTX 4080 (CUDA) | 95 tok/s |
| i9-13900K (CPU) | 25 tok/s |

---

**License**: Apache 2.0 | **GitHub**: [ruvnet/ruvector](https://github.com/ruvnet/ruvector)