Enhanced model card with badges, tutorials, and documentation
Browse files
README.md
CHANGED
|
@@ -12,43 +12,90 @@ tags:
|
|
| 12 |
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
|
| 15 |
-
|
| 16 |
|
| 17 |
-
|
| 18 |
|
| 19 |
-
|
|
|
|
|
|
|
| 20 |
|
| 21 |
-
|
| 22 |
|
| 23 |
-
|
| 24 |
-
- **Extended Context**: 8192 token context window
|
| 25 |
-
- **Q4_K_M Quantization**: Efficient 4-bit quantization
|
| 26 |
|
| 27 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
|
| 29 |
| Property | Value |
|
| 30 |
|----------|-------|
|
| 31 |
-
| Parameters | 1.
|
| 32 |
-
| Quantization | Q4_K_M |
|
| 33 |
-
| Context
|
| 34 |
-
|
|
| 35 |
-
|
|
|
|
|
| 36 |
|
| 37 |
-
##
|
| 38 |
|
| 39 |
-
|
| 40 |
-
|
| 41 |
-
|
| 42 |
|
| 43 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 45 |
```rust
|
| 46 |
use ruvllm::hub::ModelDownloader;
|
| 47 |
|
| 48 |
-
let
|
| 49 |
-
|
|
|
|
| 50 |
```
|
| 51 |
|
| 52 |
-
##
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 53 |
|
| 54 |
-
Apache 2.0
|
|
|
|
| 12 |
pipeline_tag: text-generation
|
| 13 |
---
|
| 14 |
|
| 15 |
+
<div align="center">
|
| 16 |
|
| 17 |
+
# RuvLTRA Medium
|
| 18 |
|
| 19 |
+
[](https://opensource.org/licenses/Apache-2.0)
|
| 20 |
+
[](https://huggingface.co/ruv/ruvltra-medium)
|
| 21 |
+
[](https://github.com/ggerganov/ggml/blob/master/docs/gguf.md)
|
| 22 |
|
| 23 |
+
**⚖️ Balanced Model for General-Purpose Tasks**
|
| 24 |
|
| 25 |
+
</div>
|
|
|
|
|
|
|
| 26 |
|
| 27 |
+
---
|
| 28 |
+
|
| 29 |
+
## Overview
|
| 30 |
+
|
| 31 |
+
RuvLTRA Medium provides the sweet spot between capability and resource usage. Ideal for desktop applications, development workstations, and moderate-scale deployments.
|
| 32 |
+
|
| 33 |
+
## Model Card
|
| 34 |
|
| 35 |
| Property | Value |
|
| 36 |
|----------|-------|
|
| 37 |
+
| **Parameters** | 1.1 Billion |
|
| 38 |
+
| **Quantization** | Q4_K_M |
|
| 39 |
+
| **Context** | 8,192 tokens |
|
| 40 |
+
| **Size** | ~669 MB |
|
| 41 |
+
| **Min RAM** | 2 GB |
|
| 42 |
+
| **Recommended RAM** | 4 GB |
|
| 43 |
|
| 44 |
+
## 🚀 Quick Start
|
| 45 |
|
| 46 |
+
```bash
|
| 47 |
+
# Download
|
| 48 |
+
wget https://huggingface.co/ruv/ruvltra-medium/resolve/main/ruvltra-1.1b-q4_k_m.gguf
|
| 49 |
|
| 50 |
+
# Run inference
|
| 51 |
+
./llama-cli -m ruvltra-1.1b-q4_k_m.gguf \
|
| 52 |
+
-p "Explain quantum computing in simple terms:" \
|
| 53 |
+
-n 512 -c 8192
|
| 54 |
+
```
|
| 55 |
|
| 56 |
+
## 💡 Use Cases
|
| 57 |
+
|
| 58 |
+
- **Development**: Code assistance and generation
|
| 59 |
+
- **Writing**: Content creation and editing
|
| 60 |
+
- **Analysis**: Document summarization
|
| 61 |
+
- **Chat**: Conversational AI applications
|
| 62 |
+
|
| 63 |
+
## 🔧 Integration
|
| 64 |
+
|
| 65 |
+
### Rust
|
| 66 |
```rust
|
| 67 |
use ruvllm::hub::ModelDownloader;
|
| 68 |
|
| 69 |
+
let path = ModelDownloader::new()
|
| 70 |
+
.download("ruv/ruvltra-medium", None)
|
| 71 |
+
.await?;
|
| 72 |
```
|
| 73 |
|
| 74 |
+
### Python
|
| 75 |
+
```python
|
| 76 |
+
from llama_cpp import Llama
|
| 77 |
+
from huggingface_hub import hf_hub_download
|
| 78 |
+
|
| 79 |
+
model_path = hf_hub_download("ruv/ruvltra-medium", "ruvltra-1.1b-q4_k_m.gguf")
|
| 80 |
+
llm = Llama(model_path=model_path, n_ctx=8192)
|
| 81 |
+
```
|
| 82 |
+
|
| 83 |
+
### OpenAI-Compatible Server
|
| 84 |
+
|
| 85 |
+
```bash
|
| 86 |
+
python -m llama_cpp.server \
|
| 87 |
+
--model ruvltra-1.1b-q4_k_m.gguf \
|
| 88 |
+
--host 0.0.0.0 --port 8000
|
| 89 |
+
```
|
| 90 |
+
|
| 91 |
+
## Performance
|
| 92 |
+
|
| 93 |
+
| Platform | Tokens/sec |
|
| 94 |
+
|----------|------------|
|
| 95 |
+
| M2 Pro (Metal) | 65 tok/s |
|
| 96 |
+
| RTX 4080 (CUDA) | 95 tok/s |
|
| 97 |
+
| i9-13900K (CPU) | 25 tok/s |
|
| 98 |
+
|
| 99 |
+
---
|
| 100 |
|
| 101 |
+
**License**: Apache 2.0 | **GitHub**: [ruvnet/ruvector](https://github.com/ruvnet/ruvector)
|