File size: 3,947 Bytes
8629504
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5d151b9
f38c051
8629504
 
 
 
f38c051
8629504
f38c051
8629504
 
5d151b9
8629504
5d151b9
8629504
f38c051
8629504
 
 
5d151b9
8629504
5d151b9
8629504
5d151b9
8629504
f38c051
5d151b9
f38c051
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8629504
 
 
 
 
5d151b9
 
 
f38c051
 
8629504
f38c051
8629504
f38c051
 
 
8629504
 
 
f38c051
8629504
f38c051
 
 
 
 
 
 
 
8629504
f38c051
8629504
 
 
 
 
f38c051
 
 
8629504
f38c051
 
5d151b9
f38c051
8629504
f38c051
8629504
f38c051
5d151b9
 
 
8629504
 
 
 
 
 
 
 
5d151b9
8629504
5d151b9
f38c051
 
8629504
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5d151b9
f38c051
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
---
language:
  - cs
  - sk
  - en
  - de
license: apache-2.0
base_model: EuroLLM-9B
quantization: Q8_0
tags:
  - gguf
  - llama.cpp
  - offline
  - local-ai
  - multilingual
  - cli-runtime
  - ai-runtime
pipeline_tag: text-generation
library_name: llama.cpp
---

# Offline AI 2.2 – EuroLLM-9B-Q8_0 (GGUF)

Offline AI 2.2 is a fully local AI runtime environment built around digital sovereignty, privacy, and system autonomy.

No cloud.  
No telemetry.  
No tracking.  
No external dependencies.  

Everything runs locally via **llama.cpp**.

---

## 🖥️ CLI Preview

Below is the Offline AI runtime interface:

![Offline AI CLI Help Menu](cli_help_menu.png)

Offline AI is no longer just a model launcher.

It is a **local AI runtime environment** designed to manage and operate language models fully offline with a structured command interface.

Core capabilities include:

- CLI runtime environment  
- Model lifecycle management  
- Profile-based workspace system  
- Snapshot conversation archiving  
- Runtime diagnostics and monitoring  
- Administrative control layer  

The architecture is designed as a foundation for **multi-model local AI systems**.

---

## 🧠 RUNTIME ARCHITECTURE

Offline AI uses a layered architecture:

User (CLI)

Python Runtime

C++ Inference Engine (llama.cpp)

GGUF Language Model


The Python runtime acts as the **control layer**, responsible for:

- command handling  
- model orchestration  
- workspace profiles  
- snapshots and notes  
- system diagnostics  
- administrative operations  

The inference backend is a lightweight C++ wrapper around **llama.cpp** with real-time token streaming.

---

## 🔧 TECHNICAL INFORMATION

Base model: EuroLLM-9B  
Quantization: Q8_0 (GGUF)  
Format: llama.cpp compatible  
Inference engine: llama.cpp  
Offline AI Version: 2.2  
Recommended RAM: 16 GB  
Platforms: macOS, Windows, Linux  

This repository distributes a **quantized GGUF Q8_0 variant** of the EuroLLM-9B model optimized for efficient local inference.

The original model weights are **not modified and not fine-tuned** as part of this project.

---

## 🚀 WHAT'S NEW IN 2.2

- Structured CLI runtime environment  
- Model lifecycle management system  
- Model alias system  
- Workspace profiles and isolation  
- Snapshot conversation archiving  
- Runtime diagnostics and monitoring  
- Administrative control mode  
- Improved modular runtime architecture  

Offline AI 2.2 evolves the project from a simple model launcher into a **local AI runtime platform** prepared for managing multiple specialized AI models.

---

## 🔐 PROJECT PHILOSOPHY

Offline AI demonstrates that modern AI systems can operate fully offline.

The project explores the idea that:

- AI does not require cloud infrastructure  
- Open models can run independently on personal hardware  
- AI tools can respect user privacy  
- Local-first computing is a viable architecture  

Offline AI promotes:

- Digital sovereignty  
- Transparent system design  
- Offline experimentation  
- User-controlled AI environments  

---

## 📄 MODEL ORIGIN & LICENSE

Model: EuroLLM-9B  
Original authors: EuroLLM Project consortium  
Funded by: European Union research initiatives  
Base model license: Apache License 2.0  

Quantized distribution: GGUF Q8_0  
Runtime engine: llama.cpp (MIT License)  
Offline AI runtime interface: © David Káninský  

All components are used in compliance with their respective licenses.

---

## ⚠️ DISCLAIMER

This project is an educational and experimental implementation.

It is not a commercial AI service and does not replace professional advice.  
Outputs are not intended for legal, medical, financial, or critical decision-making use.

Use beyond personal, research, or educational purposes is at your own responsibility.

---

## 🌍 PROJECT

Website: https://OfflineAI.online  
Domains: .cz / .sk / .de  

Offline AI Runtime  
Author: David Káninský