output = llm(
"Once upon a time,",
max_tokens=512,
echo=True
)
print(output)SafeAgent 7B v1
The AI model powering SafeAgent ? your personal AI agent that runs entirely on your machine.
SafeAgent lets you chat with AI, read your emails, search the web, manage GitHub, post to Slack, and automate workflows ? all running locally. Your data never leaves your network. No subscriptions. No cloud dependency.
Run It Now
ollama run SafeAgent/safeagent-7b-v1
What SafeAgent Can Do With This Model
- Read and send emails (Gmail, Outlook, Yahoo)
- Search the web and summarise results
- Write, fix, and explain code
- Plan trips, book restaurants, compare products
- Automate workflows and scheduled tasks
- Everything stays on your machine -- AES-256 encrypted
Install SafeAgent
curl -sL https://www.safeagent.dev/docker-compose.yml -o docker-compose.yml && docker compose up
Open http://localhost:3000 -- running in 30 seconds.
Model Details
| Property | Value |
|---|---|
| Base model | Mistral 7B v0.1 |
| Fine-tuning method | QLoRA 4-bit |
| Training data | 80,000 OpenOrca instruction examples |
| Training time | 20 hours |
| Final loss | 0.5564 |
| Format | GGUF f16 |
Why Local AI?
| Cloud AI | SafeAgent |
|---|---|
| Your data on their servers | Everything on your machine |
| $20-100/month | Free forever |
| Rate limited | No limits -- your hardware |
| Vendor lock-in | Open source, always |
| They train on your data | Your data stays yours |
Links
- Website: https://safeagent.dev
- DockerHub: https://hub.docker.com/u/gurmukhs
- Built by Gurmukh Singh: https://www.linkedin.com/in/gurmukh-singh-38a1a4246/
- Downloads last month
- 58
Hardware compatibility
Log In to add your hardware
We're not able to determine the quantization variants.
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐ Ask for provider support
# !pip install llama-cpp-python from llama_cpp import Llama llm = Llama.from_pretrained( repo_id="SafeAgent/safeagent-7b-v1", filename="safeagent-7b-v1.gguf", )