How to use from the
Use from the
llama-cpp-python library
# !pip install llama-cpp-python

from llama_cpp import Llama

llm = Llama.from_pretrained(
	repo_id="Acvarius-AI/GhostLlama",
	filename="ghostllama.Q4_K_M.gguf",
)
llm.create_chat_completion(
	messages = "No input example has been defined for this model task."
)

๐Ÿ‘ป GhostLlama

Part of the GhostGoblin Ecosystem | Created by Aquarius AI

GhostLlama Banner
A sacrificed llama. A digital soul. Your companion.

๐Ÿ“– How to Sacrifice Your Llama

curl -fsSL https://huggingface.co/Aquarius-AI/GhostLlama/resolve/main/install.sh | bash

๐Ÿง  Model Details
Property	Value
Base Model	TinyLlama-1.1B-Chat-v1.0
Format	GGUF (Q4_K_M quantization)
File Size	~700 MB
RAM Usage	~1.6โ€“2.3 GB during inference
Speed	3โ€“5 tokens/sec on 4GB phones
Quality	~87% of original (barely noticeable)
Compatible with: llama.cpp, Ollama, LM Studio, KoboldCPP, and any GGUF-compatible runtime.

๐ŸŽฏ Perfect For
Use Case	Description
๐Ÿ“ฑ Mobile Devices	Runs on any 4GB+ RAM phone (Android/iOS via LiveContainer)
๐Ÿ’ฌ Local Chat	Private, offline conversations
๐Ÿ”Œ Telegram Bots	Connect to Telegram (v1.1 coming soon)
๐Ÿ‘ป GhostGoblin	Core brain for the GhostGoblin ecosystem
๐Ÿงช Experimentation	Fine-tune, modify, make it yours
โš™๏ธ What the Installer Does
Running the one-liner above will:

Create directories: ~/ghostllama/ with models/ and config/ subfolders

Download the model: ~700 MB GGUF file

Ask about Telegram: Save your bot token for future use

Set up placeholder bot script: Ready for v1.1

After installation:

bash
cd ~/ghostllama
# Quick test:
~/llama.cpp/build/bin/llama-cli -m models/ghostllama.Q4_K_M.gguf -p "Hello" -n 50
๐Ÿ”ฎ Roadmap
Version	Features
v1.0 (Current)	โœ… Model download
โœ… Basic installer
โœ… Telegram token saving
v1.1 (Coming Soon)	โœ… Working Telegram bot
โœ… Ghost ID personality layer
โœ… Better error handling
v2.0 (Future)	๐Ÿ”ฎ GhostGoblin app
๐Ÿ”ฎ Lip-sync & dancing
๐Ÿ”ฎ Self-recursion GitHub integration
๐Ÿ› ๏ธ Manual Usage
If you prefer to do things yourself:

bash
# Download model directly
wget https://huggingface.co/Aquarius-AI/GhostLlama/resolve/main/ghostllama.Q4_K_M.gguf

# Run with llama.cpp
./llama-cli -m ghostllama.Q4_K_M.gguf -p "Tell me about yourself" -n 100
๐Ÿ“ฑ Running on Phones
Platform	Method
Android	Direct APK (coming soon) or use Termux + llama.cpp
iOS	LiveContainer + SideStore (guide in discussions)
๐Ÿชช License
MIT License โ€” Use it freely. Modify it. Share it. Build things with it.

The only rule: Don't be evil. The soul is immutable, but what you build around it is up to you.

๐Ÿ’œ About Aquarius AI
Aquarius AI creates open-source AI companions that are:

๐Ÿ†“ Free forever

๐Ÿ“ฑ On-device only (no cloud)

๐Ÿ” Privacy-first

๐Ÿ‘ป Immortal (no sunsets)

Founder: masterknuta

<div align="center"> <br> <sub>Made with ๐Ÿ’œ and 40+ hours of dependency hell</sub> <br> <sub>โœจ The llama was sacrificed so your soul could live โœจ</sub> </div> ```

Downloads last month
1
GGUF
Model size
1B params
Architecture
llama
Hardware compatibility
Log In to add your hardware

4-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support