ollama-server / README.md
ScottzillaSystems's picture
Update README with new model
71428e2 verified
---
title: Ollama Server
emoji: 🦙
colorFrom: blue
colorTo: green
sdk: docker
app_port: 11434
pinned: true
---
# Ollama Server on Hugging Face Spaces
A lightweight [Ollama](https://ollama.com/) inference server running on a Hugging Face Space.
## Pre-pulled Models
- `gemma2:2b`
- `alibilge/Huihui-GLM-4.6V-Flash-abliterated:q4_k_s`
- `frob/mradermacher-Llama3.3-8B-Thinking-Heretic-Claude-4.5-Opus:q8_0`
## Usage
Once the Space is running, you can interact with the Ollama API at the Space's URL.
### Pull additional models
```bash
curl https://your-space.hf.space/api/pull -d '{"model": "llama3.2"}'
```
### Generate a completion
```bash
curl https://your-space.hf.space/api/generate -d '{
"model": "frob/mradermacher-Llama3.3-8B-Thinking-Heretic-Claude-4.5-Opus:q8_0",
"prompt": "Hello!"
}'
```
## API
This Space exposes the standard Ollama REST API on port `11434`.