ollama-api-proxy / README.md
GitHub Actions
Sync from GitHub
1d32142
---
title: Ollama API Proxy
emoji: 🦙
colorFrom: purple
colorTo: blue
sdk: docker
pinned: false
---
# Ollama API Proxy
A FastAPI-based proxy for the Ollama API hosted on Hugging Face Spaces.
## Endpoints
### GET /
Health check endpoint returning service status.
### GET /health
Simple health check endpoint.
### POST /chat
Send a chat message to the Ollama API.
**Request Body:**
```json
{
"message": "Your message here",
"model": "gpt-oss:120b",
"stream": true
}
```
**Response (non-streaming):**
```json
{
"response": "The AI response"
}
```
## Environment Variables
- `OLLAMA_API_KEY`: Your Ollama API key (set as a secret in HF Spaces)
## Setup
1. Create a new Space on Hugging Face with Docker SDK
2. Add `OLLAMA_API_KEY` as a repository secret
3. Push this code to the Space repository