victor's picture
victor HF Staff
formatting
b25ac89
# Running Locally
## Quick Start
1. Create a `.env.local` file with your API credentials:
```ini
OPENAI_BASE_URL=https://router.huggingface.co/v1
OPENAI_API_KEY=hf_************************
```
2. Install and run:
```bash
npm install
npm run dev -- --open
```
That's it! Chat UI will discover available models automatically from your endpoint.
## Configuration
Chat UI connects to any OpenAI-compatible API. Set `OPENAI_BASE_URL` to your provider:
| Provider | `OPENAI_BASE_URL` |
| ------------ | ---------------------------------- |
| Hugging Face | `https://router.huggingface.co/v1` |
| Ollama | `http://127.0.0.1:11434/v1` |
| llama.cpp | `http://127.0.0.1:8080/v1` |
| OpenRouter | `https://openrouter.ai/api/v1` |
See the [configuration overview](../configuration/overview) for all available options.
## Database
For **development**, MongoDB is optional. When `MONGODB_URL` is not set, Chat UI uses an embedded MongoDB server that persists data to the `./db` folder.
For **production**, you should use a dedicated MongoDB instance:
### Option 1: Local MongoDB (Docker)
```bash
docker run -d -p 27017:27017 -v mongo-chat-ui:/data --name mongo-chat-ui mongo:latest
```
Then set `MONGODB_URL=mongodb://localhost:27017` in `.env.local`.
### Option 2: MongoDB Atlas (Managed)
Use [MongoDB Atlas free tier](https://www.mongodb.com/pricing) for a managed database. Copy the connection string to `MONGODB_URL`.
## Running in Production
For production deployments:
```bash
npm install
npm run build
npm run preview
```
The server listens on `http://localhost:4173` by default.