krish-mind-chat / README.md
Krishkanth's picture
Deploy Krish Mind with all updates
f8b4220
---
title: Krish Mind Chat
emoji: 🧠
colorFrom: blue
colorTo: blue
sdk: docker
app_file: app.py
pinned: false
---
# Deployment Instructions (Hugging Face Spaces)
Your deployment package is ready! Follow these steps to put Krish Mind online for free.
## 1. No Manual Model Upload Needed!
Great news! Since your model is already on Hugging Face (`Krishkanth/krish-mind-gguf-Q4`), the app will download it automatically when it starts.
You **DO NOT** need to copy the 5GB model file into this folder.
## 2. Create Hugging Face Space
1. Login to [Hugging Face](https://huggingface.co).
2. Click **New Space**.
3. Name: `krish-mind-chat` (or similar).
4. SDK: **Docker** (Select "Blank" template).
5. Hardware: **CPU Basic (Free)** (2 vCPU, 16GB RAM).
6. Create Space.
## 3. Upload Files
You have two options:
### Option A: Web Upload (Easiest)
1. Go to the **Files** tab of your new Space.
2. Click **Add file > Upload files**.
3. Drag and drop these files/folders from `d:\Krish Mind\deployment`:
- `Dockerfile`
- `requirements.txt`
- `app.py`
- `static/` folder
- `data/` folder
4. Commit changes.
### Option B: Git (Recommended for Code)
If you prefer the command line:
1. **Clone your empty Space:**
```bash
git clone https://huggingface.co/spaces/YOUR_USERNAME/krish-mind-chat
cd krish-mind-chat
```
2. **Copy files:**
- Copy all files from `d:\Krish Mind\deployment` (except the `model` folder!) into this new `krish-mind-chat` folder.
3. **Push:**
```bash
git add .
git commit -m "Deploy Krish Mind"
git push
```
> **🔑 Authentication Required:**
> When asked for credentials:
> - **Username**: Your Hugging Face username (e.g., `Krishkanth`)
> - **Password**: Your **Hugging Face Access Token** (Not your password!)
> - Go to [Settings > Access Tokens](https://huggingface.co/settings/tokens)
> - Create a new token with **Write** permissions.
> - Copy and paste it as the password.
## FAQ: Why GGUF? Why not Safetensors (HF Fixed)?
You asked about using your `hf_fixed` model. Here is why we **MUST use GGUF** for the Free Tier:
- **RAM Limit:** Hugging Face Free Tier gives you **16GB RAM**.
- **Safetensors (HF Fixed)**: A 7B model takes **~15GB+ RAM**. It will likely **crash** (Out of Memory) or be extremely slow.
- **GGUF (Q4)**: Takes only **~5-6GB RAM**. It runs fast and stable on the Free Tier.
- **Speed:** GGUF is optimized for CPU. Safetensors is built for GPU. On a free CPU Space, Safetensors would be unusably slow (like 1 word per 5 seconds).
**Deploy with confidence!** The current setup downloads the GGUF model directly from your repo `Krishkanth/krish-mind` on startup. purely cloud-to-cloud. negligible upload time for you!