Spaces:
Sleeping
Sleeping
| title: Krish Mind Chat | |
| emoji: 🧠 | |
| colorFrom: blue | |
| colorTo: blue | |
| sdk: docker | |
| app_file: app.py | |
| pinned: false | |
| # Deployment Instructions (Hugging Face Spaces) | |
| Your deployment package is ready! Follow these steps to put Krish Mind online for free. | |
| ## 1. No Manual Model Upload Needed! | |
| Great news! Since your model is already on Hugging Face (`Krishkanth/krish-mind-gguf-Q4`), the app will download it automatically when it starts. | |
| You **DO NOT** need to copy the 5GB model file into this folder. | |
| ## 2. Create Hugging Face Space | |
| 1. Login to [Hugging Face](https://huggingface.co). | |
| 2. Click **New Space**. | |
| 3. Name: `krish-mind-chat` (or similar). | |
| 4. SDK: **Docker** (Select "Blank" template). | |
| 5. Hardware: **CPU Basic (Free)** (2 vCPU, 16GB RAM). | |
| 6. Create Space. | |
| ## 3. Upload Files | |
| You have two options: | |
| ### Option A: Web Upload (Easiest) | |
| 1. Go to the **Files** tab of your new Space. | |
| 2. Click **Add file > Upload files**. | |
| 3. Drag and drop these files/folders from `d:\Krish Mind\deployment`: | |
| - `Dockerfile` | |
| - `requirements.txt` | |
| - `app.py` | |
| - `static/` folder | |
| - `data/` folder | |
| 4. Commit changes. | |
| ### Option B: Git (Recommended for Code) | |
| If you prefer the command line: | |
| 1. **Clone your empty Space:** | |
| ```bash | |
| git clone https://huggingface.co/spaces/YOUR_USERNAME/krish-mind-chat | |
| cd krish-mind-chat | |
| ``` | |
| 2. **Copy files:** | |
| - Copy all files from `d:\Krish Mind\deployment` (except the `model` folder!) into this new `krish-mind-chat` folder. | |
| 3. **Push:** | |
| ```bash | |
| git add . | |
| git commit -m "Deploy Krish Mind" | |
| git push | |
| ``` | |
| > **🔑 Authentication Required:** | |
| > When asked for credentials: | |
| > - **Username**: Your Hugging Face username (e.g., `Krishkanth`) | |
| > - **Password**: Your **Hugging Face Access Token** (Not your password!) | |
| > - Go to [Settings > Access Tokens](https://huggingface.co/settings/tokens) | |
| > - Create a new token with **Write** permissions. | |
| > - Copy and paste it as the password. | |
| ## FAQ: Why GGUF? Why not Safetensors (HF Fixed)? | |
| You asked about using your `hf_fixed` model. Here is why we **MUST use GGUF** for the Free Tier: | |
| - **RAM Limit:** Hugging Face Free Tier gives you **16GB RAM**. | |
| - **Safetensors (HF Fixed)**: A 7B model takes **~15GB+ RAM**. It will likely **crash** (Out of Memory) or be extremely slow. | |
| - **GGUF (Q4)**: Takes only **~5-6GB RAM**. It runs fast and stable on the Free Tier. | |
| - **Speed:** GGUF is optimized for CPU. Safetensors is built for GPU. On a free CPU Space, Safetensors would be unusably slow (like 1 word per 5 seconds). | |
| **Deploy with confidence!** The current setup downloads the GGUF model directly from your repo `Krishkanth/krish-mind` on startup. purely cloud-to-cloud. negligible upload time for you! | |