| ---
|
| title: Mistral Fine-tuned Model
|
| emoji: π€
|
| colorFrom: blue
|
| colorTo: purple
|
| sdk: docker
|
| app_port: 7860
|
| ---
|
|
|
| # π€ Mistral Fine-tuned Model
|
|
|
| Flask API with separate HTML/CSS/JS frontend for `KASHH-4/mistral_fine-tuned` model.
|
|
|
| ## π What This Is
|
|
|
| A **Flask API server** with **separate frontend files**:
|
| - Backend: Python Flask with CORS
|
| - Frontend: HTML + CSS + JavaScript
|
| - Clean separation of concerns
|
| - API-first design
|
|
|
| ## π Project Structure
|
|
|
| ```
|
| e:\EDI\hf-node-app\
|
| βββ app.py # Main Gradio application
|
| βββ requirements.txt # Python dependencies
|
| βββ README.md # This file
|
| βββ .gitignore # Git ignore rules
|
| ```
|
|
|
| ## π§ Deploy to Hugging Face Spaces
|
|
|
| ### Step 1: Create a Space
|
|
|
| 1. Go to https://huggingface.co/spaces
|
| 2. Click **"Create new Space"**
|
| 3. Configure:
|
| - **Owner:** KASHH-4 (or your account)
|
| - **Space name:** `mistral-api` (or any name)
|
| - **SDK:** Gradio
|
| - **Hardware:** CPU basic (Free)
|
| - **Visibility:** Public
|
| 4. Click **"Create Space"**
|
|
|
| ### Step 2: Upload Files
|
|
|
| Upload these 3 files to your Space:
|
| - `app.py`
|
| - `requirements.txt`
|
| - `README.md` (optional)
|
|
|
| **Via Web UI:**
|
| 1. Click "Files" tab
|
| 2. Click "Add file" β "Upload files"
|
| 3. Drag and drop the files
|
| 4. Commit changes
|
|
|
| **Via Git:**
|
| ```bash
|
| git init
|
| git remote add origin https://huggingface.co/spaces/KASHH-4/mistral-api
|
| git add app.py requirements.txt README.md .gitignore
|
| git commit -m "Initial deployment"
|
| git push origin main
|
| ```
|
|
|
| ### Step 3: Wait for Deployment
|
|
|
| - First build takes 5-10 minutes
|
| - Watch the logs for "Running on..."
|
| - Your Space will be live at: `https://kashh-4-mistral-api.hf.space`
|
|
|
| ## π§ͺ Test Your Space
|
|
|
| ### Web Interface
|
| Visit: `https://huggingface.co/spaces/KASHH-4/mistral-api`
|
|
|
| ### API Endpoint
|
| ```bash
|
| curl -X POST "https://kashh-4-mistral-api.hf.space/api/predict" \
|
| -H "Content-Type: application/json" \
|
| -d '{"data":["Hello, how are you?"]}'
|
| ```
|
|
|
| ### From JavaScript/Node.js
|
| ```javascript
|
| const response = await fetch('https://kashh-4-mistral-api.hf.space/api/predict', {
|
| method: 'POST',
|
| headers: { 'Content-Type': 'application/json' },
|
| body: JSON.stringify({ data: ["Your prompt here"] })
|
| });
|
|
|
| const result = await response.json();
|
| console.log(result.data[0]); // Generated text
|
| ```
|
|
|
| ### From Python
|
| ```python
|
| import requests
|
|
|
| response = requests.post(
|
| 'https://kashh-4-mistral-api.hf.space/api/predict',
|
| json={'data': ['Your prompt here']}
|
| )
|
|
|
| print(response.json()['data'][0])
|
| ```
|
|
|
| ## π° Cost
|
|
|
| **100% FREE** on HF Spaces:
|
| - Free CPU tier (slower, ~10-30 sec per request)
|
| - Sleeps after 48h inactivity (30 sec wake-up)
|
| - Perfect for demos, personal projects, testing
|
|
|
| **Optional Upgrades:**
|
| - GPU T4 Small: $0.60/hour (much faster, 2-5 sec)
|
| - GPU A10G: $3.15/hour (very fast, 1-2 sec)
|
|
|
| Upgrade in: Space Settings β Hardware
|
|
|
| ## π§ Local Testing (Optional)
|
|
|
| If you have Python installed and want to test locally before deploying:
|
|
|
| ```bash
|
| # Install dependencies
|
| pip install -r requirements.txt
|
|
|
| # Run locally
|
| python app.py
|
|
|
| # Visit: http://localhost:7860
|
| ```
|
|
|
| **Requirements:**
|
| - Python 3.9+
|
| - 16GB+ RAM (for model loading)
|
| - GPU recommended but not required
|
|
|
| ## π Model Configuration
|
|
|
| The app is configured for `KASHH-4/mistral_fine-tuned`. To use a different model, edit `app.py`:
|
|
|
| ```python
|
| MODEL_NAME = "your-org/your-model"
|
| ```
|
|
|
| ## π Troubleshooting
|
|
|
| **Space stuck on "Building":**
|
| - Check logs for errors
|
| - Model might be too large for free CPU
|
| - Try: Restart Space in Settings
|
|
|
| **Space shows "Runtime Error":**
|
| - Check if model exists and is public
|
| - Verify model format is compatible with transformers
|
| - Try smaller model first to test
|
|
|
| **Slow responses:**
|
| - Normal on free CPU tier
|
| - Upgrade to GPU for faster inference
|
| - Or use smaller model
|
|
|
| ## π Support
|
|
|
| Issues? Check the deployment guide in `huggingface-space/DEPLOYMENT-GUIDE.md`
|
|
|
| ---
|
|
|
| ## ποΈ Cleanup Old Files
|
|
|
| If you followed earlier Node.js instructions, delete unnecessary files:
|
|
|
| See `CLEANUP.md` for full list of files to remove.
|
|
|
| ## License
|
| MIT
|
|
|