Spaces:
Sleeping
HuggingFace Spaces Upload Checklist
β Pre-Upload Checklist
Your app is ready! Just upload these files:
Required Files (Check off as you upload)
-
app.pyβ MAIN FILE - HuggingFace Spaces needs this exact name -
llm.py -
extractors.py -
tagging.py -
chunking.py -
validation.py -
reporting.py -
dashboard.py -
production_logger.py -
quote_extractor.py -
requirements.txt
Total: 11 files
π« DO NOT Upload
- β
.envfile - β
test_*.pyfiles - β
*.logfiles - β
logs/folder - β
outputs/folder - β
__pycache__/folder
π― Upload Steps
1. Create Your Space
- Go to: https://huggingface.co/new-space
- Enter a name (e.g.,
transcriptor-ai) - Choose Gradio as SDK
- Select GPU hardware (T4 minimum) β οΈ IMPORTANT!
- Click "Create Space"
2. Upload Files
Method A: Drag & Drop
- Click "Files" tab in your Space
- Click "Upload files"
- Drag all 11 files from the checklist above
- Click "Commit"
Method B: Git Repository
- Create a new Git repo
- Copy the 11 files above
- Add
.gitignore(already created for you) - Push to repo
- Connect repo to Space in Settings
3. Configure Space (Optional)
Go to Settings β Variables and add (all optional):
| Variable | Value | Why |
|---|---|---|
DEBUG_MODE |
True |
See detailed logs |
LLM_TEMPERATURE |
0.7 |
Already the default |
You don't need to configure anything - it works out of the box!
β±οΈ What to Expect
First Startup
- Installing dependencies: 2-5 minutes
- Downloading Phi-3-mini model: 2-5 minutes
- Total: ~5-10 minutes
Watch the Logs tab - you'll see:
Installing dependencies...
β
Configuration loaded for HuggingFace Spaces
π TranscriptorAI Enterprise - LLM Backend: local
[Local Model] Loading microsoft/Phi-3-mini-4k-instruct...
Downloading model files...
[Local Model] β
Model loaded on cuda:0
Running on local URL: http://0.0.0.0:7860
Subsequent Startups
- Only 30-60 seconds (model is cached)
β Verify It's Working
1. Check Startup Logs
Look for these lines in the Logs tab:
β
Configuration loaded for HuggingFace Spaces
β
LLM Backend: local
β
Model loaded on cuda:0 β GPU confirmed!
β
Running on local URL
2. Test with Sample
- Click "Upload Files"
- Upload a DOCX transcript
- Select "HCP" as interviewee type
- Click "Analyze Transcripts"
- Wait 5-10 minutes for processing
Expected Result:
- Quality Score: 0.7-1.0 (not 0.00!)
- CSV and PDF downloads available
- Dashboard shows charts
π Common Issues
Issue: ModuleNotFoundError: No module named 'xyz'
Solution: Upload the missing xyz.py file
Issue: Very slow or hangs
Check: Did you select GPU hardware?
- Go to Settings
- Under Hardware, choose "GPU (T4)"
- Restart Space
Issue: Quality Score 0.00
Solution:
- Add Variable:
DEBUG_MODE=True - Check logs for error messages
- Look for "[Local Model] β Generated" to confirm it's working
Issue: Out of memory
Solution:
- Add Variable:
LOCAL_MODEL=TinyLlama/TinyLlama-1.1B-Chat-v1.0 - OR upgrade to larger GPU
π° Cost
Free Tier (CPU)
- β οΈ Very slow (10+ minutes per transcript)
- Not recommended
GPU (T4) - ~$0.60/hour
- β Recommended
- Fast processing (~5-10 min per transcript)
- Space sleeps after inactivity (saves money)
- Only charged when active
π Quick Reference
Space must have:
app.pyas main file β (already correct)requirements.txtwith dependencies β (already correct)- GPU hardware selected β οΈ (you must select this)
No .env file needed - everything configured in code β
No terminal commands needed - all automatic β
π Ready to Deploy!
- β Check you have all 11 files
- β Create Space with GPU hardware
- β Upload files via drag & drop
- β Wait for build (watch Logs tab)
- β Test with a transcript
See FILES_TO_UPLOAD.txt for the complete list of files.
π Still Stuck?
Common causes:
- Forgot to upload a file - Check all 11 files are uploaded
- Selected CPU instead of GPU - Change in Settings
- Uploaded .env file - Delete it, not needed on Spaces
Last Updated: October 2025
You're ready - just upload the 11 files and you're done! π