Spaces:
Running
Running
A newer version of the Gradio SDK is available:
6.1.0
Quick Start Guide
π Deploy in 5 Minutes
Step 1: Complete the Inference (REQUIRED)
β οΈ The code has placeholders for actual video generation
See TODO.md for details on integrating the inference logic.
Step 2: Create HuggingFace Space
- Go to https://huggingface.co/new-space
- Fill in:
- Name:
infinitetalk(or your choice) - License:
apache-2.0 - SDK:
Gradio - Hardware:
ZeroGPU⨠(FREE tier available!)
- Name:
- Click Create Space
Step 3: Upload Files
Via Web UI (easiest):
- Click "Files" tab in your Space
- Drag and drop all files from this directory:
README.md app.py requirements.txt packages.txt .gitignore LICENSE.txt src/ (folder) wan/ (folder) utils/ (folder) assets/ (folder) examples/ (folder) - Click "Commit changes"
Via Git:
git clone https://huggingface.co/spaces/YOUR_USERNAME/YOUR_SPACE_NAME
cd YOUR_SPACE_NAME
cp -r /path/to/infinitetalk-hf-space/* .
git add .
git commit -m "Initial deployment"
git push
Step 4: Wait for Build
- Build time: 5-10 minutes
- Check "Logs" tab for progress
- Flash-attn compilation takes longest
Step 5: Test
- Space shows "Running" β
- First generation downloads models (2-3 min)
- Try image-to-video example
- Try video dubbing example
β‘ Quick Commands
# View directory structure
ls -la
# Check file sizes
du -sh *
# Count lines of code
find . -name "*.py" | xargs wc -l
# Test Python syntax
python -m py_compile app.py
# View logs (after deployment)
# Go to your Space β Logs tab
π― Common Issues & Fixes
Build Fails
- Check Logs tab for specific error
- Flash-attn timeout? Normal, wait 10-15 min
- Still failing? Try Dockerfile approach (see DEPLOYMENT.md)
Models Don't Download
- Check https://status.huggingface.co
- Verify model repo IDs in
utils/model_loader.py - Add HF_TOKEN in Space settings if needed
Out of Memory
- Use 480p instead of 720p
- Reduce steps to 30
- Process shorter videos (<10s)
Space Stuck
- Refresh page
- Check if in queue (ZeroGPU)
- Wait for quota to refill
π Files Overview
| File/Folder | Purpose | Lines | Critical? |
|---|---|---|---|
README.md |
Space metadata | ~50 | β Yes |
app.py |
Main application | ~350 | β Yes |
requirements.txt |
Python packages | ~30 | β Yes |
packages.txt |
System packages | ~4 | β Yes |
utils/model_loader.py |
Model management | ~200 | β Yes |
utils/gpu_manager.py |
Memory management | ~150 | β Yes |
src/ |
Audio analysis | - | β Yes |
wan/ |
Model code | - | β Yes |
assets/ |
UI assets | - | Optional |
examples/ |
Sample data | - | Optional |
π§ Pre-Deployment Checklist
- All files present
- README.md has YAML metadata
- requirements.txt is properly ordered
- ZeroGPU hardware configured
- Inference logic integrated β¬ οΈ CRITICAL
- Tested locally (if possible)
- Examples prepared
π° Cost Breakdown
Free Tier
- Cost: $0
- GPU: H200 (70GB VRAM)
- Quota: 300s per session, 600s max
- Usage: ~3-5 generations per quota
- Best for: Testing, demos, light use
PRO Tier
- Cost: $9/month
- GPU: Same H200
- Quota: 8Γ more (1500s)
- Spaces: Up to 10
- Best for: Regular use, public demos
π Performance Expectations
| Task | Resolution | Time | VRAM |
|---|---|---|---|
| Model download | - | 2-3 min | - |
| 10s video | 480p | ~40s | ~38GB |
| 10s video | 720p | ~70s | ~55GB |
| 30s video | 480p | ~90s | ~45GB |
π Learning Resources
β Success Checklist
After deployment:
- Space builds successfully
- No errors in Logs
- UI loads properly
- Models download on first run
- Image-to-video works
- Video dubbing works
- No OOM errors
- Memory cleanup works
- Can run multiple generations
- Results look good!
π Need Help?
- Check TODO.md for implementation details
- Read DEPLOYMENT.md for troubleshooting
- Review PROJECT_SUMMARY.md for overview
- Ask on HuggingFace Forums: https://discuss.huggingface.co
- File issue on InfiniteTalk GitHub: https://github.com/MeiGen-AI/InfiniteTalk
Ready? Complete the inference integration, then deploy! π