| | --- |
| | title: 0.6wen |
| | emoji: ๐ |
| | colorFrom: gray |
| | colorTo: gray |
| | sdk: gradio |
| | sdk_version: 5.30.0 |
| | app_file: app.py |
| | pinned: false |
| | --- |
| | |
| | # Qwen3-0.6B Text-to-Text Chat Interface |
| |
|
| | A Gradio-powered web interface for interacting with the Qwen3-0.6B language model from HuggingFace. |
| |
|
| | ## Features |
| | - Interactive chat interface |
| | - Adjustable generation parameters |
| | - Input validation and error handling |
| | - Mobile-responsive design |
| | - Example prompts for inspiration |
| |
|
| | ## Requirements |
| | - Python 3.8+ |
| | - CUDA-compatible GPU recommended but not required |
| | - At least 8GB RAM (16GB recommended) |
| |
|
| | ## Installation |
| | 1. Install dependencies: `pip install -r requirements.txt` |
| | 2. Run the app: `python app.py` |
| |
|
| | ## Deployment on HuggingFace Spaces |
| | 1. Create new Space with Gradio SDK |
| | 2. Upload all files |
| | 3. Add `/model_cache` to `.gitignore` |
| | 4. Enable GPU hardware in Space settings |
| |
|
| | ## Tips |
| | - For production use, consider adding rate limiting |
| | - Can be extended to support multi-turn conversations |
| | - Add system prompt functionality for role-playing |