--- title: Technical Description Assistant emoji: 📝 colorFrom: blue colorTo: purple sdk: gradio sdk_version: 4.29.0 app_file: app.py pinned: false license: mit short_description: rewrite rough project descriptions into technical summaries --- # Technical Description Assistant A zero-shot Hugging Face Space that uses LLM to help developers instantly rewrite rough project descriptions into clean, professional technical summaries. ## ⚠️ Required: Hugging Face Token This Space **requires** a Hugging Face access token to function. The application uses only LLM-based processing with no fallback mechanisms. 1. Go to your Hugging Face profile and create a new access token with **`read`** permissions: [**huggingface.co/settings/tokens**](https://huggingface.co/settings/tokens) 2. Go to your Space settings: [**huggingface.co/spaces/Snaseem2026/technical-description-assistant/settings**](https://huggingface.co/spaces/Snaseem2026/technical-description-assistant/settings) 3. Scroll down to **"Secrets and variables"** and click **"New secret"**. 4. For the **Name**, enter `HUGGING_FACE_HUB_TOKEN`. 5. For the **Value**, paste your `hf_...` access token. 6. Click **"Save secret"** and restart the Space from the settings menu. ## How It Works This application uses the Hugging Face OpenAI-compatible API (`https://router.huggingface.co/v1/chat/completions`) to call a pre-trained large language model (`meta-llama/Llama-3.2-3B-Instruct`). It requires no local model download and no rule-based fallbacks, relying entirely on the LLM for professional rewriting. 1. **Input**: A developer enters a casual or rough description of their project. 2. **Processing**: The description is sent to the LLM with a system prompt instructing it to act as an expert technical writer. The output is always concise and limited to a maximum of 3 sentences. 3. **Output**: The model returns a polished, professional, and technically accurate summary (never more than 3 sentences). ## Tech Stack - **UI**: Gradio - **API**: OpenAI-compatible Chat API (`requests` library) - **Language Model**: `meta-llama/Llama-3.2-3B-Instruct` (LLM-only, no rule-based fallback) - **Hosting**: Hugging Face Spaces ## Running Locally To run this project on your own machine: 1. **Clone the repository:** ```bash git clone https://huggingface.co/spaces/Snaseem2026/technical-description-assistant cd technical-description-assistant ``` 2. **Install dependencies:** ```bash pip install -r requirements.txt ``` (The app now requires the `requests` library for direct API calls.) 3. **Set your Hugging Face token (required):** The application requires a valid token to access the Inference API. ```bash export HUGGING_FACE_HUB_TOKEN="hf_YOUR_TOKEN_HERE" ``` 4. **Run the application:** ```bash python app.py ```