--- license: mit title: Fine-Tuned GPT2 For Azure DevOps Q&A sdk: gradio emoji: ⚡ colorFrom: red colorTo: yellow short_description: Bot trained on custom AZ DevOps dataset sdk_version: 5.41.0 --- ### NOTE: The large fine-tuned GPT-2 model weights are hosted on Hugging Face: https://huggingface.co/heramb04/GPT2-Azure-DevOps No manual download needed. The model is automatically pulled when `app.py` runs. If you want to download manually: bash wget https://huggingface.co/heramb04/GPT2-Azure-DevOps/resolve/main/model.safetensors ## About This project deploys a fine-tuned GPT-2 model for Azure DevOps Q&A as a web API using huggingFace and gradio. The model is loaded locally and provides a Q&A interface. ## Features - **Local Model Inference:** Uses a fine-tuned GPT-2 model loaded from local files. - **Gradio Interface:** A simple web API for text-based question and answer. - **Easy Deployment:** Run locally and test using a public link via Gradio. ### Prerequisites - Python 3.8+ - Git ## How to Run Locally 1. Clone this repository: git clone https://github.com/Heramb04/Fine_Tuned_GPT-2.git cd PRODIGY_GA_1 Navigate to the directory where you have cloned the repository.. 2. Set up a virtual environment: On windows: python -m venv venv venv\Scripts\activate On macOS/Linux: python3 -m venv venv source venv/bin/activate 3. Install the dependencies: pip install -r requirements.txt 4. Run the application: python app.py