Spaces:
Sleeping
Sleeping
| title: testingtodeploy | |
| emoji: 🚀 | |
| colorFrom: blue | |
| colorTo: green | |
| sdk: docker | |
| pinned: false | |
| license: mit | |
| # Huggingface Spaces for Docker with FastAPI | |
| ## Overview | |
| This repository contains a simple example of how to deploy a Huggingface model using Docker and FastAPI. The model used is the `google-flan-t5-base` model from the Huggingface model hub. | |
| ## Usage | |
| To run the FastAPI server, you can use the following command: | |
| ```bash | |
| docker-compose up | |
| ``` | |
| then visit `http://localhost:8000/docs` to see the API documentation. | |
| ## Useful Links | |
| - [Google Flan T5 Base Model](https://huggingface.co/google/flan-t5-base) | |
| - [Files](https://huggingface.co/spaces/sarthaksavvy/google-flan-fastapi/tree/main)] |