image_generator / README.md
Iratze's picture
Update README.md
6740fa2 verified
---
language: en
license: mit
library_name: diffusers
tags:
- stable-diffusion
- text-to-image
- image-generation
inference: true
model-index:
- name: Stable Diffusion Image Generator
results:
- task:
name: Text-to-Image
type: text-to-image
dataset:
name: LAION-400M
type: image
metrics:
- name: Inception Score
type: inception_score
value: 9.0
- name: FID
type: fid
value: 25.0
---
# Stable Diffusion Image Generator with Inception Score
This repository uses the `Stable Diffusion` model from the `diffusers` library to generate images based on a text prompt and returns the generated image in base64 format.
## How It Works
1. The user sends a prompt (e.g., "A red apple on a wooden table").
2. The `Stable Diffusion` model generates images based on the provided prompt.
3. The first generated image is returned as a base64-encoded PNG image.
## Model Used
- **Model**: `CompVis/stable-diffusion-v1-4`
- **Library**: [diffusers](https://huggingface.co/docs/diffusers)
- The model is pre-trained, and inference is run on a GPU (if available) or CPU.
## How to Use the Inference API
You can use this model via the Hugging Face Inference API by making a POST request with the following format:
```bash
curl -X POST https://api-inference.huggingface.co/models/YOUR_USERNAME/stable-diffusion-make -H "Authorization: Bearer YOUR_API_TOKEN" -H "Content-Type: application/json" -d '{"prompt": "A red apple on a wooden table", "num_images": 1}'
```
### Parameters:
- `prompt`: The text prompt for image generation.
- `num_images`: Number of images to generate (default is 1).
The response will return the first image encoded in base64 format.
## License
MIT License