metadata
language: en
license: mit
library_name: diffusers
tags:
- stable-diffusion
- text-to-image
- image-generation
inference: true
model-index:
- name: Stable Diffusion Image Generator
results:
- task:
name: Text-to-Image
type: text-to-image
dataset:
name: LAION-400M
type: image
metrics:
- name: Inception Score
type: inception_score
value: 9
- name: FID
type: fid
value: 25
Stable Diffusion Image Generator with Inception Score
This repository uses the Stable Diffusion
model from the diffusers
library to generate images based on a text prompt and returns the generated image in base64 format.
How It Works
- The user sends a prompt (e.g., "A red apple on a wooden table").
- The
Stable Diffusion
model generates images based on the provided prompt. - The first generated image is returned as a base64-encoded PNG image.
Model Used
- Model:
CompVis/stable-diffusion-v1-4
- Library: diffusers
- The model is pre-trained, and inference is run on a GPU (if available) or CPU.
How to Use the Inference API
You can use this model via the Hugging Face Inference API by making a POST request with the following format:
curl -X POST https://api-inference.huggingface.co/models/YOUR_USERNAME/stable-diffusion-make -H "Authorization: Bearer YOUR_API_TOKEN" -H "Content-Type: application/json" -d '{"prompt": "A red apple on a wooden table", "num_images": 1}'
Parameters:
prompt
: The text prompt for image generation.num_images
: Number of images to generate (default is 1).
The response will return the first image encoded in base64 format.
License
MIT License