image_generator / README.md
Iratze's picture
Update README.md
6740fa2 verified
metadata
language: en
license: mit
library_name: diffusers
tags:
  - stable-diffusion
  - text-to-image
  - image-generation
inference: true
model-index:
  - name: Stable Diffusion Image Generator
    results:
      - task:
          name: Text-to-Image
          type: text-to-image
        dataset:
          name: LAION-400M
          type: image
        metrics:
          - name: Inception Score
            type: inception_score
            value: 9
          - name: FID
            type: fid
            value: 25

Stable Diffusion Image Generator with Inception Score

This repository uses the Stable Diffusion model from the diffusers library to generate images based on a text prompt and returns the generated image in base64 format.

How It Works

  1. The user sends a prompt (e.g., "A red apple on a wooden table").
  2. The Stable Diffusion model generates images based on the provided prompt.
  3. The first generated image is returned as a base64-encoded PNG image.

Model Used

  • Model: CompVis/stable-diffusion-v1-4
  • Library: diffusers
  • The model is pre-trained, and inference is run on a GPU (if available) or CPU.

How to Use the Inference API

You can use this model via the Hugging Face Inference API by making a POST request with the following format:

curl -X POST https://api-inference.huggingface.co/models/YOUR_USERNAME/stable-diffusion-make -H "Authorization: Bearer YOUR_API_TOKEN" -H "Content-Type: application/json" -d '{"prompt": "A red apple on a wooden table", "num_images": 1}'

Parameters:

  • prompt: The text prompt for image generation.
  • num_images: Number of images to generate (default is 1).

The response will return the first image encoded in base64 format.

License

MIT License