Text-to-Image
Diffusers
flux
flux-diffusers
simpletuner
Not-For-All-Audiences
lora
template:sd-lora
standard
license: other | |
base_model: "black-forest-labs/FLUX.1-dev" | |
tags: | |
- flux | |
- flux-diffusers | |
- text-to-image | |
- diffusers | |
- simpletuner | |
- not-for-all-audiences | |
- lora | |
- template:sd-lora | |
- standard | |
inference: true | |
widget: | |
- text: 'unconditional (blank prompt)' | |
parameters: | |
negative_prompt: '''' | |
output: | |
url: ./assets/image_0_0.png | |
- text: 'm1nchen, a girl taking a selfie' | |
parameters: | |
negative_prompt: '''' | |
output: | |
url: ./assets/image_1_0.png | |
# clayygodd/flux-lora-min | |
This is a standard PEFT LoRA derived from [black-forest-labs/FLUX.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev). | |
The main validation prompt used during training was: | |
``` | |
m1nchen, a girl taking a selfie | |
``` | |
## Validation settings | |
- CFG: `3.5` | |
- CFG Rescale: `0.0` | |
- Steps: `15` | |
- Sampler: `None` | |
- Seed: `42` | |
- Resolution: `1024x1024` | |
Note: The validation settings are not necessarily the same as the [training settings](#training-settings). | |
You can find some example images in the following gallery: | |
<Gallery /> | |
The text encoder **was not** trained. | |
You may reuse the base model text encoder for inference. | |
## Training settings | |
- Training epochs: 18 | |
- Training steps: 200 | |
- Learning rate: 0.0001 | |
- Effective batch size: 2 | |
- Micro-batch size: 2 | |
- Gradient accumulation steps: 1 | |
- Number of GPUs: 1 | |
- Prediction type: flow-matching | |
- Rescaled betas zero SNR: False | |
- Optimizer: adamw_bf16 | |
- Precision: Pure BF16 | |
- Quantised: No | |
- Xformers: Not used | |
- LoRA Rank: 16 | |
- LoRA Alpha: None | |
- LoRA Dropout: 0.1 | |
- LoRA initialisation style: default | |
## Datasets | |
### default_dataset | |
- Repeats: 0 | |
- Total number of images: 22 | |
- Total number of aspect buckets: 1 | |
- Resolution: 1.048576 megapixels | |
- Cropped: True | |
- Crop style: center | |
- Crop aspect: square | |
## Inference | |
```python | |
import torch | |
from diffusers import DiffusionPipeline | |
model_id = 'black-forest-labs/FLUX.1-dev' | |
adapter_id = 'clayygodd/clayygodd/flux-lora-min' | |
pipeline = DiffusionPipeline.from_pretrained(model_id) | |
pipeline.load_lora_weights(adapter_id) | |
prompt = "m1nchen, a girl taking a selfie" | |
pipeline.to('cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu') | |
image = pipeline( | |
prompt=prompt, | |
num_inference_steps=15, | |
generator=torch.Generator(device='cuda' if torch.cuda.is_available() else 'mps' if torch.backends.mps.is_available() else 'cpu').manual_seed(1641421826), | |
width=1024, | |
height=1024, | |
guidance_scale=3.5, | |
).images[0] | |
image.save("output.png", format="PNG") | |
``` | |