|
--- |
|
base_model: |
|
- black-forest-labs/FLUX.1-dev |
|
library_name: diffusers |
|
license_name: flux-1-dev-non-commercial-license |
|
license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md |
|
pipeline_tag: image-to-image |
|
tags: |
|
- ControlNet |
|
license: cc-by-nc-4.0 |
|
--- |
|
# ⚡ Flux.1-dev: Depth ControlNet ⚡ |
|
|
|
This is [Flux.1-dev](https://huggingface.co/black-forest-labs/FLUX.1-dev) ControlNet for Depth map developped by Jasper research team. |
|
|
|
<p align="center"> |
|
<img style="width:700px;" src="examples/showcase.jpg"> |
|
</p> |
|
|
|
# How to use |
|
This model can be used directly with the `diffusers` library |
|
|
|
```python |
|
import torch |
|
from diffusers.utils import load_image |
|
from diffusers import FluxControlNetModel |
|
from diffusers.pipelines import FluxControlNetPipeline |
|
|
|
# Load pipeline |
|
controlnet = FluxControlNetModel.from_pretrained( |
|
"jasperai/Flux.1-dev-Controlnet-Depth", |
|
torch_dtype=torch.bfloat16 |
|
) |
|
pipe = FluxControlNetPipeline.from_pretrained( |
|
"black-forest-labs/FLUX.1-dev", |
|
controlnet=controlnet, |
|
torch_dtype=torch.bfloat16 |
|
) |
|
|
|
|
|
# Load a control image |
|
control_image = load_image( |
|
"https://huggingface.co/jasperai/Flux.1-dev-Controlnet-Depth/resolve/main/examples/depth.jpg" |
|
) |
|
|
|
prompt = "a statue of a gnome in a field of purple tulips" |
|
|
|
image = pipe( |
|
prompt, |
|
control_image=control_image, |
|
controlnet_conditioning_scale=0.6, |
|
num_inference_steps=28, |
|
guidance_scale=3.5, |
|
height=control_image.size[1], |
|
width=control_image.size[0] |
|
).images[0] |
|
image |
|
``` |
|
|
|
<p align="center"> |
|
<img style="width:500px;" src="examples/output.jpg"> |
|
</p> |
|
|
|
💡 Note: You can compute the conditioning map using for instance the `MidasDetector` from the `controlnet_aux` library |
|
|
|
```python |
|
from controlnet_aux import MidasDetector |
|
from diffusers.utils import load_image |
|
|
|
midas = MidasDetector.from_pretrained("lllyasviel/Annotators") |
|
|
|
# Load an image |
|
im = load_image( |
|
"https://huggingface.co/jasperai/jasperai/Flux.1-dev-Controlnet-Depth/resolve/main/examples/output.jpg" |
|
) |
|
|
|
surface = midas(im) |
|
``` |
|
|
|
# Training |
|
This model was trained with depth maps computed with [Clipdrop's depth estimator model](https://clipdrop.co/apis/docs/portrait-depth-estimation) as well as open-souce depth estimation models such as Midas or Leres. |
|
|
|
# Licence |
|
This model is released under the the Creative Commons BY-NC license. |
|
Note that if used with Flux.1-dev, the model under the Flux.1-dev model also applies to this model. |