Spaces:
Runtime error
Runtime error
# Taken from https://github.com/comfyanonymous/ComfyUI | |
# This file is only for reference, and not used in the backend or runtime. | |
import ldm_patched.modules.samplers | |
import ldm_patched.modules.sample | |
from ldm_patched.k_diffusion import sampling as k_diffusion_sampling | |
import ldm_patched.utils.latent_visualization | |
import torch | |
import ldm_patched.modules.utils | |
class BasicScheduler: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"model": ("MODEL",), | |
"scheduler": (ldm_patched.modules.samplers.SCHEDULER_NAMES, ), | |
"steps": ("INT", {"default": 20, "min": 1, "max": 10000}), | |
"denoise": ("FLOAT", {"default": 1.0, "min": 0.0, "max": 1.0, "step": 0.01}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/schedulers" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, model, scheduler, steps, denoise): | |
total_steps = steps | |
if denoise < 1.0: | |
total_steps = int(steps/denoise) | |
ldm_patched.modules.model_management.load_models_gpu([model]) | |
sigmas = ldm_patched.modules.samplers.calculate_sigmas_scheduler(model.model, scheduler, total_steps).cpu() | |
sigmas = sigmas[-(steps + 1):] | |
return (sigmas, ) | |
class KarrasScheduler: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"steps": ("INT", {"default": 20, "min": 1, "max": 10000}), | |
"sigma_max": ("FLOAT", {"default": 14.614642, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
"sigma_min": ("FLOAT", {"default": 0.0291675, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
"rho": ("FLOAT", {"default": 7.0, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/schedulers" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, steps, sigma_max, sigma_min, rho): | |
sigmas = k_diffusion_sampling.get_sigmas_karras(n=steps, sigma_min=sigma_min, sigma_max=sigma_max, rho=rho) | |
return (sigmas, ) | |
class ExponentialScheduler: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"steps": ("INT", {"default": 20, "min": 1, "max": 10000}), | |
"sigma_max": ("FLOAT", {"default": 14.614642, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
"sigma_min": ("FLOAT", {"default": 0.0291675, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/schedulers" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, steps, sigma_max, sigma_min): | |
sigmas = k_diffusion_sampling.get_sigmas_exponential(n=steps, sigma_min=sigma_min, sigma_max=sigma_max) | |
return (sigmas, ) | |
class PolyexponentialScheduler: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"steps": ("INT", {"default": 20, "min": 1, "max": 10000}), | |
"sigma_max": ("FLOAT", {"default": 14.614642, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
"sigma_min": ("FLOAT", {"default": 0.0291675, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
"rho": ("FLOAT", {"default": 1.0, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/schedulers" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, steps, sigma_max, sigma_min, rho): | |
sigmas = k_diffusion_sampling.get_sigmas_polyexponential(n=steps, sigma_min=sigma_min, sigma_max=sigma_max, rho=rho) | |
return (sigmas, ) | |
class SDTurboScheduler: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"model": ("MODEL",), | |
"steps": ("INT", {"default": 1, "min": 1, "max": 10}), | |
"denoise": ("FLOAT", {"default": 1.0, "min": 0, "max": 1.0, "step": 0.01}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/schedulers" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, model, steps, denoise): | |
start_step = 10 - int(10 * denoise) | |
timesteps = torch.flip(torch.arange(1, 11) * 100 - 1, (0,))[start_step:start_step + steps] | |
ldm_patched.modules.model_management.load_models_gpu([model]) | |
sigmas = model.model.model_sampling.sigma(timesteps) | |
sigmas = torch.cat([sigmas, sigmas.new_zeros([1])]) | |
return (sigmas, ) | |
class VPScheduler: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"steps": ("INT", {"default": 20, "min": 1, "max": 10000}), | |
"beta_d": ("FLOAT", {"default": 19.9, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), #TODO: fix default values | |
"beta_min": ("FLOAT", {"default": 0.1, "min": 0.0, "max": 1000.0, "step":0.01, "round": False}), | |
"eps_s": ("FLOAT", {"default": 0.001, "min": 0.0, "max": 1.0, "step":0.0001, "round": False}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/schedulers" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, steps, beta_d, beta_min, eps_s): | |
sigmas = k_diffusion_sampling.get_sigmas_vp(n=steps, beta_d=beta_d, beta_min=beta_min, eps_s=eps_s) | |
return (sigmas, ) | |
class SplitSigmas: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"sigmas": ("SIGMAS", ), | |
"step": ("INT", {"default": 0, "min": 0, "max": 10000}), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS","SIGMAS") | |
CATEGORY = "sampling/custom_sampling/sigmas" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, sigmas, step): | |
sigmas1 = sigmas[:step + 1] | |
sigmas2 = sigmas[step:] | |
return (sigmas1, sigmas2) | |
class FlipSigmas: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"sigmas": ("SIGMAS", ), | |
} | |
} | |
RETURN_TYPES = ("SIGMAS",) | |
CATEGORY = "sampling/custom_sampling/sigmas" | |
FUNCTION = "get_sigmas" | |
def get_sigmas(self, sigmas): | |
sigmas = sigmas.flip(0) | |
if sigmas[0] == 0: | |
sigmas[0] = 0.0001 | |
return (sigmas,) | |
class KSamplerSelect: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"sampler_name": (ldm_patched.modules.samplers.SAMPLER_NAMES, ), | |
} | |
} | |
RETURN_TYPES = ("SAMPLER",) | |
CATEGORY = "sampling/custom_sampling/samplers" | |
FUNCTION = "get_sampler" | |
def get_sampler(self, sampler_name): | |
sampler = ldm_patched.modules.samplers.sampler_object(sampler_name) | |
return (sampler, ) | |
class SamplerDPMPP_2M_SDE: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"solver_type": (['midpoint', 'heun'], ), | |
"eta": ("FLOAT", {"default": 1.0, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
"s_noise": ("FLOAT", {"default": 1.0, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
"noise_device": (['gpu', 'cpu'], ), | |
} | |
} | |
RETURN_TYPES = ("SAMPLER",) | |
CATEGORY = "sampling/custom_sampling/samplers" | |
FUNCTION = "get_sampler" | |
def get_sampler(self, solver_type, eta, s_noise, noise_device): | |
if noise_device == 'cpu': | |
sampler_name = "dpmpp_2m_sde" | |
else: | |
sampler_name = "dpmpp_2m_sde_gpu" | |
sampler = ldm_patched.modules.samplers.ksampler(sampler_name, {"eta": eta, "s_noise": s_noise, "solver_type": solver_type}) | |
return (sampler, ) | |
class SamplerDPMPP_SDE: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"eta": ("FLOAT", {"default": 1.0, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
"s_noise": ("FLOAT", {"default": 1.0, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
"r": ("FLOAT", {"default": 0.5, "min": 0.0, "max": 100.0, "step":0.01, "round": False}), | |
"noise_device": (['gpu', 'cpu'], ), | |
} | |
} | |
RETURN_TYPES = ("SAMPLER",) | |
CATEGORY = "sampling/custom_sampling/samplers" | |
FUNCTION = "get_sampler" | |
def get_sampler(self, eta, s_noise, r, noise_device): | |
if noise_device == 'cpu': | |
sampler_name = "dpmpp_sde" | |
else: | |
sampler_name = "dpmpp_sde_gpu" | |
sampler = ldm_patched.modules.samplers.ksampler(sampler_name, {"eta": eta, "s_noise": s_noise, "r": r}) | |
return (sampler, ) | |
class SamplerCustom: | |
def INPUT_TYPES(s): | |
return {"required": | |
{"model": ("MODEL",), | |
"add_noise": ("BOOLEAN", {"default": True}), | |
"noise_seed": ("INT", {"default": 0, "min": 0, "max": 0xffffffffffffffff}), | |
"cfg": ("FLOAT", {"default": 8.0, "min": 0.0, "max": 100.0, "step":0.1, "round": 0.01}), | |
"positive": ("CONDITIONING", ), | |
"negative": ("CONDITIONING", ), | |
"sampler": ("SAMPLER", ), | |
"sigmas": ("SIGMAS", ), | |
"latent_image": ("LATENT", ), | |
} | |
} | |
RETURN_TYPES = ("LATENT","LATENT") | |
RETURN_NAMES = ("output", "denoised_output") | |
FUNCTION = "sample" | |
CATEGORY = "sampling/custom_sampling" | |
def sample(self, model, add_noise, noise_seed, cfg, positive, negative, sampler, sigmas, latent_image): | |
latent = latent_image | |
latent_image = latent["samples"] | |
if not add_noise: | |
noise = torch.zeros(latent_image.size(), dtype=latent_image.dtype, layout=latent_image.layout, device="cpu") | |
else: | |
batch_inds = latent["batch_index"] if "batch_index" in latent else None | |
noise = ldm_patched.modules.sample.prepare_noise(latent_image, noise_seed, batch_inds) | |
noise_mask = None | |
if "noise_mask" in latent: | |
noise_mask = latent["noise_mask"] | |
x0_output = {} | |
callback = ldm_patched.utils.latent_visualization.prepare_callback(model, sigmas.shape[-1] - 1, x0_output) | |
disable_pbar = not ldm_patched.modules.utils.PROGRESS_BAR_ENABLED | |
samples = ldm_patched.modules.sample.sample_custom(model, noise, cfg, sampler, sigmas, positive, negative, latent_image, noise_mask=noise_mask, callback=callback, disable_pbar=disable_pbar, seed=noise_seed) | |
out = latent.copy() | |
out["samples"] = samples | |
if "x0" in x0_output: | |
out_denoised = latent.copy() | |
out_denoised["samples"] = model.model.process_latent_out(x0_output["x0"].cpu()) | |
else: | |
out_denoised = out | |
return (out, out_denoised) | |
NODE_CLASS_MAPPINGS = { | |
"SamplerCustom": SamplerCustom, | |
"BasicScheduler": BasicScheduler, | |
"KarrasScheduler": KarrasScheduler, | |
"ExponentialScheduler": ExponentialScheduler, | |
"PolyexponentialScheduler": PolyexponentialScheduler, | |
"VPScheduler": VPScheduler, | |
"SDTurboScheduler": SDTurboScheduler, | |
"KSamplerSelect": KSamplerSelect, | |
"SamplerDPMPP_2M_SDE": SamplerDPMPP_2M_SDE, | |
"SamplerDPMPP_SDE": SamplerDPMPP_SDE, | |
"SplitSigmas": SplitSigmas, | |
"FlipSigmas": FlipSigmas, | |
} | |