| import ipdb |
| import torch |
| from typing import Optional, Union, List, Tuple, Any |
|
|
| from torch import Tensor |
|
|
| from diffusers.pipelines import FluxPipeline |
| from PIL import Image, ImageFilter |
| import numpy as np |
| import cv2 |
|
|
| condition_dict = { |
| "depth": 0, |
| "canny": 1, |
| "subject": 4, |
| "coloring": 6, |
| "deblurring": 7, |
| "fill": 9, |
| } |
|
|
|
|
| class Condition(object): |
| def __init__( |
| self, |
| condition_type: str, |
| raw_img: Union[Image.Image, torch.Tensor] = None, |
| no_process: bool = False, |
| condition: Union[Image.Image, torch.Tensor] = None, |
| condition_ids = None, |
| mask=None, |
| ) -> None: |
| self.condition_type = condition_type |
| assert raw_img is not None or condition is not None |
| if raw_img is not None: |
| if no_process: |
| self.condition = raw_img.convert("RGB") |
| else: |
| self.condition = self.get_condition(condition_type, raw_img) |
| self.condition_ids = None |
| else: |
| self.condition = condition |
| self.condition_ids = condition_ids |
| |
| assert mask is None, "Mask not supported yet" |
|
|
| def get_condition( |
| self, condition_type: str, raw_img: Union[Image.Image, torch.Tensor] |
| ) -> Union[Image.Image, torch.Tensor]: |
| """ |
| Returns the condition image. |
| """ |
| if condition_type == "depth": |
| from transformers import pipeline |
|
|
| depth_pipe = pipeline( |
| task="depth-estimation", |
| model="ckpt/depth-anything-small-hf", |
| device="cuda", |
| ) |
| source_image = raw_img.convert("RGB") |
| condition_img = depth_pipe(source_image)["depth"].convert("RGB") |
| return condition_img |
| elif condition_type == "canny": |
| img = np.array(raw_img) |
| edges = cv2.Canny(img, 100, 200) |
| edges = Image.fromarray(edges).convert("RGB") |
| return edges |
| elif condition_type == "subject": |
| return raw_img |
| elif condition_type == "coloring": |
| return raw_img.convert("L").convert("RGB") |
| elif condition_type == "deblurring": |
| condition_image = ( |
| raw_img.convert("RGB") |
| .filter(ImageFilter.GaussianBlur(10)) |
| .convert("RGB") |
| ) |
| return condition_image |
| elif condition_type == "fill": |
| return raw_img.convert("RGB") |
| return self.condition |
|
|
| @property |
| def type_id(self) -> int: |
| """ |
| Returns the type id of the condition. |
| """ |
| return condition_dict[self.condition_type] |
|
|
| def _encode_image(self, pipe: FluxPipeline, cond_img: Image.Image) -> tuple[Any, Any]: |
| """ |
| Encodes an image condition into tokens using the pipeline. |
| """ |
| cond_img = pipe.image_processor.preprocess(cond_img) |
| cond_img = cond_img.to("cuda").to(pipe.dtype) |
| |
| cond_img = pipe.vae.encode(cond_img).latent_dist.sample() |
| cond_img = ( |
| cond_img - pipe.vae.config.shift_factor |
| ) * pipe.vae.config.scaling_factor |
| cond_tokens = pipe._pack_latents(cond_img, *cond_img.shape) |
| cond_ids = pipe._prepare_latent_image_ids( |
| cond_img.shape[0], |
| cond_img.shape[2]//2, |
| cond_img.shape[3]//2, |
| pipe.device, |
| pipe.dtype, |
| ) |
| if self.condition_type == "subject": |
| cond_ids[:, 2] += cond_img.shape[2] //2 |
| return cond_tokens, cond_ids |
|
|
| def encode(self, pipe: FluxPipeline) -> tuple[Any, Any, Tensor]: |
| """ |
| Encodes the condition into tokens, ids and type_id. |
| """ |
| if self.condition_ids is not None: |
| tokens, ids = self.condition, self.condition_ids |
| elif self.condition_ids is None and self.condition_type in [ |
| "depth", |
| "canny", |
| "subject", |
| "coloring", |
| "deblurring", |
| "fill", |
| ]: |
| tokens, ids = self._encode_image(pipe, self.condition) |
| else: |
| raise NotImplementedError( |
| f"There are two ways to use it: \n" |
| f"(1) Give the condition tensor to the 'self.condition' and the condition_ids to the 'self.condition_ids' manually.\n" |
| f"(2) Give the raw_image to the 'self.raw_img' and process the rest operations with a pipeline automatically.\n" |
| ) |
| type_id = torch.ones_like(ids[:, :1]) * self.type_id |
| return tokens, ids, type_id |
|
|