Spaces:
Paused
Paused
import gc | |
import os | |
import random | |
import numpy as np | |
import json | |
import torch | |
from PIL import Image, PngImagePlugin | |
from datetime import datetime | |
from dataclasses import dataclass | |
from typing import Callable, Dict, Optional, Tuple | |
from diffusers import ( | |
DDIMScheduler, | |
DPMSolverMultistepScheduler, | |
DPMSolverSinglestepScheduler, | |
EulerAncestralDiscreteScheduler, | |
EulerDiscreteScheduler, | |
) | |
MAX_SEED = np.iinfo(np.int32).max | |
class StyleConfig: | |
prompt: str | |
negative_prompt: str | |
def randomize_seed_fn(seed: int, randomize_seed: bool) -> int: | |
if randomize_seed: | |
seed = random.randint(0, MAX_SEED) | |
return seed | |
def seed_everything(seed: int) -> torch.Generator: | |
torch.manual_seed(seed) | |
torch.cuda.manual_seed_all(seed) | |
np.random.seed(seed) | |
generator = torch.Generator() | |
generator.manual_seed(seed) | |
return generator | |
def parse_aspect_ratio(aspect_ratio: str) -> Optional[Tuple[int, int]]: | |
if aspect_ratio == "Custom": | |
return None | |
width, height = aspect_ratio.split(" x ") | |
return int(width), int(height) | |
def aspect_ratio_handler( | |
aspect_ratio: str, custom_width: int, custom_height: int | |
) -> Tuple[int, int]: | |
if aspect_ratio == "Custom": | |
return custom_width, custom_height | |
else: | |
width, height = parse_aspect_ratio(aspect_ratio) | |
return width, height | |
def get_scheduler(scheduler_config: Dict, name: str) -> Optional[Callable]: | |
scheduler_factory_map = { | |
"DPM++ 2M Karras": lambda: DPMSolverMultistepScheduler.from_config( | |
scheduler_config, use_karras_sigmas=True | |
), | |
"DPM++ SDE Karras": lambda: DPMSolverSinglestepScheduler.from_config( | |
scheduler_config, use_karras_sigmas=True | |
), | |
"DPM++ 2M SDE Karras": lambda: DPMSolverMultistepScheduler.from_config( | |
scheduler_config, use_karras_sigmas=True, algorithm_type="sde-dpmsolver++" | |
), | |
"Euler": lambda: EulerDiscreteScheduler.from_config(scheduler_config), | |
"Euler a": lambda: EulerAncestralDiscreteScheduler.from_config( | |
scheduler_config | |
), | |
"DDIM": lambda: DDIMScheduler.from_config(scheduler_config), | |
} | |
return scheduler_factory_map.get(name, lambda: None)() | |
def free_memory() -> None: | |
torch.cuda.empty_cache() | |
gc.collect() | |
def preprocess_prompt( | |
style_dict, | |
style_name: str, | |
positive: str, | |
negative: str = "", | |
add_style: bool = True, | |
) -> Tuple[str, str]: | |
p, n = style_dict.get(style_name, style_dict["(None)"]) | |
if add_style and positive.strip(): | |
formatted_positive = p.format(prompt=positive) | |
else: | |
formatted_positive = positive | |
combined_negative = n | |
if negative.strip(): | |
if combined_negative: | |
combined_negative += ", " + negative | |
else: | |
combined_negative = negative | |
return formatted_positive, combined_negative | |
def common_upscale( | |
samples: torch.Tensor, | |
width: int, | |
height: int, | |
upscale_method: str, | |
) -> torch.Tensor: | |
return torch.nn.functional.interpolate( | |
samples, size=(height, width), mode=upscale_method | |
) | |
def upscale( | |
samples: torch.Tensor, upscale_method: str, scale_by: float | |
) -> torch.Tensor: | |
width = round(samples.shape[3] * scale_by) | |
height = round(samples.shape[2] * scale_by) | |
return common_upscale(samples, width, height, upscale_method) | |
def load_wildcard_files(wildcard_dir: str) -> Dict[str, str]: | |
wildcard_files = {} | |
for file in os.listdir(wildcard_dir): | |
if file.endswith(".txt"): | |
key = f"__{file.split('.')[0]}__" # Create a key like __character__ | |
wildcard_files[key] = os.path.join(wildcard_dir, file) | |
return wildcard_files | |
def get_random_line_from_file(file_path: str) -> str: | |
with open(file_path, "r") as file: | |
lines = file.readlines() | |
if not lines: | |
return "" | |
return random.choice(lines).strip() | |
def add_wildcard(prompt: str, wildcard_files: Dict[str, str]) -> str: | |
for key, file_path in wildcard_files.items(): | |
if key in prompt: | |
wildcard_line = get_random_line_from_file(file_path) | |
prompt = prompt.replace(key, wildcard_line) | |
return prompt | |
def preprocess_image_dimensions(width, height): | |
if width % 8 != 0: | |
width = width - (width % 8) | |
if height % 8 != 0: | |
height = height - (height % 8) | |
return width, height | |
def save_image(image, metadata, output_dir): | |
current_time = datetime.now().strftime("%Y%m%d_%H%M%S") | |
os.makedirs(output_dir, exist_ok=True) | |
filename = f"image_{current_time}.png" | |
filepath = os.path.join(output_dir, filename) | |
metadata_str = json.dumps(metadata) | |
info = PngImagePlugin.PngInfo() | |
info.add_text("metadata", metadata_str) | |
image.save(filepath, "PNG", pnginfo=info) | |
return filepath | |
def is_google_colab(): | |
try: | |
import google.colab | |
return True | |
except: | |
return False | |