| from typing import Tuple, Dict |
| import requests |
| import random |
| import numpy as np |
| import gradio as gr |
| import torch |
| from PIL import Image |
| from diffusers import FluxInpaintPipeline |
|
|
| |
| MARKDOWN_TEXT = """ |
| # FLUX.1 Inpainting 🔥 |
| Shoutout to [Black Forest Labs](https://huggingface.co/black-forest-labs) for |
| creating this amazing model, and a big thanks to [Gothos](https://github.com/Gothos) |
| for taking it to the next level by enabling inpainting with the FLUX. |
| """ |
|
|
| MAX_SEED_VALUE = np.iinfo(np.int32).max |
| DEFAULT_IMAGE_SIZE = 1024 |
| DEVICE_TYPE = "cuda" if torch.cuda.is_available() else "cpu" |
|
|
| |
| pipeline = FluxInpaintPipeline.from_pretrained( |
| "black-forest-labs/FLUX.1-schnell", torch_dtype=torch.bfloat16).to(DEVICE_TYPE) |
|
|
| def adjust_image_size( |
| original_size: Tuple[int, int], max_dimension: int = DEFAULT_IMAGE_SIZE |
| ) -> Tuple[int, int]: |
| width, height = original_size |
| scaling_factor = max_dimension / max(width, height) |
| new_width = int(width * scaling_factor) - (int(width * scaling_factor) % 32) |
| new_height = int(height * scaling_factor) - (int(height * scaling_factor) % 32) |
| return new_width, new_height |
|
|
| def process_images( |
| input_data: Dict, |
| prompt: str, |
| seed: int, |
| randomize_seed: bool, |
| strength: float, |
| num_steps: int, |
| progress=gr.Progress(track_tqdm=True) |
| ): |
| if not prompt: |
| gr.Info("Please enter a text prompt.") |
| return None, None |
|
|
| background_img = input_data['background'] |
| mask_img = input_data['layers'][0] |
|
|
| if background_img is None: |
| gr.Info("Please upload an image.") |
| return None, None |
|
|
| if mask_img is None: |
| gr.Info("Please draw a mask on the image.") |
| return None, None |
|
|
| new_width, new_height = adjust_image_size(background_img.size) |
| resized_bg = background_img.resize((new_width, new_height), Image.LANCZOS) |
| resized_mask = mask_img.resize((new_width, new_height), Image.LANCZOS) |
|
|
| if randomize_seed: |
| seed = random.randint(0, MAX_SEED_VALUE) |
| generator = torch.Generator().manual_seed(seed) |
|
|
| result_image = pipeline( |
| prompt=prompt, |
| image=resized_bg, |
| mask_image=resized_mask, |
| width=new_width, |
| height=new_height, |
| strength=strength, |
| generator=generator, |
| num_inference_steps=num_steps |
| ).images[0] |
| |
| return result_image, resized_mask |
|
|
| |
| with gr.Blocks() as demo: |
| gr.Markdown(MARKDOWN_TEXT) |
| |
| with gr.Row(): |
| with gr.Column(): |
| img_editor = gr.ImageEditor( |
| label='Image', |
| type='pil', |
| sources=["upload", "webcam"], |
| image_mode='RGB', |
| layers=False, |
| brush=gr.Brush(colors=["#FFFFFF"], color_mode="fixed") |
| ) |
|
|
| with gr.Row(): |
| text_input = gr.Text( |
| label="Prompt", |
| show_label=False, |
| max_lines=1, |
| placeholder="Enter your prompt", |
| container=False |
| ) |
| submit_btn = gr.Button( |
| value='Submit', variant='primary', scale=0 |
| ) |
|
|
| with gr.Accordion("Advanced Settings", open=False): |
| seed_slider = gr.Slider( |
| label="Seed", |
| minimum=0, |
| maximum=MAX_SEED_VALUE, |
| step=1, |
| value=42 |
| ) |
| random_seed_chkbox = gr.Checkbox( |
| label="Randomize seed", value=True |
| ) |
|
|
| with gr.Row(): |
| strength_slider = gr.Slider( |
| label="Strength", |
| info="Indicates extent to transform the reference `image`.", |
| minimum=0, |
| maximum=1, |
| step=0.01, |
| value=0.85 |
| ) |
| steps_slider = gr.Slider( |
| label="Number of inference steps", |
| info="The number of denoising steps.", |
| minimum=1, |
| maximum=50, |
| step=1, |
| value=20 |
| ) |
| |
| with gr.Column(): |
| output_img = gr.Image( |
| type='pil', image_mode='RGB', label='Generated Image', format="png" |
| ) |
| with gr.Accordion("Debug", open=False): |
| output_mask = gr.Image( |
| type='pil', image_mode='RGB', label='Input Mask', format="png" |
| ) |
|
|
| gr.Examples( |
| fn=process_images, |
| examples=[ |
| [ |
| { |
| "background": Image.open(requests.get("https://media.roboflow.com/spaces/doge-2-image.png", stream=True).raw), |
| "layers": [Image.open(requests.get("https://media.roboflow.com/spaces/doge-2-mask-2.png", stream=True).raw).convert("RGBA")], |
| "composite": Image.open(requests.get("https://media.roboflow.com/spaces/doge-2-composite-2.png", stream=True).raw), |
| }, |
| "little lion", |
| 42, |
| False, |
| 0.85, |
| 30 |
| ], |
| [ |
| { |
| "background": Image.open(requests.get("https://media.roboflow.com/spaces/doge-2-image.png", stream=True).raw), |
| "layers": [Image.open(requests.get("https://media.roboflow.com/spaces/doge-2-mask-3.png", stream=True).raw).convert("RGBA")], |
| "composite": Image.open(requests.get("https://media.roboflow.com/spaces/doge-2-composite-3.png", stream=True).raw), |
| }, |
| "tribal tattoos", |
| 42, |
| False, |
| 0.85, |
| 30 |
| ] |
| ], |
| inputs=[ |
| img_editor, |
| text_input, |
| seed_slider, |
| random_seed_chkbox, |
| strength_slider, |
| steps_slider |
| ], |
| outputs=[ |
| output_img, |
| output_mask |
| ], |
| run_on_click=True, |
| cache_examples=True |
| ) |
|
|
| submit_btn.click( |
| fn=process_images, |
| inputs=[ |
| img_editor, |
| text_input, |
| seed_slider, |
| random_seed_chkbox, |
| strength_slider, |
| steps_slider |
| ], |
| outputs=[ |
| output_img, |
| output_mask |
| ] |
| ) |
|
|
| demo.launch(debug=False, show_error=True) |
|
|
|
|