Logo_test / app.py
fantaxy's picture
Update app.py
797e81e verified
raw
history blame
3.84 kB
import gradio as gr
import numpy as np
import random
import spaces
import torch
from diffusers import DiffusionPipeline
from transformers import pipeline
# λ²ˆμ—­ νŒŒμ΄ν”„λΌμΈ 및 ν•˜λ“œμ›¨μ–΄ μ„€μ •
device = "cuda" if torch.cuda.is_available() else "cpu"
translator = pipeline("translation", model="Helsinki-NLP/opus-mt-ko-en", device=device)
dtype = torch.bfloat16
pipe = DiffusionPipeline.from_pretrained("black-forest-labs/FLUX.1-schnell", torch_dtype=dtype).to(device)
MAX_SEED = np.iinfo(np.int32).max
MAX_IMAGE_SIZE = 2048
@spaces.GPU()
def infer(prompt, seed=42, randomize_seed=False, width=1024, height=1024, num_inference_steps=4, progress=gr.Progress(track_tqdm=True)):
if randomize_seed:
seed = random.randint(0, MAX_SEED)
generator = torch.Generator().manual_seed(seed)
# ν•œκΈ€ μž…λ ₯ 감지 및 λ²ˆμ—­
if any('\uAC00' <= char <= '\uD7A3' for char in prompt):
print("ν•œκ΅­μ–΄ ν”„λ‘¬ν”„νŠΈ λ²ˆμ—­ 쀑...")
translated_prompt = translator(prompt, max_length=512)[0]['translation_text']
print("λ²ˆμ—­λœ ν”„λ‘¬ν”„νŠΈ:", translated_prompt)
prompt = translated_prompt
image = pipe(
prompt = prompt,
width = width,
height = height,
num_inference_steps = num_inference_steps,
generator = generator,
guidance_scale=0.0
).images[0]
return image, seed
# (이전 import ꡬ문 및 νŒŒμ΄ν”„λΌμΈ μ„€μ • μœ μ§€)
examples = [
["[ν•œκΈ€] [μŠ€νƒ€μΌ: λͺ¨λ˜] [색상: λΉ¨κ°•κ³Ό κ²€μ •] [컨셉: 식당] [ν…μŠ€νŠΈ: 'λ§›μžˆλŠ”μ§‘'] [λ°°κ²½: μ‹¬ν”Œ]"],
["[Style: Corporate] [Color: Navy and Silver] [Concept: Finance] [Text: 'TRUST'] [Background: Professional]"],
["[Style: Dynamic] [Color: Purple and Orange] [Concept: Creative Agency] [Text: 'SPARK'] [Background: Abstract]"],
["[Style: Minimalist] [Color: Red and White] [Concept: Sports] [Text: 'POWER'] [Background: Clean]"]
]
css = """
footer {visibility: hidden}
.container {max-width: 850px; margin: auto; padding: 20px}
.title {text-align: center; margin-bottom: 20px}
#prompt {min-height: 50px}
#result {min-height: 400px}
.gr-box {border-radius: 10px; border: 1px solid #ddd}
"""
with gr.Blocks(theme="Yntec/HaleyCH_Theme_Orange", css=css) as demo:
gr.HTML("<h1 class='title'>LOGO Generator AI</h1>")
with gr.Column(elem_id="container"):
with gr.Group():
prompt = gr.Text(
label="둜고 μ„€λͺ…",
placeholder="둜고 λ””μžμΈμ„ μƒμ„Ένžˆ μ„€λͺ…ν•΄μ£Όμ„Έμš” (ν•œκΈ€ μž…λ ₯ κ°€λŠ₯)",
lines=2
)
run_button = gr.Button("둜고 생성", variant="primary")
with gr.Row():
result = gr.Image(label="μƒμ„±λœ 둜고", show_label=True)
with gr.Accordion("κ³ κΈ‰ μ„€μ •", open=False):
with gr.Row():
seed = gr.Slider(label="μ‹œλ“œ", minimum=0, maximum=MAX_SEED, step=1, value=0)
randomize_seed = gr.Checkbox(label="랜덀 μ‹œλ“œ", value=True)
with gr.Row():
width = gr.Slider(label="λ„ˆλΉ„", minimum=256, maximum=MAX_IMAGE_SIZE, step=32, value=512)
height = gr.Slider(label="높이", minimum=256, maximum=MAX_IMAGE_SIZE, step=32, value=512)
num_inference_steps = gr.Slider(label="ν’ˆμ§ˆ", minimum=1, maximum=50, step=1, value=4)
gr.Examples(
examples=examples,
fn=infer,
inputs=[prompt],
outputs=[result, seed],
cache_examples="lazy"
)
gr.on(
triggers=[run_button.click, prompt.submit],
fn=infer,
inputs=[prompt, seed, randomize_seed, width, height, num_inference_steps],
outputs=[result, seed]
)
demo.launch()