Spaces:
Runtime error
Runtime error
more
Browse files- app-controlnet.py +10 -4
- app-img2img.py +8 -2
- app-txt2img.py +8 -2
- requirements.txt +1 -1
app-controlnet.py
CHANGED
|
@@ -35,6 +35,7 @@ import psutil
|
|
| 35 |
MAX_QUEUE_SIZE = int(os.environ.get("MAX_QUEUE_SIZE", 0))
|
| 36 |
TIMEOUT = float(os.environ.get("TIMEOUT", 0))
|
| 37 |
SAFETY_CHECKER = os.environ.get("SAFETY_CHECKER", None)
|
|
|
|
| 38 |
WIDTH = 512
|
| 39 |
HEIGHT = 512
|
| 40 |
# disable tiny autoencoder for better quality speed tradeoff
|
|
@@ -100,15 +101,20 @@ pipe.unet.to(memory_format=torch.channels_last)
|
|
| 100 |
if psutil.virtual_memory().total < 64 * 1024**3:
|
| 101 |
pipe.enable_attention_slicing()
|
| 102 |
|
| 103 |
-
if not mps_available and not xpu_available:
|
| 104 |
-
pipe.unet = torch.compile(pipe.unet, mode="reduce-overhead", fullgraph=True)
|
| 105 |
-
pipe(prompt="warmup", image=[Image.new("RGB", (768, 768))], control_image=[Image.new("RGB", (768, 768))])
|
| 106 |
-
|
| 107 |
compel_proc = Compel(
|
| 108 |
tokenizer=pipe.tokenizer,
|
| 109 |
text_encoder=pipe.text_encoder,
|
| 110 |
truncate_long_prompts=False,
|
| 111 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 112 |
user_queue_map = {}
|
| 113 |
|
| 114 |
|
|
|
|
| 35 |
MAX_QUEUE_SIZE = int(os.environ.get("MAX_QUEUE_SIZE", 0))
|
| 36 |
TIMEOUT = float(os.environ.get("TIMEOUT", 0))
|
| 37 |
SAFETY_CHECKER = os.environ.get("SAFETY_CHECKER", None)
|
| 38 |
+
TORCH_COMPILE = os.environ.get("TORCH_COMPILE", None)
|
| 39 |
WIDTH = 512
|
| 40 |
HEIGHT = 512
|
| 41 |
# disable tiny autoencoder for better quality speed tradeoff
|
|
|
|
| 101 |
if psutil.virtual_memory().total < 64 * 1024**3:
|
| 102 |
pipe.enable_attention_slicing()
|
| 103 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 104 |
compel_proc = Compel(
|
| 105 |
tokenizer=pipe.tokenizer,
|
| 106 |
text_encoder=pipe.text_encoder,
|
| 107 |
truncate_long_prompts=False,
|
| 108 |
)
|
| 109 |
+
if TORCH_COMPILE:
|
| 110 |
+
pipe.text_encoder = torch.compile(pipe.text_encoder, mode="max-autotune", fullgraph=False)
|
| 111 |
+
pipe.tokenizer = torch.compile(pipe.tokenizer, mode="max-autotune", fullgraph=False)
|
| 112 |
+
pipe.unet = torch.compile(pipe.unet, mode="max-autotune", fullgraph=False)
|
| 113 |
+
pipe.vae = torch.compile(pipe.vae, mode="max-autotune", fullgraph=False)
|
| 114 |
+
|
| 115 |
+
pipe(prompt="warmup", image=[Image.new("RGB", (768, 768))], control_image=[Image.new("RGB", (768, 768))])
|
| 116 |
+
|
| 117 |
+
|
| 118 |
user_queue_map = {}
|
| 119 |
|
| 120 |
|
app-img2img.py
CHANGED
|
@@ -29,6 +29,8 @@ import psutil
|
|
| 29 |
MAX_QUEUE_SIZE = int(os.environ.get("MAX_QUEUE_SIZE", 0))
|
| 30 |
TIMEOUT = float(os.environ.get("TIMEOUT", 0))
|
| 31 |
SAFETY_CHECKER = os.environ.get("SAFETY_CHECKER", None)
|
|
|
|
|
|
|
| 32 |
WIDTH = 512
|
| 33 |
HEIGHT = 512
|
| 34 |
# disable tiny autoencoder for better quality speed tradeoff
|
|
@@ -76,8 +78,12 @@ pipe.unet.to(memory_format=torch.channels_last)
|
|
| 76 |
if psutil.virtual_memory().total < 64 * 1024**3:
|
| 77 |
pipe.enable_attention_slicing()
|
| 78 |
|
| 79 |
-
if
|
| 80 |
-
pipe.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 81 |
pipe(prompt="warmup", image=[Image.new("RGB", (512, 512))])
|
| 82 |
|
| 83 |
compel_proc = Compel(
|
|
|
|
| 29 |
MAX_QUEUE_SIZE = int(os.environ.get("MAX_QUEUE_SIZE", 0))
|
| 30 |
TIMEOUT = float(os.environ.get("TIMEOUT", 0))
|
| 31 |
SAFETY_CHECKER = os.environ.get("SAFETY_CHECKER", None)
|
| 32 |
+
TORCH_COMPILE = os.environ.get("TORCH_COMPILE", None)
|
| 33 |
+
|
| 34 |
WIDTH = 512
|
| 35 |
HEIGHT = 512
|
| 36 |
# disable tiny autoencoder for better quality speed tradeoff
|
|
|
|
| 78 |
if psutil.virtual_memory().total < 64 * 1024**3:
|
| 79 |
pipe.enable_attention_slicing()
|
| 80 |
|
| 81 |
+
if TORCH_COMPILE:
|
| 82 |
+
pipe.text_encoder = torch.compile(pipe.text_encoder, mode="max-autotune", fullgraph=False)
|
| 83 |
+
pipe.tokenizer = torch.compile(pipe.tokenizer, mode="max-autotune", fullgraph=False)
|
| 84 |
+
pipe.unet = torch.compile(pipe.unet, mode="max-autotune", fullgraph=False)
|
| 85 |
+
pipe.vae = torch.compile(pipe.vae, mode="max-autotune", fullgraph=False)
|
| 86 |
+
|
| 87 |
pipe(prompt="warmup", image=[Image.new("RGB", (512, 512))])
|
| 88 |
|
| 89 |
compel_proc = Compel(
|
app-txt2img.py
CHANGED
|
@@ -30,6 +30,8 @@ import psutil
|
|
| 30 |
MAX_QUEUE_SIZE = int(os.environ.get("MAX_QUEUE_SIZE", 0))
|
| 31 |
TIMEOUT = float(os.environ.get("TIMEOUT", 0))
|
| 32 |
SAFETY_CHECKER = os.environ.get("SAFETY_CHECKER", None)
|
|
|
|
|
|
|
| 33 |
WIDTH = 768
|
| 34 |
HEIGHT = 768
|
| 35 |
# disable tiny autoencoder for better quality speed tradeoff
|
|
@@ -76,8 +78,12 @@ pipe.unet.to(memory_format=torch.channels_last)
|
|
| 76 |
if psutil.virtual_memory().total < 64 * 1024**3:
|
| 77 |
pipe.enable_attention_slicing()
|
| 78 |
|
| 79 |
-
if
|
| 80 |
-
pipe.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 81 |
pipe(prompt="warmup", num_inference_steps=1, guidance_scale=8.0)
|
| 82 |
|
| 83 |
compel_proc = Compel(
|
|
|
|
| 30 |
MAX_QUEUE_SIZE = int(os.environ.get("MAX_QUEUE_SIZE", 0))
|
| 31 |
TIMEOUT = float(os.environ.get("TIMEOUT", 0))
|
| 32 |
SAFETY_CHECKER = os.environ.get("SAFETY_CHECKER", None)
|
| 33 |
+
TORCH_COMPILE = os.environ.get("TORCH_COMPILE", None)
|
| 34 |
+
|
| 35 |
WIDTH = 768
|
| 36 |
HEIGHT = 768
|
| 37 |
# disable tiny autoencoder for better quality speed tradeoff
|
|
|
|
| 78 |
if psutil.virtual_memory().total < 64 * 1024**3:
|
| 79 |
pipe.enable_attention_slicing()
|
| 80 |
|
| 81 |
+
if TORCH_COMPILE:
|
| 82 |
+
pipe.text_encoder = torch.compile(pipe.text_encoder, mode="max-autotune", fullgraph=False)
|
| 83 |
+
pipe.tokenizer = torch.compile(pipe.tokenizer, mode="max-autotune", fullgraph=False)
|
| 84 |
+
pipe.unet = torch.compile(pipe.unet, mode="max-autotune", fullgraph=False)
|
| 85 |
+
pipe.vae = torch.compile(pipe.vae, mode="max-autotune", fullgraph=False)
|
| 86 |
+
|
| 87 |
pipe(prompt="warmup", num_inference_steps=1, guidance_scale=8.0)
|
| 88 |
|
| 89 |
compel_proc = Compel(
|
requirements.txt
CHANGED
|
@@ -1,4 +1,4 @@
|
|
| 1 |
-
diffusers==0.22.
|
| 2 |
transformers==4.34.1
|
| 3 |
gradio==3.50.2
|
| 4 |
--extra-index-url https://download.pytorch.org/whl/cu121
|
|
|
|
| 1 |
+
diffusers==0.22.2
|
| 2 |
transformers==4.34.1
|
| 3 |
gradio==3.50.2
|
| 4 |
--extra-index-url https://download.pytorch.org/whl/cu121
|