Diffusers Bot
commited on
Upload folder using huggingface_hub
Browse files- main/lpw_stable_diffusion_xl.py +2 -12
- main/pipeline_demofusion_sdxl.py +2 -12
- main/pipeline_sdxl_style_aligned.py +0 -4
- main/pipeline_stable_diffusion_xl_controlnet_adapter.py +2 -12
- main/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py +2 -12
- main/pipeline_stable_diffusion_xl_differential_img2img.py +0 -4
- main/pipeline_stable_diffusion_xl_ipex.py +0 -4
main/lpw_stable_diffusion_xl.py
CHANGED
|
@@ -24,12 +24,7 @@ from diffusers import DiffusionPipeline, StableDiffusionXLPipeline
|
|
| 24 |
from diffusers.image_processor import PipelineImageInput, VaeImageProcessor
|
| 25 |
from diffusers.loaders import FromSingleFileMixin, IPAdapterMixin, LoraLoaderMixin, TextualInversionLoaderMixin
|
| 26 |
from diffusers.models import AutoencoderKL, ImageProjection, UNet2DConditionModel
|
| 27 |
-
from diffusers.models.attention_processor import
|
| 28 |
-
AttnProcessor2_0,
|
| 29 |
-
LoRAAttnProcessor2_0,
|
| 30 |
-
LoRAXFormersAttnProcessor,
|
| 31 |
-
XFormersAttnProcessor,
|
| 32 |
-
)
|
| 33 |
from diffusers.pipelines.pipeline_utils import StableDiffusionMixin
|
| 34 |
from diffusers.pipelines.stable_diffusion_xl.pipeline_output import StableDiffusionXLPipelineOutput
|
| 35 |
from diffusers.schedulers import KarrasDiffusionSchedulers
|
|
@@ -1292,12 +1287,7 @@ class SDXLLongPromptWeightingPipeline(
|
|
| 1292 |
self.vae.to(dtype=torch.float32)
|
| 1293 |
use_torch_2_0_or_xformers = isinstance(
|
| 1294 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 1295 |
-
(
|
| 1296 |
-
AttnProcessor2_0,
|
| 1297 |
-
XFormersAttnProcessor,
|
| 1298 |
-
LoRAXFormersAttnProcessor,
|
| 1299 |
-
LoRAAttnProcessor2_0,
|
| 1300 |
-
),
|
| 1301 |
)
|
| 1302 |
# if xformers or torch_2_0 is used attention block does not need
|
| 1303 |
# to be in float32 which can save lots of memory
|
|
|
|
| 24 |
from diffusers.image_processor import PipelineImageInput, VaeImageProcessor
|
| 25 |
from diffusers.loaders import FromSingleFileMixin, IPAdapterMixin, LoraLoaderMixin, TextualInversionLoaderMixin
|
| 26 |
from diffusers.models import AutoencoderKL, ImageProjection, UNet2DConditionModel
|
| 27 |
+
from diffusers.models.attention_processor import AttnProcessor2_0, XFormersAttnProcessor
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 28 |
from diffusers.pipelines.pipeline_utils import StableDiffusionMixin
|
| 29 |
from diffusers.pipelines.stable_diffusion_xl.pipeline_output import StableDiffusionXLPipelineOutput
|
| 30 |
from diffusers.schedulers import KarrasDiffusionSchedulers
|
|
|
|
| 1287 |
self.vae.to(dtype=torch.float32)
|
| 1288 |
use_torch_2_0_or_xformers = isinstance(
|
| 1289 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 1290 |
+
(AttnProcessor2_0, XFormersAttnProcessor),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1291 |
)
|
| 1292 |
# if xformers or torch_2_0 is used attention block does not need
|
| 1293 |
# to be in float32 which can save lots of memory
|
main/pipeline_demofusion_sdxl.py
CHANGED
|
@@ -16,12 +16,7 @@ from diffusers.loaders import (
|
|
| 16 |
TextualInversionLoaderMixin,
|
| 17 |
)
|
| 18 |
from diffusers.models import AutoencoderKL, UNet2DConditionModel
|
| 19 |
-
from diffusers.models.attention_processor import
|
| 20 |
-
AttnProcessor2_0,
|
| 21 |
-
LoRAAttnProcessor2_0,
|
| 22 |
-
LoRAXFormersAttnProcessor,
|
| 23 |
-
XFormersAttnProcessor,
|
| 24 |
-
)
|
| 25 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
| 26 |
from diffusers.pipelines.pipeline_utils import DiffusionPipeline, StableDiffusionMixin
|
| 27 |
from diffusers.schedulers import KarrasDiffusionSchedulers
|
|
@@ -612,12 +607,7 @@ class DemoFusionSDXLPipeline(
|
|
| 612 |
self.vae.to(dtype=torch.float32)
|
| 613 |
use_torch_2_0_or_xformers = isinstance(
|
| 614 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 615 |
-
(
|
| 616 |
-
AttnProcessor2_0,
|
| 617 |
-
XFormersAttnProcessor,
|
| 618 |
-
LoRAXFormersAttnProcessor,
|
| 619 |
-
LoRAAttnProcessor2_0,
|
| 620 |
-
),
|
| 621 |
)
|
| 622 |
# if xformers or torch_2_0 is used attention block does not need
|
| 623 |
# to be in float32 which can save lots of memory
|
|
|
|
| 16 |
TextualInversionLoaderMixin,
|
| 17 |
)
|
| 18 |
from diffusers.models import AutoencoderKL, UNet2DConditionModel
|
| 19 |
+
from diffusers.models.attention_processor import AttnProcessor2_0, XFormersAttnProcessor
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 20 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
| 21 |
from diffusers.pipelines.pipeline_utils import DiffusionPipeline, StableDiffusionMixin
|
| 22 |
from diffusers.schedulers import KarrasDiffusionSchedulers
|
|
|
|
| 607 |
self.vae.to(dtype=torch.float32)
|
| 608 |
use_torch_2_0_or_xformers = isinstance(
|
| 609 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 610 |
+
(AttnProcessor2_0, XFormersAttnProcessor),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 611 |
)
|
| 612 |
# if xformers or torch_2_0 is used attention block does not need
|
| 613 |
# to be in float32 which can save lots of memory
|
main/pipeline_sdxl_style_aligned.py
CHANGED
|
@@ -46,8 +46,6 @@ from diffusers.models.attention_processor import (
|
|
| 46 |
Attention,
|
| 47 |
AttnProcessor2_0,
|
| 48 |
FusedAttnProcessor2_0,
|
| 49 |
-
LoRAAttnProcessor2_0,
|
| 50 |
-
LoRAXFormersAttnProcessor,
|
| 51 |
XFormersAttnProcessor,
|
| 52 |
)
|
| 53 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
|
@@ -1153,8 +1151,6 @@ class StyleAlignedSDXLPipeline(
|
|
| 1153 |
(
|
| 1154 |
AttnProcessor2_0,
|
| 1155 |
XFormersAttnProcessor,
|
| 1156 |
-
LoRAXFormersAttnProcessor,
|
| 1157 |
-
LoRAAttnProcessor2_0,
|
| 1158 |
FusedAttnProcessor2_0,
|
| 1159 |
),
|
| 1160 |
)
|
|
|
|
| 46 |
Attention,
|
| 47 |
AttnProcessor2_0,
|
| 48 |
FusedAttnProcessor2_0,
|
|
|
|
|
|
|
| 49 |
XFormersAttnProcessor,
|
| 50 |
)
|
| 51 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
|
|
|
| 1151 |
(
|
| 1152 |
AttnProcessor2_0,
|
| 1153 |
XFormersAttnProcessor,
|
|
|
|
|
|
|
| 1154 |
FusedAttnProcessor2_0,
|
| 1155 |
),
|
| 1156 |
)
|
main/pipeline_stable_diffusion_xl_controlnet_adapter.py
CHANGED
|
@@ -25,12 +25,7 @@ from transformers import CLIPTextModel, CLIPTextModelWithProjection, CLIPTokeniz
|
|
| 25 |
from diffusers.image_processor import PipelineImageInput, VaeImageProcessor
|
| 26 |
from diffusers.loaders import FromSingleFileMixin, StableDiffusionXLLoraLoaderMixin, TextualInversionLoaderMixin
|
| 27 |
from diffusers.models import AutoencoderKL, ControlNetModel, MultiAdapter, T2IAdapter, UNet2DConditionModel
|
| 28 |
-
from diffusers.models.attention_processor import
|
| 29 |
-
AttnProcessor2_0,
|
| 30 |
-
LoRAAttnProcessor2_0,
|
| 31 |
-
LoRAXFormersAttnProcessor,
|
| 32 |
-
XFormersAttnProcessor,
|
| 33 |
-
)
|
| 34 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
| 35 |
from diffusers.pipelines.controlnet.multicontrolnet import MultiControlNetModel
|
| 36 |
from diffusers.pipelines.pipeline_utils import DiffusionPipeline, StableDiffusionMixin
|
|
@@ -797,12 +792,7 @@ class StableDiffusionXLControlNetAdapterPipeline(
|
|
| 797 |
self.vae.to(dtype=torch.float32)
|
| 798 |
use_torch_2_0_or_xformers = isinstance(
|
| 799 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 800 |
-
(
|
| 801 |
-
AttnProcessor2_0,
|
| 802 |
-
XFormersAttnProcessor,
|
| 803 |
-
LoRAXFormersAttnProcessor,
|
| 804 |
-
LoRAAttnProcessor2_0,
|
| 805 |
-
),
|
| 806 |
)
|
| 807 |
# if xformers or torch_2_0 is used attention block does not need
|
| 808 |
# to be in float32 which can save lots of memory
|
|
|
|
| 25 |
from diffusers.image_processor import PipelineImageInput, VaeImageProcessor
|
| 26 |
from diffusers.loaders import FromSingleFileMixin, StableDiffusionXLLoraLoaderMixin, TextualInversionLoaderMixin
|
| 27 |
from diffusers.models import AutoencoderKL, ControlNetModel, MultiAdapter, T2IAdapter, UNet2DConditionModel
|
| 28 |
+
from diffusers.models.attention_processor import AttnProcessor2_0, XFormersAttnProcessor
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
| 30 |
from diffusers.pipelines.controlnet.multicontrolnet import MultiControlNetModel
|
| 31 |
from diffusers.pipelines.pipeline_utils import DiffusionPipeline, StableDiffusionMixin
|
|
|
|
| 792 |
self.vae.to(dtype=torch.float32)
|
| 793 |
use_torch_2_0_or_xformers = isinstance(
|
| 794 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 795 |
+
(AttnProcessor2_0, XFormersAttnProcessor),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 796 |
)
|
| 797 |
# if xformers or torch_2_0 is used attention block does not need
|
| 798 |
# to be in float32 which can save lots of memory
|
main/pipeline_stable_diffusion_xl_controlnet_adapter_inpaint.py
CHANGED
|
@@ -44,12 +44,7 @@ from diffusers.models import (
|
|
| 44 |
T2IAdapter,
|
| 45 |
UNet2DConditionModel,
|
| 46 |
)
|
| 47 |
-
from diffusers.models.attention_processor import
|
| 48 |
-
AttnProcessor2_0,
|
| 49 |
-
LoRAAttnProcessor2_0,
|
| 50 |
-
LoRAXFormersAttnProcessor,
|
| 51 |
-
XFormersAttnProcessor,
|
| 52 |
-
)
|
| 53 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
| 54 |
from diffusers.pipelines.controlnet.multicontrolnet import MultiControlNetModel
|
| 55 |
from diffusers.pipelines.pipeline_utils import StableDiffusionMixin
|
|
@@ -1135,12 +1130,7 @@ class StableDiffusionXLControlNetAdapterInpaintPipeline(
|
|
| 1135 |
self.vae.to(dtype=torch.float32)
|
| 1136 |
use_torch_2_0_or_xformers = isinstance(
|
| 1137 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 1138 |
-
(
|
| 1139 |
-
AttnProcessor2_0,
|
| 1140 |
-
XFormersAttnProcessor,
|
| 1141 |
-
LoRAXFormersAttnProcessor,
|
| 1142 |
-
LoRAAttnProcessor2_0,
|
| 1143 |
-
),
|
| 1144 |
)
|
| 1145 |
# if xformers or torch_2_0 is used attention block does not need
|
| 1146 |
# to be in float32 which can save lots of memory
|
|
|
|
| 44 |
T2IAdapter,
|
| 45 |
UNet2DConditionModel,
|
| 46 |
)
|
| 47 |
+
from diffusers.models.attention_processor import AttnProcessor2_0, XFormersAttnProcessor
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
| 49 |
from diffusers.pipelines.controlnet.multicontrolnet import MultiControlNetModel
|
| 50 |
from diffusers.pipelines.pipeline_utils import StableDiffusionMixin
|
|
|
|
| 1130 |
self.vae.to(dtype=torch.float32)
|
| 1131 |
use_torch_2_0_or_xformers = isinstance(
|
| 1132 |
self.vae.decoder.mid_block.attentions[0].processor,
|
| 1133 |
+
(AttnProcessor2_0, XFormersAttnProcessor),
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1134 |
)
|
| 1135 |
# if xformers or torch_2_0 is used attention block does not need
|
| 1136 |
# to be in float32 which can save lots of memory
|
main/pipeline_stable_diffusion_xl_differential_img2img.py
CHANGED
|
@@ -37,8 +37,6 @@ from diffusers.loaders import (
|
|
| 37 |
from diffusers.models import AutoencoderKL, ImageProjection, UNet2DConditionModel
|
| 38 |
from diffusers.models.attention_processor import (
|
| 39 |
AttnProcessor2_0,
|
| 40 |
-
LoRAAttnProcessor2_0,
|
| 41 |
-
LoRAXFormersAttnProcessor,
|
| 42 |
XFormersAttnProcessor,
|
| 43 |
)
|
| 44 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
|
@@ -854,8 +852,6 @@ class StableDiffusionXLDifferentialImg2ImgPipeline(
|
|
| 854 |
(
|
| 855 |
AttnProcessor2_0,
|
| 856 |
XFormersAttnProcessor,
|
| 857 |
-
LoRAXFormersAttnProcessor,
|
| 858 |
-
LoRAAttnProcessor2_0,
|
| 859 |
),
|
| 860 |
)
|
| 861 |
# if xformers or torch_2_0 is used attention block does not need
|
|
|
|
| 37 |
from diffusers.models import AutoencoderKL, ImageProjection, UNet2DConditionModel
|
| 38 |
from diffusers.models.attention_processor import (
|
| 39 |
AttnProcessor2_0,
|
|
|
|
|
|
|
| 40 |
XFormersAttnProcessor,
|
| 41 |
)
|
| 42 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
|
|
|
| 852 |
(
|
| 853 |
AttnProcessor2_0,
|
| 854 |
XFormersAttnProcessor,
|
|
|
|
|
|
|
| 855 |
),
|
| 856 |
)
|
| 857 |
# if xformers or torch_2_0 is used attention block does not need
|
main/pipeline_stable_diffusion_xl_ipex.py
CHANGED
|
@@ -34,8 +34,6 @@ from diffusers.loaders import (
|
|
| 34 |
from diffusers.models import AutoencoderKL, UNet2DConditionModel
|
| 35 |
from diffusers.models.attention_processor import (
|
| 36 |
AttnProcessor2_0,
|
| 37 |
-
LoRAAttnProcessor2_0,
|
| 38 |
-
LoRAXFormersAttnProcessor,
|
| 39 |
XFormersAttnProcessor,
|
| 40 |
)
|
| 41 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
|
@@ -662,8 +660,6 @@ class StableDiffusionXLPipelineIpex(
|
|
| 662 |
(
|
| 663 |
AttnProcessor2_0,
|
| 664 |
XFormersAttnProcessor,
|
| 665 |
-
LoRAXFormersAttnProcessor,
|
| 666 |
-
LoRAAttnProcessor2_0,
|
| 667 |
),
|
| 668 |
)
|
| 669 |
# if xformers or torch_2_0 is used attention block does not need
|
|
|
|
| 34 |
from diffusers.models import AutoencoderKL, UNet2DConditionModel
|
| 35 |
from diffusers.models.attention_processor import (
|
| 36 |
AttnProcessor2_0,
|
|
|
|
|
|
|
| 37 |
XFormersAttnProcessor,
|
| 38 |
)
|
| 39 |
from diffusers.models.lora import adjust_lora_scale_text_encoder
|
|
|
|
| 660 |
(
|
| 661 |
AttnProcessor2_0,
|
| 662 |
XFormersAttnProcessor,
|
|
|
|
|
|
|
| 663 |
),
|
| 664 |
)
|
| 665 |
# if xformers or torch_2_0 is used attention block does not need
|