Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -59,10 +59,7 @@ state_dict = torch.load(filepath, map_location="cpu")
|
|
59 |
model.load_state_dict(state_dict)
|
60 |
model = model.to(DEVICE).eval()
|
61 |
|
62 |
-
from huggingface_hub import hf_hub_download
|
63 |
-
import os
|
64 |
import torch
|
65 |
-
from diffusers.utils import load_image
|
66 |
from controlnet_bria import BriaControlNetModel, BriaMultiControlNetModel
|
67 |
from pipeline_bria_controlnet import BriaControlNetPipeline
|
68 |
import PIL.Image as Image
|
@@ -71,9 +68,17 @@ base_model = 'briaai/BRIA-4B-Adapt'
|
|
71 |
controlnet_model = 'briaai/BRIA-4B-Adapt-ControlNet-Union'
|
72 |
|
73 |
controlnet = BriaControlNetModel.from_pretrained(controlnet_model, torch_dtype=torch.bfloat16)
|
74 |
-
controlnet = BriaMultiControlNetModel([controlnet])
|
75 |
-
|
76 |
pipe = BriaControlNetPipeline.from_pretrained(base_model, controlnet=controlnet, torch_dtype=torch.bfloat16, trust_remote_code=True)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
77 |
pipe.to("cuda")
|
78 |
|
79 |
mode_mapping = {
|
|
|
59 |
model.load_state_dict(state_dict)
|
60 |
model = model.to(DEVICE).eval()
|
61 |
|
|
|
|
|
62 |
import torch
|
|
|
63 |
from controlnet_bria import BriaControlNetModel, BriaMultiControlNetModel
|
64 |
from pipeline_bria_controlnet import BriaControlNetPipeline
|
65 |
import PIL.Image as Image
|
|
|
68 |
controlnet_model = 'briaai/BRIA-4B-Adapt-ControlNet-Union'
|
69 |
|
70 |
controlnet = BriaControlNetModel.from_pretrained(controlnet_model, torch_dtype=torch.bfloat16)
|
|
|
|
|
71 |
pipe = BriaControlNetPipeline.from_pretrained(base_model, controlnet=controlnet, torch_dtype=torch.bfloat16, trust_remote_code=True)
|
72 |
+
|
73 |
+
tokenizer = T5TokenizerFast.from_pretrained(
|
74 |
+
base_model, subfolder="tokenizer"
|
75 |
+
)
|
76 |
+
text_encoder = T5EncoderModel.from_pretrained(
|
77 |
+
base_model, subfolder="text_encoder", torch_dtype=torch.float16
|
78 |
+
)
|
79 |
+
|
80 |
+
pipe.text_encoder = text_encoder
|
81 |
+
pipe.tokenizer = tokenizer
|
82 |
pipe.to("cuda")
|
83 |
|
84 |
mode_mapping = {
|