Spaces:
Configuration error
Configuration error
Update models/depth_normal_pipeline_clip_cfg.py
Browse files
models/depth_normal_pipeline_clip_cfg.py
CHANGED
|
@@ -73,7 +73,7 @@ class DepthNormalEstimationPipeline(DiffusionPipeline):
|
|
| 73 |
@torch.no_grad()
|
| 74 |
def __call__(self,
|
| 75 |
input_image:Image,
|
| 76 |
-
|
| 77 |
ensemble_size: int = 10,
|
| 78 |
guidance_scale: int = 1,
|
| 79 |
processing_res: int = 768,
|
|
@@ -96,7 +96,7 @@ class DepthNormalEstimationPipeline(DiffusionPipeline):
|
|
| 96 |
)," Value Error: `resize_output_back` is only valid with "
|
| 97 |
|
| 98 |
assert processing_res >=0
|
| 99 |
-
assert
|
| 100 |
assert ensemble_size >=1
|
| 101 |
|
| 102 |
# --------------- Image Processing ------------------------
|
|
@@ -146,7 +146,7 @@ class DepthNormalEstimationPipeline(DiffusionPipeline):
|
|
| 146 |
|
| 147 |
depth_pred_raw, normal_pred_raw = self.single_infer(
|
| 148 |
input_rgb=batched_image,
|
| 149 |
-
num_inference_steps=
|
| 150 |
domain=domain,
|
| 151 |
show_pbar=show_progress_bar,
|
| 152 |
)
|
|
|
|
| 73 |
@torch.no_grad()
|
| 74 |
def __call__(self,
|
| 75 |
input_image:Image,
|
| 76 |
+
denoising_steps: int = 10,
|
| 77 |
ensemble_size: int = 10,
|
| 78 |
guidance_scale: int = 1,
|
| 79 |
processing_res: int = 768,
|
|
|
|
| 96 |
)," Value Error: `resize_output_back` is only valid with "
|
| 97 |
|
| 98 |
assert processing_res >=0
|
| 99 |
+
assert denoising_steps >=1
|
| 100 |
assert ensemble_size >=1
|
| 101 |
|
| 102 |
# --------------- Image Processing ------------------------
|
|
|
|
| 146 |
|
| 147 |
depth_pred_raw, normal_pred_raw = self.single_infer(
|
| 148 |
input_rgb=batched_image,
|
| 149 |
+
num_inference_steps=denoising_steps,
|
| 150 |
domain=domain,
|
| 151 |
show_pbar=show_progress_bar,
|
| 152 |
)
|