leeway.zlw
commited on
Commit
·
23a0ba6
1
Parent(s):
69c71b8
update
Browse files
app.py
CHANGED
|
@@ -7,12 +7,13 @@ from scripts.inference import inference_process
|
|
| 7 |
import argparse
|
| 8 |
import uuid
|
| 9 |
|
| 10 |
-
is_shared_ui = True if "fudan-generative-ai/hallo" in os.environ['SPACE_ID'] else False
|
|
|
|
| 11 |
|
| 12 |
if(not is_shared_ui):
|
| 13 |
hallo_dir = snapshot_download(repo_id="fudan-generative-ai/hallo", local_dir="pretrained_models")
|
| 14 |
|
| 15 |
-
def run_inference(source_image, driving_audio,
|
| 16 |
if is_shared_ui:
|
| 17 |
raise gr.Error("This Space only works in duplicated instances")
|
| 18 |
|
|
@@ -23,10 +24,10 @@ def run_inference(source_image, driving_audio, pose_weight, face_weight, lip_wei
|
|
| 23 |
source_image=source_image,
|
| 24 |
driving_audio=driving_audio,
|
| 25 |
output=f'output-{unique_id}.mp4',
|
| 26 |
-
pose_weight=
|
| 27 |
-
face_weight=
|
| 28 |
-
lip_weight=
|
| 29 |
-
face_expand_ratio=
|
| 30 |
checkpoint=None
|
| 31 |
)
|
| 32 |
|
|
@@ -112,17 +113,13 @@ We have provided some [samples](https://huggingface.co/datasets/fudan-generative
|
|
| 112 |
with gr.Column():
|
| 113 |
avatar_face = gr.Image(type="filepath", label="Face")
|
| 114 |
driving_audio = gr.Audio(type="filepath", label="Driving audio")
|
| 115 |
-
pose_weight = gr.Number(label="pose weight", value=1.0),
|
| 116 |
-
face_weight = gr.Number(label="face weight", value=1.0),
|
| 117 |
-
lip_weight = gr.Number(label="lip weight", value=1.0),
|
| 118 |
-
face_expand_ratio = gr.Number(label="face expand ratio", value=1.2),
|
| 119 |
generate = gr.Button("Generate")
|
| 120 |
with gr.Column():
|
| 121 |
output_video = gr.Video(label="Your talking head")
|
| 122 |
|
| 123 |
generate.click(
|
| 124 |
fn=run_inference,
|
| 125 |
-
inputs=[avatar_face, driving_audio
|
| 126 |
outputs=output_video
|
| 127 |
)
|
| 128 |
|
|
|
|
| 7 |
import argparse
|
| 8 |
import uuid
|
| 9 |
|
| 10 |
+
# is_shared_ui = True if "fudan-generative-ai/hallo" in os.environ['SPACE_ID'] else False
|
| 11 |
+
is_shared_ui = False
|
| 12 |
|
| 13 |
if(not is_shared_ui):
|
| 14 |
hallo_dir = snapshot_download(repo_id="fudan-generative-ai/hallo", local_dir="pretrained_models")
|
| 15 |
|
| 16 |
+
def run_inference(source_image, driving_audio, progress=gr.Progress(track_tqdm=True)):
|
| 17 |
if is_shared_ui:
|
| 18 |
raise gr.Error("This Space only works in duplicated instances")
|
| 19 |
|
|
|
|
| 24 |
source_image=source_image,
|
| 25 |
driving_audio=driving_audio,
|
| 26 |
output=f'output-{unique_id}.mp4',
|
| 27 |
+
pose_weight=1.0,
|
| 28 |
+
face_weight=1.0,
|
| 29 |
+
lip_weight=1.0,
|
| 30 |
+
face_expand_ratio=1.2,
|
| 31 |
checkpoint=None
|
| 32 |
)
|
| 33 |
|
|
|
|
| 113 |
with gr.Column():
|
| 114 |
avatar_face = gr.Image(type="filepath", label="Face")
|
| 115 |
driving_audio = gr.Audio(type="filepath", label="Driving audio")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 116 |
generate = gr.Button("Generate")
|
| 117 |
with gr.Column():
|
| 118 |
output_video = gr.Video(label="Your talking head")
|
| 119 |
|
| 120 |
generate.click(
|
| 121 |
fn=run_inference,
|
| 122 |
+
inputs=[avatar_face, driving_audio],
|
| 123 |
outputs=output_video
|
| 124 |
)
|
| 125 |
|