Bils commited on
Commit
0a2ec1a
·
verified ·
1 Parent(s): cc2901f

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -6
app.py CHANGED
@@ -36,7 +36,7 @@ OUTPUTS_DIR = Path(os.environ.get("OUTPUTS_DIR", str(ROOT / "outputs" / "autosav
36
  OUTPUTS_DIR.mkdir(parents=True, exist_ok=True)
37
 
38
  SPACE_TITLE = "🎵 ShortiFoley — HunyuanVideo-Foley"
39
- SPACE_TAGLINE = "Text/Video Audio Foley · Created by bilsimaging.com"
40
  WATERMARK_NOTE = "Made with ❤️ by bilsimaging.com"
41
 
42
  # ZeroGPU limit
@@ -262,7 +262,7 @@ def infer_single_video(
262
 
263
 
264
  # -------------
265
- # Gradio UI (with MCP+API inside the same app)
266
  # -------------
267
  def _about_html() -> str:
268
  return f"""
@@ -459,7 +459,7 @@ Works great with media-automation in tools like **n8n**: call `load_model_tool`
459
  """
460
  )
461
 
462
- # ---- REST + MCP endpoints (inside Blocks) ----
463
  def _download_to_tmp(url: str) -> str:
464
  try:
465
  import requests
@@ -499,7 +499,7 @@ Works great with media-automation in tools like **n8n**: call `load_model_tool`
499
  sample_nums: int = 1,
500
  ) -> Dict[str, List[str]]:
501
  if _model_dict is None or _cfg is None:
502
- msg = auto_load_models(device_str="cpu") # safe in HTTP context; GPU will be used inside infer
503
  if not str(msg).startswith("✅"):
504
  raise RuntimeError(msg)
505
  local = _normalize_video_input(video_url_or_b64)
@@ -546,7 +546,7 @@ if __name__ == "__main__":
546
  logger.info("===== Application Startup =====\n")
547
  prepare_once()
548
 
549
- # Probe imports (early surfacing)
550
  sys.path.append(str(REPO_DIR))
551
  try:
552
  from hunyuanvideo_foley.utils.model_utils import load_model, denoise_process # noqa: F401
@@ -557,7 +557,7 @@ if __name__ == "__main__":
557
 
558
  ui = create_ui()
559
 
560
- # Enable MCP server so tools/resources/prompts are discoverable
561
  ui.launch(
562
  server_name="0.0.0.0",
563
  share=False,
 
36
  OUTPUTS_DIR.mkdir(parents=True, exist_ok=True)
37
 
38
  SPACE_TITLE = "🎵 ShortiFoley — HunyuanVideo-Foley"
39
+ SPACE_TAGLINE = "Bring your videos to life with AI-powered Foley"
40
  WATERMARK_NOTE = "Made with ❤️ by bilsimaging.com"
41
 
42
  # ZeroGPU limit
 
262
 
263
 
264
  # -------------
265
+ # Gradio UI
266
  # -------------
267
  def _about_html() -> str:
268
  return f"""
 
459
  """
460
  )
461
 
462
+ # ---- REST + MCP endpoints ----
463
  def _download_to_tmp(url: str) -> str:
464
  try:
465
  import requests
 
499
  sample_nums: int = 1,
500
  ) -> Dict[str, List[str]]:
501
  if _model_dict is None or _cfg is None:
502
+ msg = auto_load_models(device_str="cpu")
503
  if not str(msg).startswith("✅"):
504
  raise RuntimeError(msg)
505
  local = _normalize_video_input(video_url_or_b64)
 
546
  logger.info("===== Application Startup =====\n")
547
  prepare_once()
548
 
549
+ # Probe imports
550
  sys.path.append(str(REPO_DIR))
551
  try:
552
  from hunyuanvideo_foley.utils.model_utils import load_model, denoise_process # noqa: F401
 
557
 
558
  ui = create_ui()
559
 
560
+ # Enable MCP server
561
  ui.launch(
562
  server_name="0.0.0.0",
563
  share=False,