Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -11,7 +11,7 @@ model_path = "Qwen3-0.6B-int4-ov"
|
|
11 |
hf_hub.snapshot_download(model_id, local_dir=model_path)
|
12 |
|
13 |
pipe = ov_genai.LLMPipeline(model_path, "CPU")
|
14 |
-
pipe.start_chat()
|
15 |
|
16 |
# 建立推論函式:使用 streamer 並回傳 generator 結果
|
17 |
def generate_stream(prompt):
|
@@ -42,7 +42,7 @@ with gr.Blocks() as demo:
|
|
42 |
gr.Markdown("## 🧠 OpenVINO Streaming Demo with Gradio Textbox")
|
43 |
|
44 |
textbox_input = gr.Textbox(label="Prompt", lines=1, placeholder="Enter prompt here...")
|
45 |
-
textbox_output = gr.Textbox(label="Output", lines=
|
46 |
|
47 |
# 按鈕控制觸發推論
|
48 |
button = gr.Button("Submit")
|
|
|
11 |
hf_hub.snapshot_download(model_id, local_dir=model_path)
|
12 |
|
13 |
pipe = ov_genai.LLMPipeline(model_path, "CPU")
|
14 |
+
#pipe.start_chat()
|
15 |
|
16 |
# 建立推論函式:使用 streamer 並回傳 generator 結果
|
17 |
def generate_stream(prompt):
|
|
|
42 |
gr.Markdown("## 🧠 OpenVINO Streaming Demo with Gradio Textbox")
|
43 |
|
44 |
textbox_input = gr.Textbox(label="Prompt", lines=1, placeholder="Enter prompt here...")
|
45 |
+
textbox_output = gr.Textbox(label="Output", lines=5)
|
46 |
|
47 |
# 按鈕控制觸發推論
|
48 |
button = gr.Button("Submit")
|