Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -40,7 +40,7 @@ def generate(prompt,history):
|
|
40 |
for response in stream:
|
41 |
output += response.token.text
|
42 |
yield [(prompt,output)], ""
|
43 |
-
|
44 |
|
45 |
def load_mod(model):
|
46 |
yield f"Loading: {model}"
|
@@ -68,7 +68,7 @@ with gr.Blocks() as iface:
|
|
68 |
iface.load(load_mod,names,msg)
|
69 |
sub_b = submit_b.click(generate, [prompt,chatbot],[chatbot,stt])
|
70 |
sub_e = prompt.submit(generate, [prompt, chatbot], [chatbot,stt])
|
71 |
-
|
72 |
names_change=names.change(load_mod,names,msg)
|
73 |
-
stop_b.click(None,None,None, cancels=[sub_b,sub_e,
|
74 |
iface.queue(default_concurrency_limit=10).launch()
|
|
|
40 |
for response in stream:
|
41 |
output += response.token.text
|
42 |
yield [(prompt,output)], ""
|
43 |
+
yield [(prompt,output)], output
|
44 |
|
45 |
def load_mod(model):
|
46 |
yield f"Loading: {model}"
|
|
|
68 |
iface.load(load_mod,names,msg)
|
69 |
sub_b = submit_b.click(generate, [prompt,chatbot],[chatbot,stt])
|
70 |
sub_e = prompt.submit(generate, [prompt, chatbot], [chatbot,stt])
|
71 |
+
stt.change(pp.stream_tts,[stt,names,length,noise,width,sen_pause],aud)
|
72 |
names_change=names.change(load_mod,names,msg)
|
73 |
+
stop_b.click(None,None,None, cancels=[sub_b,sub_e,names_change])
|
74 |
iface.queue(default_concurrency_limit=10).launch()
|