Spaces:
Running
Running
雷娃
commited on
Commit
·
37ff5ad
1
Parent(s):
0b90a57
fix stream output
Browse files
app.py
CHANGED
@@ -35,22 +35,13 @@ def chat(user_input, max_new_tokens=512):
|
|
35 |
thread = Thread(target=generate)
|
36 |
thread.start()
|
37 |
|
|
|
38 |
generated_text = ""
|
39 |
for new_text in streamer:
|
40 |
generated_text += new_text
|
41 |
-
yield generated_text
|
42 |
|
43 |
thread.join()
|
44 |
-
|
45 |
-
# generate response
|
46 |
-
#with torch.no_grad():
|
47 |
-
# outputs = model.generate(
|
48 |
-
# **inputs,
|
49 |
-
# max_new_tokens=max_new_tokens,
|
50 |
-
# pad_token_id=tokenizer.eos_token_id
|
51 |
-
# )
|
52 |
-
#response = tokenizer.decode(outputs[0][inputs['input_ids'].shape[-1]:], skip_special_tokens=True)
|
53 |
-
#return response
|
54 |
|
55 |
# Construct Gradio Interface
|
56 |
interface = gr.Interface(
|
|
|
35 |
thread = Thread(target=generate)
|
36 |
thread.start()
|
37 |
|
38 |
+
prompt_len = len(prompt)
|
39 |
generated_text = ""
|
40 |
for new_text in streamer:
|
41 |
generated_text += new_text
|
42 |
+
yield generated_text[prompt_len:]
|
43 |
|
44 |
thread.join()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
45 |
|
46 |
# Construct Gradio Interface
|
47 |
interface = gr.Interface(
|