雷娃 commited on
Commit
37ff5ad
·
1 Parent(s): 0b90a57

fix stream output

Browse files
Files changed (1) hide show
  1. app.py +2 -11
app.py CHANGED
@@ -35,22 +35,13 @@ def chat(user_input, max_new_tokens=512):
35
  thread = Thread(target=generate)
36
  thread.start()
37
 
 
38
  generated_text = ""
39
  for new_text in streamer:
40
  generated_text += new_text
41
- yield generated_text
42
 
43
  thread.join()
44
-
45
- # generate response
46
- #with torch.no_grad():
47
- # outputs = model.generate(
48
- # **inputs,
49
- # max_new_tokens=max_new_tokens,
50
- # pad_token_id=tokenizer.eos_token_id
51
- # )
52
- #response = tokenizer.decode(outputs[0][inputs['input_ids'].shape[-1]:], skip_special_tokens=True)
53
- #return response
54
 
55
  # Construct Gradio Interface
56
  interface = gr.Interface(
 
35
  thread = Thread(target=generate)
36
  thread.start()
37
 
38
+ prompt_len = len(prompt)
39
  generated_text = ""
40
  for new_text in streamer:
41
  generated_text += new_text
42
+ yield generated_text[prompt_len:]
43
 
44
  thread.join()
 
 
 
 
 
 
 
 
 
 
45
 
46
  # Construct Gradio Interface
47
  interface = gr.Interface(