Update app.py
Browse files
app.py
CHANGED
@@ -47,7 +47,7 @@ def inference(audio):
|
|
47 |
# Added prompt two 2 lines below
|
48 |
prompt = "The following is a transcript of someone talking, please predict what they will say next. \n"
|
49 |
### code
|
50 |
-
input_ids = tokenizer(
|
51 |
print("inputs ", input_ids)
|
52 |
|
53 |
# prompt length
|
@@ -55,7 +55,7 @@ def inference(audio):
|
|
55 |
|
56 |
# length penalty for gpt2.generate???
|
57 |
#Prompt
|
58 |
-
generated_outputs =
|
59 |
print("outputs generated ", generated_outputs[0])
|
60 |
# only use id's that were generated
|
61 |
# gen_sequences has shape [3, 15]
|
@@ -86,7 +86,8 @@ def inference(audio):
|
|
86 |
### end code
|
87 |
# print audio data as text
|
88 |
# print(result.text)
|
89 |
-
|
|
|
90 |
# pprint("getText: ", getText)
|
91 |
# pprint("text.result: ", result.text)
|
92 |
# result.text
|
|
|
47 |
# Added prompt two 2 lines below
|
48 |
prompt = "The following is a transcript of someone talking, please predict what they will say next. \n"
|
49 |
### code
|
50 |
+
input_ids = tokenizer(result.text, return_tensors="pt").input_ids
|
51 |
print("inputs ", input_ids)
|
52 |
|
53 |
# prompt length
|
|
|
55 |
|
56 |
# length penalty for gpt2.generate???
|
57 |
#Prompt
|
58 |
+
generated_outputs = gpt2.generate(input_ids, do_sample=True, num_return_sequences=3, output_scores=True)
|
59 |
print("outputs generated ", generated_outputs[0])
|
60 |
# only use id's that were generated
|
61 |
# gen_sequences has shape [3, 15]
|
|
|
86 |
### end code
|
87 |
# print audio data as text
|
88 |
# print(result.text)
|
89 |
+
# prompt
|
90 |
+
getText = generator(prompt + result.text, max_new_tokens=10, num_return_sequences=5)
|
91 |
# pprint("getText: ", getText)
|
92 |
# pprint("text.result: ", result.text)
|
93 |
# result.text
|