Spaces:
Runtime error
Runtime error
request_llm/bridge_tgui.py
CHANGED
|
@@ -26,7 +26,7 @@ def random_hash():
|
|
| 26 |
|
| 27 |
async def run(context):
|
| 28 |
params = {
|
| 29 |
-
'max_new_tokens':
|
| 30 |
'do_sample': True,
|
| 31 |
'temperature': 0.5,
|
| 32 |
'top_p': 0.9,
|
|
@@ -39,7 +39,7 @@ async def run(context):
|
|
| 39 |
'num_beams': 1,
|
| 40 |
'penalty_alpha': 0,
|
| 41 |
'length_penalty': 1,
|
| 42 |
-
'early_stopping':
|
| 43 |
'seed': -1,
|
| 44 |
}
|
| 45 |
session = random_hash()
|
|
@@ -144,7 +144,7 @@ def predict_tgui_no_ui(inputs, top_p, temperature, history=[], sys_prompt=""):
|
|
| 144 |
raw_input = "What I would like to say is the following: " + inputs
|
| 145 |
prompt = inputs
|
| 146 |
tgui_say = ""
|
| 147 |
-
mutable = [""]
|
| 148 |
def run_coorotine(mutable):
|
| 149 |
async def get_result(mutable):
|
| 150 |
async for response in run(prompt):
|
|
|
|
| 26 |
|
| 27 |
async def run(context):
|
| 28 |
params = {
|
| 29 |
+
'max_new_tokens': 512,
|
| 30 |
'do_sample': True,
|
| 31 |
'temperature': 0.5,
|
| 32 |
'top_p': 0.9,
|
|
|
|
| 39 |
'num_beams': 1,
|
| 40 |
'penalty_alpha': 0,
|
| 41 |
'length_penalty': 1,
|
| 42 |
+
'early_stopping': True,
|
| 43 |
'seed': -1,
|
| 44 |
}
|
| 45 |
session = random_hash()
|
|
|
|
| 144 |
raw_input = "What I would like to say is the following: " + inputs
|
| 145 |
prompt = inputs
|
| 146 |
tgui_say = ""
|
| 147 |
+
mutable = ["", time.time()]
|
| 148 |
def run_coorotine(mutable):
|
| 149 |
async def get_result(mutable):
|
| 150 |
async for response in run(prompt):
|