Spaces:
Running
on
Zero
Running
on
Zero
extend max tokens
Browse files
app.py
CHANGED
@@ -279,7 +279,7 @@ with gr.Blocks(title="LLM Inference with ZeroGPU") as demo:
|
|
279 |
search_chk = gr.Checkbox(label="Enable Web Search", value=True)
|
280 |
sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
|
281 |
gr.Markdown("### Generation Parameters")
|
282 |
-
max_tok = gr.Slider(64,
|
283 |
temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
|
284 |
k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
|
285 |
p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")
|
|
|
279 |
search_chk = gr.Checkbox(label="Enable Web Search", value=True)
|
280 |
sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
|
281 |
gr.Markdown("### Generation Parameters")
|
282 |
+
max_tok = gr.Slider(64, 16384, value=2048, step=32, label="Max Tokens")
|
283 |
temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
|
284 |
k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
|
285 |
p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")
|