Luigi commited on
Commit
d730ffe
·
1 Parent(s): 4911925

extend max tokens

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -279,7 +279,7 @@ with gr.Blocks(title="LLM Inference with ZeroGPU") as demo:
279
  search_chk = gr.Checkbox(label="Enable Web Search", value=True)
280
  sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
281
  gr.Markdown("### Generation Parameters")
282
- max_tok = gr.Slider(64, 1024, value=512, step=32, label="Max Tokens")
283
  temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
284
  k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
285
  p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")
 
279
  search_chk = gr.Checkbox(label="Enable Web Search", value=True)
280
  sys_prompt = gr.Textbox(label="System Prompt", lines=3, value=update_default_prompt(search_chk.value))
281
  gr.Markdown("### Generation Parameters")
282
+ max_tok = gr.Slider(64, 16384, value=2048, step=32, label="Max Tokens")
283
  temp = gr.Slider(0.1, 2.0, value=0.7, step=0.1, label="Temperature")
284
  k = gr.Slider(1, 100, value=40, step=1, label="Top-K")
285
  p = gr.Slider(0.1, 1.0, value=0.9, step=0.05, label="Top-P")