SkyNetWalker commited on
Commit
f672a9e
·
verified ·
1 Parent(s): c88240b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -13,7 +13,8 @@ import ollama
13
  # Model from run.sh
14
  AVAILABLE_MODELS = [
15
  'smollm2:360m-instruct-q5_K_M',
16
- 'hf.co/unsloth/LFM2-1.2B-GGUF:Q5_K_M',
 
17
  'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
18
  'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M', #Slow.
19
  'hf.co/bartowski/Qwen_Qwen3-1.7B-GGUF:Q5_K_M',
@@ -34,7 +35,7 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
34
  with gr.Row():
35
  selected_model = gr.Radio(
36
  choices=AVAILABLE_MODELS,
37
- value=AVAILABLE_MODELS[1], # Default to the first model in the list
38
  label="Select Model",
39
  info="Choose the LLM model to chat with.",
40
  interactive=True
 
13
  # Model from run.sh
14
  AVAILABLE_MODELS = [
15
  'smollm2:360m-instruct-q5_K_M',
16
+ 'hf.co/ggml-org/SmolLM3-3B-GGUF:Q4_K_M,
17
+ 'hf.co/bartowski/nvidia_OpenReasoning-Nemotron-1.5B-GGUF:Q5_K_M,
18
  'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
19
  'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M', #Slow.
20
  'hf.co/bartowski/Qwen_Qwen3-1.7B-GGUF:Q5_K_M',
 
35
  with gr.Row():
36
  selected_model = gr.Radio(
37
  choices=AVAILABLE_MODELS,
38
+ value=AVAILABLE_MODELS[0], # Default to the first model in the list
39
  label="Select Model",
40
  info="Choose the LLM model to chat with.",
41
  interactive=True