Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -13,7 +13,8 @@ import ollama
|
|
13 |
# Model from run.sh
|
14 |
AVAILABLE_MODELS = [
|
15 |
'smollm2:360m-instruct-q5_K_M',
|
16 |
-
'hf.co/
|
|
|
17 |
'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
|
18 |
'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M', #Slow.
|
19 |
'hf.co/bartowski/Qwen_Qwen3-1.7B-GGUF:Q5_K_M',
|
@@ -34,7 +35,7 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
|
|
34 |
with gr.Row():
|
35 |
selected_model = gr.Radio(
|
36 |
choices=AVAILABLE_MODELS,
|
37 |
-
value=AVAILABLE_MODELS[
|
38 |
label="Select Model",
|
39 |
info="Choose the LLM model to chat with.",
|
40 |
interactive=True
|
|
|
13 |
# Model from run.sh
|
14 |
AVAILABLE_MODELS = [
|
15 |
'smollm2:360m-instruct-q5_K_M',
|
16 |
+
'hf.co/ggml-org/SmolLM3-3B-GGUF:Q4_K_M,
|
17 |
+
'hf.co/bartowski/nvidia_OpenReasoning-Nemotron-1.5B-GGUF:Q5_K_M,
|
18 |
'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
|
19 |
'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M', #Slow.
|
20 |
'hf.co/bartowski/Qwen_Qwen3-1.7B-GGUF:Q5_K_M',
|
|
|
35 |
with gr.Row():
|
36 |
selected_model = gr.Radio(
|
37 |
choices=AVAILABLE_MODELS,
|
38 |
+
value=AVAILABLE_MODELS[0], # Default to the first model in the list
|
39 |
label="Select Model",
|
40 |
info="Choose the LLM model to chat with.",
|
41 |
interactive=True
|