Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -13,8 +13,6 @@ import ollama
|
|
13 |
# Model from run.sh
|
14 |
AVAILABLE_MODELS = [
|
15 |
'smollm2:360m-instruct-q5_K_M',
|
16 |
-
'hf.co/ggml-org/SmolLM3-3B-GGUF:Q4_K_M',
|
17 |
-
'hf.co/bartowski/nvidia_OpenReasoning-Nemotron-1.5B-GGUF:Q5_K_M',
|
18 |
'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
|
19 |
'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M', #Slow.
|
20 |
'hf.co/bartowski/Qwen_Qwen3-1.7B-GGUF:Q5_K_M',
|
@@ -22,12 +20,17 @@ AVAILABLE_MODELS = [
|
|
22 |
'granite3.3:2b'
|
23 |
]
|
24 |
|
|
|
|
|
|
|
|
|
|
|
25 |
# Default System Prompt
|
26 |
DEFAULT_SYSTEM_PROMPT = "Answer everything in simple, smart, relevant and accurate style. No chatty!"
|
27 |
|
28 |
# --- Gradio Interface ---
|
29 |
with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutral")) as demo:
|
30 |
-
gr.Markdown(f"##
|
31 |
gr.Markdown(f"(Run-Location-As: `{check_ipinfo}`)")
|
32 |
gr.Markdown("Chat with the model, customize its behavior with a system prompt, and toggle streaming output.")
|
33 |
|
|
|
13 |
# Model from run.sh
|
14 |
AVAILABLE_MODELS = [
|
15 |
'smollm2:360m-instruct-q5_K_M',
|
|
|
|
|
16 |
'hf.co/bartowski/Llama-3.2-3B-Instruct-GGUF:Q4_K_M', # OK speed with CPU
|
17 |
'hf.co/unsloth/gemma-3-4b-it-qat-GGUF:Q4_K_M', #Slow.
|
18 |
'hf.co/bartowski/Qwen_Qwen3-1.7B-GGUF:Q5_K_M',
|
|
|
20 |
'granite3.3:2b'
|
21 |
]
|
22 |
|
23 |
+
#---fail to run
|
24 |
+
#'hf.co/ggml-org/SmolLM3-3B-GGUF:Q4_K_M',
|
25 |
+
#'hf.co/bartowski/nvidia_OpenReasoning-Nemotron-1.5B-GGUF:Q5_K_M',
|
26 |
+
|
27 |
+
|
28 |
# Default System Prompt
|
29 |
DEFAULT_SYSTEM_PROMPT = "Answer everything in simple, smart, relevant and accurate style. No chatty!"
|
30 |
|
31 |
# --- Gradio Interface ---
|
32 |
with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutral")) as demo:
|
33 |
+
gr.Markdown(f"## Small Language Model (SLM) run with CPU") # Changed title to be more generic
|
34 |
gr.Markdown(f"(Run-Location-As: `{check_ipinfo}`)")
|
35 |
gr.Markdown("Chat with the model, customize its behavior with a system prompt, and toggle streaming output.")
|
36 |
|