Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse filesesbieta-ec-qwen-2.5-3B-F16.gguf
app.py
CHANGED
@@ -18,8 +18,8 @@ hf_hub_download(
|
|
18 |
local_dir = "./models"
|
19 |
)
|
20 |
hf_hub_download(
|
21 |
-
repo_id="
|
22 |
-
filename="
|
23 |
local_dir = "./models"
|
24 |
)
|
25 |
|
@@ -77,8 +77,8 @@ css = """
|
|
77 |
def get_messages_formatter_type(model_name):
|
78 |
if "Llama" in model_name:
|
79 |
return MessagesFormatterType.LLAMA_3
|
80 |
-
elif "
|
81 |
-
return MessagesFormatterType.
|
82 |
else:
|
83 |
raise ValueError(f"Unsupported model: {model_name}")
|
84 |
|
@@ -196,7 +196,7 @@ demo = gr.ChatInterface(
|
|
196 |
'Meta-Llama-3-70B-Instruct-Q3_K_M.gguf',
|
197 |
'gemma-2-27b-it-Q8_0.gguf'
|
198 |
],
|
199 |
-
value="
|
200 |
label="Model"
|
201 |
),
|
202 |
gr.Textbox(value="You are a helpful assistant.", label="System message"),
|
|
|
18 |
local_dir = "./models"
|
19 |
)
|
20 |
hf_hub_download(
|
21 |
+
repo_id="pabloce/esbieta-ec-qwen-2.5-3B",
|
22 |
+
filename="esbieta-ec-qwen-2.5-3B-F16.gguf",
|
23 |
local_dir = "./models"
|
24 |
)
|
25 |
|
|
|
77 |
def get_messages_formatter_type(model_name):
|
78 |
if "Llama" in model_name:
|
79 |
return MessagesFormatterType.LLAMA_3
|
80 |
+
elif "esbieta" in model_name:
|
81 |
+
return MessagesFormatterType.CHATML
|
82 |
else:
|
83 |
raise ValueError(f"Unsupported model: {model_name}")
|
84 |
|
|
|
196 |
'Meta-Llama-3-70B-Instruct-Q3_K_M.gguf',
|
197 |
'gemma-2-27b-it-Q8_0.gguf'
|
198 |
],
|
199 |
+
value="esbieta-ec-qwen-2.5-3B-F16.gguf",
|
200 |
label="Model"
|
201 |
),
|
202 |
gr.Textbox(value="You are a helpful assistant.", label="System message"),
|