Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -23,11 +23,6 @@ hf_hub_download(
|
|
| 23 |
filename="Meta-Llama-3-8B-Instruct-Q6_K.gguf",
|
| 24 |
local_dir="./models"
|
| 25 |
)
|
| 26 |
-
hf_hub_download(
|
| 27 |
-
repo_id="bartowski/aya-23-8B-GGUF",
|
| 28 |
-
filename="aya-23-8B-Q6_K.gguf",
|
| 29 |
-
local_dir="./models"
|
| 30 |
-
)
|
| 31 |
|
| 32 |
css = """
|
| 33 |
.message-row {
|
|
@@ -76,8 +71,7 @@ PLACEHOLDER = """
|
|
| 76 |
def get_context_by_model(model_name):
|
| 77 |
model_context_limits = {
|
| 78 |
"Mistral-7B-Instruct-v0.3-Q6_K.gguf": 32768,
|
| 79 |
-
"Meta-Llama-3-8B-Instruct-Q6_K.gguf": 8192
|
| 80 |
-
"aya-23-8B-Q6_K.gguf": 8192
|
| 81 |
}
|
| 82 |
return model_context_limits.get(model_name, None)
|
| 83 |
|
|
@@ -254,8 +248,7 @@ demo = gr.ChatInterface(
|
|
| 254 |
),
|
| 255 |
gr.Dropdown([
|
| 256 |
'Mistral-7B-Instruct-v0.3-Q6_K.gguf',
|
| 257 |
-
'Meta-Llama-3-8B-Instruct-Q6_K.gguf'
|
| 258 |
-
'aya-23-8B-Q6_K.gguf'
|
| 259 |
],
|
| 260 |
value="Mistral-7B-Instruct-v0.3-Q6_K.gguf",
|
| 261 |
label="Model"
|
|
|
|
| 23 |
filename="Meta-Llama-3-8B-Instruct-Q6_K.gguf",
|
| 24 |
local_dir="./models"
|
| 25 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 26 |
|
| 27 |
css = """
|
| 28 |
.message-row {
|
|
|
|
| 71 |
def get_context_by_model(model_name):
|
| 72 |
model_context_limits = {
|
| 73 |
"Mistral-7B-Instruct-v0.3-Q6_K.gguf": 32768,
|
| 74 |
+
"Meta-Llama-3-8B-Instruct-Q6_K.gguf": 8192
|
|
|
|
| 75 |
}
|
| 76 |
return model_context_limits.get(model_name, None)
|
| 77 |
|
|
|
|
| 248 |
),
|
| 249 |
gr.Dropdown([
|
| 250 |
'Mistral-7B-Instruct-v0.3-Q6_K.gguf',
|
| 251 |
+
'Meta-Llama-3-8B-Instruct-Q6_K.gguf'
|
|
|
|
| 252 |
],
|
| 253 |
value="Mistral-7B-Instruct-v0.3-Q6_K.gguf",
|
| 254 |
label="Model"
|