Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -7,8 +7,8 @@ from huggingface_hub import hf_hub_download
|
|
7 |
# Initialize Llama model from Hugging Face
|
8 |
llm = Llama(
|
9 |
model_path=hf_hub_download(
|
10 |
-
repo_id=os.environ.get("REPO_ID", "mradermacher/Atlas-Chat-
|
11 |
-
filename=os.environ.get("MODEL_FILE", "Atlas-Chat-
|
12 |
),
|
13 |
n_ctx=4096,
|
14 |
n_gpu_layers=-1,
|
@@ -65,7 +65,7 @@ def generate_text(
|
|
65 |
# Define the Gradio interface
|
66 |
demo = gr.ChatInterface(
|
67 |
generate_text,
|
68 |
-
title="
|
69 |
description="Running LLM with https://github.com/abetlen/llama-cpp-python",
|
70 |
examples=[
|
71 |
['How to setup a human base on Mars? Give short answer.'],
|
|
|
7 |
# Initialize Llama model from Hugging Face
|
8 |
llm = Llama(
|
9 |
model_path=hf_hub_download(
|
10 |
+
repo_id=os.environ.get("REPO_ID", "mradermacher/Atlas-Chat-2B-GGUF"),
|
11 |
+
filename=os.environ.get("MODEL_FILE", "Atlas-Chat-2B.Q8_0.gguf"),
|
12 |
),
|
13 |
n_ctx=4096,
|
14 |
n_gpu_layers=-1,
|
|
|
65 |
# Define the Gradio interface
|
66 |
demo = gr.ChatInterface(
|
67 |
generate_text,
|
68 |
+
title="using Atlas-Chat-2B | I had to switch to the 2B model because the 9B was too much for this space!",
|
69 |
description="Running LLM with https://github.com/abetlen/llama-cpp-python",
|
70 |
examples=[
|
71 |
['How to setup a human base on Mars? Give short answer.'],
|