Upload 2 files
Browse files- Dockerfile +2 -2
- app.py +2 -2
Dockerfile
CHANGED
@@ -14,8 +14,8 @@ RUN pip install --no-cache-dir -r requirements.txt
|
|
14 |
RUN apt-get update && apt-get install -y curl
|
15 |
RUN curl -fsSL https://ollama.com/install.sh | sh
|
16 |
|
17 |
-
# Pre-download the llama3 model
|
18 |
-
RUN ollama pull llama3
|
19 |
|
20 |
# Copy the app code
|
21 |
COPY app.py .
|
|
|
14 |
RUN apt-get update && apt-get install -y curl
|
15 |
RUN curl -fsSL https://ollama.com/install.sh | sh
|
16 |
|
17 |
+
# Pre-download the llama3.2:1b model and debug
|
18 |
+
RUN ollama pull llama3.2:1b && ollama list
|
19 |
|
20 |
# Copy the app code
|
21 |
COPY app.py .
|
app.py
CHANGED
@@ -51,8 +51,8 @@ def setup_client(provider):
|
|
51 |
client = instructor.from_openai(
|
52 |
OllamaClient(base_url="http://localhost:11434/v1", api_key="ollama"), mode=instructor.Mode.JSON
|
53 |
)
|
54 |
-
model = "llama3"
|
55 |
-
display_model = "Ollama (llama3)"
|
56 |
else:
|
57 |
st.error(f"Unsupported provider: {provider}")
|
58 |
return None, None, None
|
|
|
51 |
client = instructor.from_openai(
|
52 |
OllamaClient(base_url="http://localhost:11434/v1", api_key="ollama"), mode=instructor.Mode.JSON
|
53 |
)
|
54 |
+
model = "llama3.2:1b" # Updated to use the smaller model
|
55 |
+
display_model = "Ollama (llama3.2:1b)"
|
56 |
else:
|
57 |
st.error(f"Unsupported provider: {provider}")
|
58 |
return None, None, None
|