Spaces:
Running
Running
MVPilgrim
commited on
Commit
·
7cb1a56
1
Parent(s):
b1bc691
debug
Browse files- Dockerfile +2 -2
- app.py +1 -1
Dockerfile
CHANGED
@@ -3,8 +3,8 @@ FROM nvidia/cuda:12.2.0-base-ubuntu22.04 AS base
|
|
3 |
|
4 |
ENV DEBIAN_FRONTEND=noninteractive
|
5 |
ENV DEBCONF_NOWARNINGS="yes"
|
6 |
-
|
7 |
-
|
8 |
|
9 |
# Install necessary dependencies and musl
|
10 |
RUN apt-get update && apt-get install -y \
|
|
|
3 |
|
4 |
ENV DEBIAN_FRONTEND=noninteractive
|
5 |
ENV DEBCONF_NOWARNINGS="yes"
|
6 |
+
ENV CUDA_VISIBLE_DEVICES=0
|
7 |
+
ENV LLAMA_CUBLAS=1
|
8 |
|
9 |
# Install necessary dependencies and musl
|
10 |
RUN apt-get update && apt-get install -y \
|
app.py
CHANGED
@@ -351,7 +351,7 @@ try:
|
|
351 |
logger.info("### Initializing LLM.")
|
352 |
llm = Llama(model_path,
|
353 |
#*,
|
354 |
-
n_gpu_layers
|
355 |
split_mode=llama_cpp.LLAMA_SPLIT_MODE_LAYER,
|
356 |
main_gpu=0,
|
357 |
tensor_split=None,
|
|
|
351 |
logger.info("### Initializing LLM.")
|
352 |
llm = Llama(model_path,
|
353 |
#*,
|
354 |
+
n_gpu_layers=-1,
|
355 |
split_mode=llama_cpp.LLAMA_SPLIT_MODE_LAYER,
|
356 |
main_gpu=0,
|
357 |
tensor_split=None,
|