Spaces:
Sleeping
Sleeping
MVPilgrim
commited on
Commit
·
7cb1a56
1
Parent(s):
b1bc691
debug
Browse files- Dockerfile +2 -2
- app.py +1 -1
Dockerfile
CHANGED
|
@@ -3,8 +3,8 @@ FROM nvidia/cuda:12.2.0-base-ubuntu22.04 AS base
|
|
| 3 |
|
| 4 |
ENV DEBIAN_FRONTEND=noninteractive
|
| 5 |
ENV DEBCONF_NOWARNINGS="yes"
|
| 6 |
-
|
| 7 |
-
|
| 8 |
|
| 9 |
# Install necessary dependencies and musl
|
| 10 |
RUN apt-get update && apt-get install -y \
|
|
|
|
| 3 |
|
| 4 |
ENV DEBIAN_FRONTEND=noninteractive
|
| 5 |
ENV DEBCONF_NOWARNINGS="yes"
|
| 6 |
+
ENV CUDA_VISIBLE_DEVICES=0
|
| 7 |
+
ENV LLAMA_CUBLAS=1
|
| 8 |
|
| 9 |
# Install necessary dependencies and musl
|
| 10 |
RUN apt-get update && apt-get install -y \
|
app.py
CHANGED
|
@@ -351,7 +351,7 @@ try:
|
|
| 351 |
logger.info("### Initializing LLM.")
|
| 352 |
llm = Llama(model_path,
|
| 353 |
#*,
|
| 354 |
-
n_gpu_layers
|
| 355 |
split_mode=llama_cpp.LLAMA_SPLIT_MODE_LAYER,
|
| 356 |
main_gpu=0,
|
| 357 |
tensor_split=None,
|
|
|
|
| 351 |
logger.info("### Initializing LLM.")
|
| 352 |
llm = Llama(model_path,
|
| 353 |
#*,
|
| 354 |
+
n_gpu_layers=-1,
|
| 355 |
split_mode=llama_cpp.LLAMA_SPLIT_MODE_LAYER,
|
| 356 |
main_gpu=0,
|
| 357 |
tensor_split=None,
|