Spaces:
Running
Running
Commit
·
c367a95
1
Parent(s):
006bcec
fix: try to fix hf model dir
Browse files
core-model-prediction/Dockerfile
CHANGED
@@ -28,7 +28,7 @@ RUN python -c "from transformers import AutoTokenizer, AutoModelForCausalLM; \
|
|
28 |
model.save_pretrained('/app/gemma-2b')"
|
29 |
|
30 |
# Model env
|
31 |
-
ENV MODEL_DIR
|
32 |
|
33 |
# Make port 8080 available to the world outside this container
|
34 |
EXPOSE 8080
|
|
|
28 |
model.save_pretrained('/app/gemma-2b')"
|
29 |
|
30 |
# Model env
|
31 |
+
ENV MODEL_DIR=gemma-2b
|
32 |
|
33 |
# Make port 8080 available to the world outside this container
|
34 |
EXPOSE 8080
|
core-model-prediction/gemma2b_dependencies.py
CHANGED
@@ -13,9 +13,9 @@ class Gemma2BDependencies:
|
|
13 |
def __new__(cls):
|
14 |
if cls._instance is None:
|
15 |
cls._instance = super(Gemma2BDependencies, cls).__new__(cls)
|
16 |
-
model_dir = os.getenv("MODEL_DIR", "
|
17 |
-
cls._instance.tokenizer = AutoTokenizer.from_pretrained(model_dir)
|
18 |
-
cls._instance.model = AutoModelForCausalLM.from_pretrained(model_dir)
|
19 |
cls._instance.device = DeviceManager()
|
20 |
cls._instance.model.to(cls._instance.device)
|
21 |
return cls._instance
|
|
|
13 |
def __new__(cls):
|
14 |
if cls._instance is None:
|
15 |
cls._instance = super(Gemma2BDependencies, cls).__new__(cls)
|
16 |
+
model_dir = os.getenv("MODEL_DIR", "gemma-2b")
|
17 |
+
cls._instance.tokenizer = AutoTokenizer.from_pretrained(model_dir, local_files_only=True)
|
18 |
+
cls._instance.model = AutoModelForCausalLM.from_pretrained(model_dir, local_files_only=True)
|
19 |
cls._instance.device = DeviceManager()
|
20 |
cls._instance.model.to(cls._instance.device)
|
21 |
return cls._instance
|