Spaces:
Running
Running
Commit
·
609fc09
1
Parent(s):
17e1736
Upd syntax
Browse files- Dockerfile +1 -0
- vlm.py +2 -3
Dockerfile
CHANGED
@@ -17,6 +17,7 @@ RUN pip install --no-cache-dir -r requirements.txt
|
|
17 |
# Set Hugging Face cache directory to persist model downloads
|
18 |
ENV HF_HOME="/home/user/.cache/huggingface"
|
19 |
ENV SENTENCE_TRANSFORMERS_HOME="/home/user/.cache/huggingface/sentence-transformers"
|
|
|
20 |
|
21 |
# Create cache directories and ensure permissions
|
22 |
RUN mkdir -p /app/model_cache /home/user/.cache/huggingface/sentence-transformers && \
|
|
|
17 |
# Set Hugging Face cache directory to persist model downloads
|
18 |
ENV HF_HOME="/home/user/.cache/huggingface"
|
19 |
ENV SENTENCE_TRANSFORMERS_HOME="/home/user/.cache/huggingface/sentence-transformers"
|
20 |
+
ENV MEDGEMMA_HOME="/home/user/.cache/huggingface/sentence-transformers"
|
21 |
|
22 |
# Create cache directories and ensure permissions
|
23 |
RUN mkdir -p /app/model_cache /home/user/.cache/huggingface/sentence-transformers && \
|
vlm.py
CHANGED
@@ -2,8 +2,7 @@
|
|
2 |
import os, logging, traceback, json, base64
|
3 |
from io import BytesIO
|
4 |
from PIL import Image
|
5 |
-
|
6 |
-
from transformers import pipeline # Render model on transformers
|
7 |
from translation import translate_query
|
8 |
|
9 |
# Initialise once
|
@@ -19,7 +18,7 @@ def load_vlm():
|
|
19 |
global vlm_pipe
|
20 |
if vlm_pipe is None:
|
21 |
logger.info("⏳ Loading MedGEMMA model via Transformers pipeline...")
|
22 |
-
vlm_pipe = pipeline("image-to-text", model="google/medgemma-4b", use_auth_token=HF_TOKEN, device_map="auto")
|
23 |
logger.info("✅ MedGEMMA model ready.")
|
24 |
return vlm_pipe
|
25 |
|
|
|
2 |
import os, logging, traceback, json, base64
|
3 |
from io import BytesIO
|
4 |
from PIL import Image
|
5 |
+
from huggingface_hub import InferenceClient # Render model on HF hub
|
|
|
6 |
from translation import translate_query
|
7 |
|
8 |
# Initialise once
|
|
|
18 |
global vlm_pipe
|
19 |
if vlm_pipe is None:
|
20 |
logger.info("⏳ Loading MedGEMMA model via Transformers pipeline...")
|
21 |
+
vlm_pipe = pipeline("image-to-text", model="google/medgemma-4b-it", use_auth_token=HF_TOKEN, device_map="auto")
|
22 |
logger.info("✅ MedGEMMA model ready.")
|
23 |
return vlm_pipe
|
24 |
|