Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
import streamlit as st
|
2 |
-
from transformers import
|
3 |
from huggingface_hub import login
|
4 |
import PyPDF2
|
5 |
import pandas as pd
|
@@ -80,17 +80,22 @@ def load_model(hf_token):
|
|
80 |
|
81 |
login(token=hf_token)
|
82 |
|
83 |
-
# Load tokenizer
|
84 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
|
|
|
|
|
|
|
|
85 |
|
86 |
# Load the model with appropriate dtype for CPU/GPU compatibility
|
87 |
dtype = torch.float16 if DEVICE == "cuda" else torch.float32
|
88 |
-
model =
|
89 |
MODEL_NAME,
|
90 |
token=hf_token,
|
91 |
torch_dtype=dtype,
|
92 |
device_map="auto", # Automatically maps to CPU or GPU
|
93 |
-
quantization_config=None # Disable FP8 quantization
|
|
|
94 |
)
|
95 |
|
96 |
return model, tokenizer
|
|
|
1 |
import streamlit as st
|
2 |
+
from transformers import AutoModel, AutoTokenizer
|
3 |
from huggingface_hub import login
|
4 |
import PyPDF2
|
5 |
import pandas as pd
|
|
|
80 |
|
81 |
login(token=hf_token)
|
82 |
|
83 |
+
# Load tokenizer with trust_remote_code=True
|
84 |
+
tokenizer = AutoTokenizer.from_pretrained(
|
85 |
+
MODEL_NAME,
|
86 |
+
token=hf_token,
|
87 |
+
trust_remote_code=True
|
88 |
+
)
|
89 |
|
90 |
# Load the model with appropriate dtype for CPU/GPU compatibility
|
91 |
dtype = torch.float16 if DEVICE == "cuda" else torch.float32
|
92 |
+
model = AutoModel.from_pretrained(
|
93 |
MODEL_NAME,
|
94 |
token=hf_token,
|
95 |
torch_dtype=dtype,
|
96 |
device_map="auto", # Automatically maps to CPU or GPU
|
97 |
+
quantization_config=None, # Disable FP8 quantization
|
98 |
+
trust_remote_code=True # Allow custom code execution
|
99 |
)
|
100 |
|
101 |
return model, tokenizer
|