amiguel commited on
Commit
d64ef24
·
verified ·
1 Parent(s): c7c3a66

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +10 -5
app.py CHANGED
@@ -1,5 +1,5 @@
1
  import streamlit as st
2
- from transformers import AutoModelForCausalLM, AutoTokenizer
3
  from huggingface_hub import login
4
  import PyPDF2
5
  import pandas as pd
@@ -80,17 +80,22 @@ def load_model(hf_token):
80
 
81
  login(token=hf_token)
82
 
83
- # Load tokenizer
84
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, token=hf_token)
 
 
 
 
85
 
86
  # Load the model with appropriate dtype for CPU/GPU compatibility
87
  dtype = torch.float16 if DEVICE == "cuda" else torch.float32
88
- model = AutoModelForCausalLM.from_pretrained(
89
  MODEL_NAME,
90
  token=hf_token,
91
  torch_dtype=dtype,
92
  device_map="auto", # Automatically maps to CPU or GPU
93
- quantization_config=None # Disable FP8 quantization
 
94
  )
95
 
96
  return model, tokenizer
 
1
  import streamlit as st
2
+ from transformers import AutoModel, AutoTokenizer
3
  from huggingface_hub import login
4
  import PyPDF2
5
  import pandas as pd
 
80
 
81
  login(token=hf_token)
82
 
83
+ # Load tokenizer with trust_remote_code=True
84
+ tokenizer = AutoTokenizer.from_pretrained(
85
+ MODEL_NAME,
86
+ token=hf_token,
87
+ trust_remote_code=True
88
+ )
89
 
90
  # Load the model with appropriate dtype for CPU/GPU compatibility
91
  dtype = torch.float16 if DEVICE == "cuda" else torch.float32
92
+ model = AutoModel.from_pretrained(
93
  MODEL_NAME,
94
  token=hf_token,
95
  torch_dtype=dtype,
96
  device_map="auto", # Automatically maps to CPU or GPU
97
+ quantization_config=None, # Disable FP8 quantization
98
+ trust_remote_code=True # Allow custom code execution
99
  )
100
 
101
  return model, tokenizer