amine_dubs
commited on
Commit
·
6f3aee6
1
Parent(s):
068c749
erroe
Browse files- backend/main.py +14 -7
backend/main.py
CHANGED
@@ -51,11 +51,12 @@ os.environ['XDG_CACHE_HOME'] = '/tmp/cache'
|
|
51 |
# --- Global model and tokenizer variables ---
|
52 |
translator = None
|
53 |
tokenizer = None
|
|
|
54 |
|
55 |
# --- Model initialization function ---
|
56 |
def initialize_model():
|
57 |
"""Initialize the translation model and tokenizer."""
|
58 |
-
global translator, tokenizer
|
59 |
|
60 |
try:
|
61 |
print("Initializing model and tokenizer...")
|
@@ -69,16 +70,22 @@ def initialize_model():
|
|
69 |
cache_dir="/tmp/transformers_cache"
|
70 |
)
|
71 |
|
72 |
-
#
|
73 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
74 |
translator = pipeline(
|
75 |
"text2text-generation",
|
76 |
-
model=
|
77 |
tokenizer=tokenizer,
|
78 |
device=-1, # Use CPU for compatibility (-1) or GPU if available (0)
|
79 |
-
|
80 |
-
max_length=512,
|
81 |
-
model_kwargs={"from_tf": True} # This is the key fix
|
82 |
)
|
83 |
|
84 |
print(f"Model {model_name} successfully initialized")
|
|
|
51 |
# --- Global model and tokenizer variables ---
|
52 |
translator = None
|
53 |
tokenizer = None
|
54 |
+
model = None
|
55 |
|
56 |
# --- Model initialization function ---
|
57 |
def initialize_model():
|
58 |
"""Initialize the translation model and tokenizer."""
|
59 |
+
global translator, tokenizer, model
|
60 |
|
61 |
try:
|
62 |
print("Initializing model and tokenizer...")
|
|
|
70 |
cache_dir="/tmp/transformers_cache"
|
71 |
)
|
72 |
|
73 |
+
# Load the model explicitly with from_tf=True
|
74 |
+
print("Loading model with from_tf=True...")
|
75 |
+
model = AutoModelForSeq2SeqLM.from_pretrained(
|
76 |
+
model_name,
|
77 |
+
from_tf=True, # Explicitly set from_tf=True
|
78 |
+
cache_dir="/tmp/transformers_cache"
|
79 |
+
)
|
80 |
+
|
81 |
+
# Create a pipeline with the loaded model and tokenizer
|
82 |
+
print("Creating pipeline with pre-loaded model...")
|
83 |
translator = pipeline(
|
84 |
"text2text-generation",
|
85 |
+
model=model, # Use the model we loaded with from_tf=True
|
86 |
tokenizer=tokenizer,
|
87 |
device=-1, # Use CPU for compatibility (-1) or GPU if available (0)
|
88 |
+
max_length=512
|
|
|
|
|
89 |
)
|
90 |
|
91 |
print(f"Model {model_name} successfully initialized")
|