Update app.py
Browse files
app.py
CHANGED
@@ -8,8 +8,7 @@ import uvicorn
|
|
8 |
app = FastAPI()
|
9 |
|
10 |
# Initialize tokenizer
|
11 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
12 |
-
"Xenova/multi-qa-mpnet-base-dot-v1")
|
13 |
|
14 |
# Load ONNX model
|
15 |
session = InferenceSession("model.onnx")
|
@@ -36,13 +35,7 @@ async def predict(request: Request):
|
|
36 |
raise HTTPException(status_code=400, detail="No text provided")
|
37 |
|
38 |
# Tokenize input
|
39 |
-
inputs = tokenizer(
|
40 |
-
text,
|
41 |
-
return_tensors="np",
|
42 |
-
padding=False, # Disable padding
|
43 |
-
truncation=False, # Disable truncation
|
44 |
-
add_special_tokens=True # Ensure CLS/SEP tokens
|
45 |
-
)
|
46 |
|
47 |
# Run model
|
48 |
outputs = session.run(None, {
|
|
|
8 |
app = FastAPI()
|
9 |
|
10 |
# Initialize tokenizer
|
11 |
+
tokenizer = AutoTokenizer.from_pretrained("Xenova/multi-qa-mpnet-base-dot-v1")
|
|
|
12 |
|
13 |
# Load ONNX model
|
14 |
session = InferenceSession("model.onnx")
|
|
|
35 |
raise HTTPException(status_code=400, detail="No text provided")
|
36 |
|
37 |
# Tokenize input
|
38 |
+
inputs = tokenizer(text)
|
|
|
|
|
|
|
|
|
|
|
|
|
39 |
|
40 |
# Run model
|
41 |
outputs = session.run(None, {
|