Update inference_test_turkcell_with_intents.py
Browse files
inference_test_turkcell_with_intents.py
CHANGED
@@ -165,7 +165,11 @@ async def generate_response(text):
|
|
165 |
output_scores=True
|
166 |
)
|
167 |
|
168 |
-
|
|
|
|
|
|
|
|
|
169 |
|
170 |
@app.post("/chat")
|
171 |
async def chat(msg: Message):
|
@@ -192,17 +196,25 @@ async def chat(msg: Message):
|
|
192 |
traceback.print_exc()
|
193 |
return JSONResponse(content={"error": str(e)}, status_code=500)
|
194 |
|
|
|
|
|
|
|
|
|
195 |
def setup_model():
|
196 |
global model, tokenizer, eos_token_id
|
197 |
try:
|
|
|
198 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
|
|
199 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_BASE, use_fast=False)
|
|
|
200 |
model = AutoModelForCausalLM.from_pretrained(MODEL_BASE, torch_dtype=torch.float32).to(device)
|
201 |
tokenizer.pad_token = tokenizer.pad_token or tokenizer.eos_token
|
202 |
model.config.pad_token_id = tokenizer.pad_token_id
|
203 |
eos_token_id = tokenizer("<|im_end|>", add_special_tokens=False)["input_ids"][0]
|
204 |
model.eval()
|
205 |
-
|
|
|
206 |
except Exception as e:
|
207 |
traceback.print_exc()
|
208 |
|
|
|
165 |
output_scores=True
|
166 |
)
|
167 |
|
168 |
+
try:
|
169 |
+
return tokenizer.decode(output.sequences[0], skip_special_tokens=True).strip()
|
170 |
+
except Exception as decode_error:
|
171 |
+
log(f"❌ Decode hatası: {decode_error}")
|
172 |
+
return random.choice(FALLBACK_ANSWERS)
|
173 |
|
174 |
@app.post("/chat")
|
175 |
async def chat(msg: Message):
|
|
|
196 |
traceback.print_exc()
|
197 |
return JSONResponse(content={"error": str(e)}, status_code=500)
|
198 |
|
199 |
+
def log(message):
|
200 |
+
timestamp = datetime.now().strftime("%H:%M:%S")
|
201 |
+
print(f"[{timestamp}] {message}", flush=True)
|
202 |
+
|
203 |
def setup_model():
|
204 |
global model, tokenizer, eos_token_id
|
205 |
try:
|
206 |
+
log("🧠 setup_model() başladı")
|
207 |
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
|
208 |
+
log("📥 Tokenizer indiriliyor...")
|
209 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_BASE, use_fast=False)
|
210 |
+
log("🧠 Model indiriliyor...")
|
211 |
model = AutoModelForCausalLM.from_pretrained(MODEL_BASE, torch_dtype=torch.float32).to(device)
|
212 |
tokenizer.pad_token = tokenizer.pad_token or tokenizer.eos_token
|
213 |
model.config.pad_token_id = tokenizer.pad_token_id
|
214 |
eos_token_id = tokenizer("<|im_end|>", add_special_tokens=False)["input_ids"][0]
|
215 |
model.eval()
|
216 |
+
log("✅ Model eval() çağrıldı")
|
217 |
+
log("✔️ Model başarıyla yüklendi ve sohbet için hazır.")
|
218 |
except Exception as e:
|
219 |
traceback.print_exc()
|
220 |
|