ciyidogan commited on
Commit
e2fcbc6
·
verified ·
1 Parent(s): 404ab93

Update inference_test_turkcell_with_intents.py

Browse files
inference_test_turkcell_with_intents.py CHANGED
@@ -165,7 +165,11 @@ async def generate_response(text):
165
  output_scores=True
166
  )
167
 
168
- return tokenizer.decode(output.sequences[0], skip_special_tokens=True).strip()
 
 
 
 
169
 
170
  @app.post("/chat")
171
  async def chat(msg: Message):
@@ -192,17 +196,25 @@ async def chat(msg: Message):
192
  traceback.print_exc()
193
  return JSONResponse(content={"error": str(e)}, status_code=500)
194
 
 
 
 
 
195
  def setup_model():
196
  global model, tokenizer, eos_token_id
197
  try:
 
198
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
 
199
  tokenizer = AutoTokenizer.from_pretrained(MODEL_BASE, use_fast=False)
 
200
  model = AutoModelForCausalLM.from_pretrained(MODEL_BASE, torch_dtype=torch.float32).to(device)
201
  tokenizer.pad_token = tokenizer.pad_token or tokenizer.eos_token
202
  model.config.pad_token_id = tokenizer.pad_token_id
203
  eos_token_id = tokenizer("<|im_end|>", add_special_tokens=False)["input_ids"][0]
204
  model.eval()
205
- print("[✔️] Model başarıyla yüklendi ve sohbet için hazır.")
 
206
  except Exception as e:
207
  traceback.print_exc()
208
 
 
165
  output_scores=True
166
  )
167
 
168
+ try:
169
+ return tokenizer.decode(output.sequences[0], skip_special_tokens=True).strip()
170
+ except Exception as decode_error:
171
+ log(f"❌ Decode hatası: {decode_error}")
172
+ return random.choice(FALLBACK_ANSWERS)
173
 
174
  @app.post("/chat")
175
  async def chat(msg: Message):
 
196
  traceback.print_exc()
197
  return JSONResponse(content={"error": str(e)}, status_code=500)
198
 
199
+ def log(message):
200
+ timestamp = datetime.now().strftime("%H:%M:%S")
201
+ print(f"[{timestamp}] {message}", flush=True)
202
+
203
  def setup_model():
204
  global model, tokenizer, eos_token_id
205
  try:
206
+ log("🧠 setup_model() başladı")
207
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
208
+ log("📥 Tokenizer indiriliyor...")
209
  tokenizer = AutoTokenizer.from_pretrained(MODEL_BASE, use_fast=False)
210
+ log("🧠 Model indiriliyor...")
211
  model = AutoModelForCausalLM.from_pretrained(MODEL_BASE, torch_dtype=torch.float32).to(device)
212
  tokenizer.pad_token = tokenizer.pad_token or tokenizer.eos_token
213
  model.config.pad_token_id = tokenizer.pad_token_id
214
  eos_token_id = tokenizer("<|im_end|>", add_special_tokens=False)["input_ids"][0]
215
  model.eval()
216
+ log(" Model eval() çağrıldı")
217
+ log("✔️ Model başarıyla yüklendi ve sohbet için hazır.")
218
  except Exception as e:
219
  traceback.print_exc()
220