Update fine_tune_inference_test.py
Browse files
fine_tune_inference_test.py
CHANGED
@@ -67,6 +67,7 @@ def root():
|
|
67 |
@app.post("/chat")
|
68 |
def chat(msg: Message):
|
69 |
try:
|
|
|
70 |
global pipe
|
71 |
if pipe is None:
|
72 |
log("🚫 Hata: Model henüz yüklenmedi.")
|
@@ -81,6 +82,7 @@ def chat(msg: Message):
|
|
81 |
full_prompt += f"Kullanıcı: {turn['user']}\nAsistan: {turn['bot']}\n"
|
82 |
full_prompt += f"Kullanıcı: {user_input}\nAsistan:"
|
83 |
|
|
|
84 |
result = pipe(full_prompt, max_new_tokens=200, do_sample=True, temperature=0.7)
|
85 |
answer = result[0]["generated_text"][len(full_prompt):].strip()
|
86 |
chat_history.append({"user": user_input, "bot": answer})
|
|
|
67 |
@app.post("/chat")
|
68 |
def chat(msg: Message):
|
69 |
try:
|
70 |
+
log(f"📦 Kullanıcı mesajı alındı: {msg}")
|
71 |
global pipe
|
72 |
if pipe is None:
|
73 |
log("🚫 Hata: Model henüz yüklenmedi.")
|
|
|
82 |
full_prompt += f"Kullanıcı: {turn['user']}\nAsistan: {turn['bot']}\n"
|
83 |
full_prompt += f"Kullanıcı: {user_input}\nAsistan:"
|
84 |
|
85 |
+
log("📦 Cevap hazırlanıyor...")
|
86 |
result = pipe(full_prompt, max_new_tokens=200, do_sample=True, temperature=0.7)
|
87 |
answer = result[0]["generated_text"][len(full_prompt):].strip()
|
88 |
chat_history.append({"user": user_input, "bot": answer})
|