Update app.py
Browse files
app.py
CHANGED
@@ -32,7 +32,7 @@ def load_model():
|
|
32 |
model = AutoModelForCausalLM.from_pretrained(
|
33 |
model_name,
|
34 |
device_map="auto",
|
35 |
-
torch_dtype="auto"
|
36 |
# Eğer istersen load_in_8bit=True parametresini ekleyebiliriz
|
37 |
)
|
38 |
|
@@ -46,6 +46,7 @@ def load_model():
|
|
46 |
|
47 |
class UserInputRequest(BaseModel):
|
48 |
user_input: str
|
|
|
49 |
|
50 |
@app.post("/generate")
|
51 |
def generate(req: UserInputRequest):
|
@@ -53,13 +54,16 @@ def generate(req: UserInputRequest):
|
|
53 |
start_time = time.time()
|
54 |
log(f"💬 Kullanıcı isteği alındı: {req.user_input}")
|
55 |
|
|
|
|
|
|
|
56 |
result = pipe(
|
57 |
-
|
58 |
-
max_new_tokens=
|
59 |
-
temperature=0.0,
|
60 |
-
top_p=1.0,
|
61 |
-
repetition_penalty=1.0,
|
62 |
-
do_sample=False
|
63 |
)
|
64 |
answer = result[0]["generated_text"]
|
65 |
|
|
|
32 |
model = AutoModelForCausalLM.from_pretrained(
|
33 |
model_name,
|
34 |
device_map="auto",
|
35 |
+
torch_dtype="auto"
|
36 |
# Eğer istersen load_in_8bit=True parametresini ekleyebiliriz
|
37 |
)
|
38 |
|
|
|
46 |
|
47 |
class UserInputRequest(BaseModel):
|
48 |
user_input: str
|
49 |
+
system_prompt: str
|
50 |
|
51 |
@app.post("/generate")
|
52 |
def generate(req: UserInputRequest):
|
|
|
54 |
start_time = time.time()
|
55 |
log(f"💬 Kullanıcı isteği alındı: {req.user_input}")
|
56 |
|
57 |
+
# System prompt + user input birleştirme
|
58 |
+
full_prompt = f"{req.system_prompt}\n\n{req.user_input}"
|
59 |
+
|
60 |
result = pipe(
|
61 |
+
full_prompt,
|
62 |
+
max_new_tokens=200,
|
63 |
+
temperature=0.0,
|
64 |
+
top_p=1.0,
|
65 |
+
repetition_penalty=1.0,
|
66 |
+
do_sample=False
|
67 |
)
|
68 |
answer = result[0]["generated_text"]
|
69 |
|