ciyidogan commited on
Commit
f8a28b3
·
verified ·
1 Parent(s): ce706b9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -7
app.py CHANGED
@@ -32,7 +32,7 @@ def load_model():
32
  model = AutoModelForCausalLM.from_pretrained(
33
  model_name,
34
  device_map="auto",
35
- torch_dtype="auto" # A100 ortamında bf16 otomatik seçer
36
  # Eğer istersen load_in_8bit=True parametresini ekleyebiliriz
37
  )
38
 
@@ -46,6 +46,7 @@ def load_model():
46
 
47
  class UserInputRequest(BaseModel):
48
  user_input: str
 
49
 
50
  @app.post("/generate")
51
  def generate(req: UserInputRequest):
@@ -53,13 +54,16 @@ def generate(req: UserInputRequest):
53
  start_time = time.time()
54
  log(f"💬 Kullanıcı isteği alındı: {req.user_input}")
55
 
 
 
 
56
  result = pipe(
57
- req.user_input,
58
- max_new_tokens=50, # Azaltıldı
59
- temperature=0.0, # Deterministic
60
- top_p=1.0, # Serbest sampling
61
- repetition_penalty=1.0, # Doğal tekrarlar izinli
62
- do_sample=False # Sampling kapalı, deterministik
63
  )
64
  answer = result[0]["generated_text"]
65
 
 
32
  model = AutoModelForCausalLM.from_pretrained(
33
  model_name,
34
  device_map="auto",
35
+ torch_dtype="auto"
36
  # Eğer istersen load_in_8bit=True parametresini ekleyebiliriz
37
  )
38
 
 
46
 
47
  class UserInputRequest(BaseModel):
48
  user_input: str
49
+ system_prompt: str
50
 
51
  @app.post("/generate")
52
  def generate(req: UserInputRequest):
 
54
  start_time = time.time()
55
  log(f"💬 Kullanıcı isteği alındı: {req.user_input}")
56
 
57
+ # System prompt + user input birleştirme
58
+ full_prompt = f"{req.system_prompt}\n\n{req.user_input}"
59
+
60
  result = pipe(
61
+ full_prompt,
62
+ max_new_tokens=200,
63
+ temperature=0.0,
64
+ top_p=1.0,
65
+ repetition_penalty=1.0,
66
+ do_sample=False
67
  )
68
  answer = result[0]["generated_text"]
69