Priyanshukr-1 commited on
Commit
6c97f34
·
verified ·
1 Parent(s): f747bda

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -166,7 +166,7 @@ async def generate(request: Request):
166
  try:
167
  response = llm.create_chat_completion(
168
  messages=messages_for_llm,
169
- max_tokens=1024, # Keep response length short for maximum speed
170
  temperature=0.7, # Adjust temperature for creativity vs. coherence (0.0-1.0)
171
  stop=["</s>"] # Stop sequence for TinyLlama Chat
172
  )
 
166
  try:
167
  response = llm.create_chat_completion(
168
  messages=messages_for_llm,
169
+ max_tokens=300, # Keep response length short for maximum speed
170
  temperature=0.7, # Adjust temperature for creativity vs. coherence (0.0-1.0)
171
  stop=["</s>"] # Stop sequence for TinyLlama Chat
172
  )