SkyNetWalker commited on
Commit
3087f19
·
verified ·
1 Parent(s): 5986c8e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -1
app.py CHANGED
@@ -100,6 +100,10 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
100
  and streams the response back to the chatbot.
101
  """
102
 
 
 
 
 
103
  # The 'history' variable from Gradio contains the entire conversation.
104
  # We prepend the system prompt to this history to form the final payload.
105
  messages = [{"role": "system", "content": system_prompt}] + history
@@ -112,7 +116,8 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
112
  response_stream = ollama.chat(
113
  model=current_selected_model, # Use the dynamically selected model
114
  messages=messages,
115
- stream=True
 
116
  )
117
 
118
  # Iterate through the stream, updating the placeholder with each new chunk.
 
100
  and streams the response back to the chatbot.
101
  """
102
 
103
+ #Disable Qwen3 thinking
104
+ #if "Qwen3".lower() in current_selected_model:
105
+ # system_prompt = system_prompt+" /no_think"
106
+
107
  # The 'history' variable from Gradio contains the entire conversation.
108
  # We prepend the system prompt to this history to form the final payload.
109
  messages = [{"role": "system", "content": system_prompt}] + history
 
116
  response_stream = ollama.chat(
117
  model=current_selected_model, # Use the dynamically selected model
118
  messages=messages,
119
+ stream=True,
120
+ think=False
121
  )
122
 
123
  # Iterate through the stream, updating the placeholder with each new chunk.