Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -101,8 +101,8 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
|
|
101 |
"""
|
102 |
|
103 |
#Disable Qwen3 thinking
|
104 |
-
|
105 |
-
|
106 |
|
107 |
# The 'history' variable from Gradio contains the entire conversation.
|
108 |
# We prepend the system prompt to this history to form the final payload.
|
@@ -116,8 +116,7 @@ with gr.Blocks(theme=gr.themes.Default(primary_hue="blue", secondary_hue="neutra
|
|
116 |
response_stream = ollama.chat(
|
117 |
model=current_selected_model, # Use the dynamically selected model
|
118 |
messages=messages,
|
119 |
-
stream=True
|
120 |
-
think=False
|
121 |
)
|
122 |
|
123 |
# Iterate through the stream, updating the placeholder with each new chunk.
|
|
|
101 |
"""
|
102 |
|
103 |
#Disable Qwen3 thinking
|
104 |
+
if "Qwen3".lower() in current_selected_model:
|
105 |
+
system_prompt = system_prompt+" /no_think"
|
106 |
|
107 |
# The 'history' variable from Gradio contains the entire conversation.
|
108 |
# We prepend the system prompt to this history to form the final payload.
|
|
|
116 |
response_stream = ollama.chat(
|
117 |
model=current_selected_model, # Use the dynamically selected model
|
118 |
messages=messages,
|
119 |
+
stream=True
|
|
|
120 |
)
|
121 |
|
122 |
# Iterate through the stream, updating the placeholder with each new chunk.
|