Spaces:
Running
on
Zero
Running
on
Zero
Update llm.py
Browse files
llm.py
CHANGED
@@ -82,7 +82,7 @@ def build_messages(
|
|
82 |
@torch.inference_mode()
|
83 |
def stream_response(
|
84 |
messages: list[dict],
|
85 |
-
enable_thinking: bool=
|
86 |
max_new_tokens: int=1_024
|
87 |
) -> tuple[str, str]:
|
88 |
"""Stream the model's response to the chat interface.
|
|
|
82 |
@torch.inference_mode()
|
83 |
def stream_response(
|
84 |
messages: list[dict],
|
85 |
+
enable_thinking: bool=False,
|
86 |
max_new_tokens: int=1_024
|
87 |
) -> tuple[str, str]:
|
88 |
"""Stream the model's response to the chat interface.
|