Spaces:
Paused
Paused
Update app_chat.py
Browse files- app_chat.py +6 -2
app_chat.py
CHANGED
|
@@ -9,8 +9,9 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStream
|
|
| 9 |
from transformers import StoppingCriteria, StoppingCriteriaList, StopStringCriteria
|
| 10 |
|
| 11 |
import subprocess
|
| 12 |
-
|
| 13 |
-
|
|
|
|
| 14 |
import torch._dynamo
|
| 15 |
torch._dynamo.config.suppress_errors = True
|
| 16 |
|
|
@@ -57,6 +58,9 @@ def generate(
|
|
| 57 |
repetition_penalty: float = 1.2,
|
| 58 |
) -> Iterator[str]:
|
| 59 |
conversation = []
|
|
|
|
|
|
|
|
|
|
| 60 |
if system_prompt:
|
| 61 |
conversation.append({"role": "system", "content": system_prompt})
|
| 62 |
conversation += chat_history
|
|
|
|
| 9 |
from transformers import StoppingCriteria, StoppingCriteriaList, StopStringCriteria
|
| 10 |
|
| 11 |
import subprocess
|
| 12 |
+
|
| 13 |
+
global model
|
| 14 |
+
|
| 15 |
import torch._dynamo
|
| 16 |
torch._dynamo.config.suppress_errors = True
|
| 17 |
|
|
|
|
| 58 |
repetition_penalty: float = 1.2,
|
| 59 |
) -> Iterator[str]:
|
| 60 |
conversation = []
|
| 61 |
+
|
| 62 |
+
global model
|
| 63 |
+
|
| 64 |
if system_prompt:
|
| 65 |
conversation.append({"role": "system", "content": system_prompt})
|
| 66 |
conversation += chat_history
|