Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -125,27 +125,33 @@ def predict_image(image_url, text, file_pref):
|
|
| 125 |
]
|
| 126 |
}
|
| 127 |
],
|
| 128 |
-
temperature=0.7,
|
| 129 |
-
max_tokens=4096,
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 130 |
top_p=1,
|
| 131 |
-
stream=
|
|
|
|
| 132 |
)
|
| 133 |
|
| 134 |
-
for chunk in completion:
|
| 135 |
-
|
| 136 |
-
|
| 137 |
-
streamer.put(None) # Signal the end
|
| 138 |
except Exception as e:
|
| 139 |
print(f"Error in generate_response: {e}")
|
| 140 |
streamer.put(None)
|
| 141 |
|
| 142 |
-
thread = Thread(target=generate_response)
|
| 143 |
-
thread.start()
|
| 144 |
|
| 145 |
-
buffer = ""
|
| 146 |
-
for new_text in streamer:
|
| 147 |
-
|
| 148 |
-
|
| 149 |
return buffer
|
| 150 |
|
| 151 |
except Exception as e:
|
|
|
|
| 125 |
]
|
| 126 |
}
|
| 127 |
],
|
| 128 |
+
# temperature=0.7,
|
| 129 |
+
# max_tokens=4096,
|
| 130 |
+
# top_p=1,
|
| 131 |
+
# stream=True
|
| 132 |
+
|
| 133 |
+
temperature=1,
|
| 134 |
+
max_completion_tokens=4096,
|
| 135 |
top_p=1,
|
| 136 |
+
stream=False,
|
| 137 |
+
stop=None,
|
| 138 |
)
|
| 139 |
|
| 140 |
+
# for chunk in completion:
|
| 141 |
+
# if chunk.choices[0].delta.content:
|
| 142 |
+
# streamer.put(chunk.choices[0].delta.content)
|
| 143 |
+
# streamer.put(None) # Signal the end
|
| 144 |
except Exception as e:
|
| 145 |
print(f"Error in generate_response: {e}")
|
| 146 |
streamer.put(None)
|
| 147 |
|
| 148 |
+
# thread = Thread(target=generate_response)
|
| 149 |
+
# thread.start()
|
| 150 |
|
| 151 |
+
# buffer = ""
|
| 152 |
+
# for new_text in streamer:
|
| 153 |
+
# buffer += new_text
|
| 154 |
+
buffer = completion.choices[0].message
|
| 155 |
return buffer
|
| 156 |
|
| 157 |
except Exception as e:
|