Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -64,19 +64,25 @@ external_log_dir = "./logs"
|
|
64 |
LOGDIR = external_log_dir
|
65 |
VOTEDIR = "./votes"
|
66 |
|
|
|
|
|
|
|
|
|
67 |
|
68 |
-
|
69 |
-
|
70 |
-
|
71 |
-
|
72 |
-
|
73 |
-
|
74 |
-
|
75 |
-
|
76 |
-
|
|
|
|
|
77 |
|
78 |
# Call the function to install Gradio 4.35.0 if needed
|
79 |
-
install_gradio_4_35_0()
|
80 |
|
81 |
import gradio as gr
|
82 |
import gradio_client
|
@@ -487,7 +493,8 @@ with gr.Blocks(
|
|
487 |
label="Max output tokens",
|
488 |
)
|
489 |
with gr.Row():
|
490 |
-
chatbot = gr.Chatbot([], elem_id="Pangea", bubble_full_width=False, height=750)
|
|
|
491 |
|
492 |
with gr.Row():
|
493 |
upvote_btn = gr.Button(value="👍 Upvote", interactive=True)
|
@@ -734,4 +741,13 @@ if __name__ == "__main__":
|
|
734 |
device = (torch.device("cuda") if torch.cuda.is_available() else torch.device("mps") if getattr(torch.backends, "mps", None) and torch.backends.mps.is_available() else torch.device("cpu"))
|
735 |
model = model.to(device)
|
736 |
chat_image_num = 0
|
737 |
-
demo.launch()
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
LOGDIR = external_log_dir
|
65 |
VOTEDIR = "./votes"
|
66 |
|
67 |
+
import gradio, uvicorn, websockets, pathlib
|
68 |
+
print("Gradio:", gradio.__version__)
|
69 |
+
print("uvicorn:", uvicorn.__version__)
|
70 |
+
print("websockets:", websockets.__version__)
|
71 |
|
72 |
+
|
73 |
+
|
74 |
+
# def install_gradio_4_35_0():
|
75 |
+
# current_version = gr.__version__
|
76 |
+
# if current_version != "4.35.0":
|
77 |
+
# print(f"Current Gradio version: {current_version}")
|
78 |
+
# print("Installing Gradio 4.35.0...")
|
79 |
+
# subprocess.check_call([sys.executable, "-m", "pip", "install", "gradio==4.35.0", "--force-reinstall"])
|
80 |
+
# print("Gradio 4.35.0 installed successfully.")
|
81 |
+
# else:
|
82 |
+
# print("Gradio 4.35.0 is already installed.")
|
83 |
|
84 |
# Call the function to install Gradio 4.35.0 if needed
|
85 |
+
# install_gradio_4_35_0()
|
86 |
|
87 |
import gradio as gr
|
88 |
import gradio_client
|
|
|
493 |
label="Max output tokens",
|
494 |
)
|
495 |
with gr.Row():
|
496 |
+
# chatbot = gr.Chatbot([], elem_id="Pangea", bubble_full_width=False, height=750)
|
497 |
+
chatbot = gr.Chatbot([], elem_id="Pangea", type="tuple", height=750)
|
498 |
|
499 |
with gr.Row():
|
500 |
upvote_btn = gr.Button(value="👍 Upvote", interactive=True)
|
|
|
741 |
device = (torch.device("cuda") if torch.cuda.is_available() else torch.device("mps") if getattr(torch.backends, "mps", None) and torch.backends.mps.is_available() else torch.device("cpu"))
|
742 |
model = model.to(device)
|
743 |
chat_image_num = 0
|
744 |
+
# demo.launch()
|
745 |
+
demo.queue()
|
746 |
+
|
747 |
+
# Spaces sets $PORT; Gradio also respects GRADIO_SERVER_PORT if you prefer.
|
748 |
+
port = int(os.environ.get("PORT", os.environ.get("GRADIO_SERVER_PORT", 7860)))
|
749 |
+
|
750 |
+
demo.launch(
|
751 |
+
server_name="0.0.0.0",
|
752 |
+
server_port=port,
|
753 |
+
)
|