Spaces:
Running
Running
Clémentine
commited on
Commit
·
fc1e99b
1
Parent(s):
df66f6e
fix col width
Browse files- app.py +41 -43
- requirements.txt +3 -3
- src/leaderboard/read_evals.py +1 -0
- src/submission/check_validity.py +2 -0
app.py
CHANGED
|
@@ -73,14 +73,13 @@ leaderboard_df = original_df.copy()
|
|
| 73 |
|
| 74 |
|
| 75 |
# Basics
|
| 76 |
-
def change_tab(query_param: str):
|
| 77 |
-
query_param = query_param.replace("'", '"')
|
| 78 |
-
|
| 79 |
-
|
| 80 |
-
|
| 81 |
-
|
| 82 |
-
|
| 83 |
-
return gr.Tabs.update(selected=0)
|
| 84 |
|
| 85 |
|
| 86 |
# Searching and filtering
|
|
@@ -192,28 +191,28 @@ with demo:
|
|
| 192 |
value=False, label="Show gated/private/deleted models", interactive=True
|
| 193 |
)
|
| 194 |
with gr.Column(min_width=320):
|
| 195 |
-
with gr.Box(elem_id="box-filter"):
|
| 196 |
-
|
| 197 |
-
|
| 198 |
-
|
| 199 |
-
|
| 200 |
-
|
| 201 |
-
|
| 202 |
-
|
| 203 |
-
|
| 204 |
-
|
| 205 |
-
|
| 206 |
-
|
| 207 |
-
|
| 208 |
-
|
| 209 |
-
|
| 210 |
-
|
| 211 |
-
|
| 212 |
-
|
| 213 |
-
|
| 214 |
-
|
| 215 |
-
|
| 216 |
-
|
| 217 |
|
| 218 |
leaderboard_table = gr.components.Dataframe(
|
| 219 |
value=leaderboard_df[
|
|
@@ -223,10 +222,10 @@ with demo:
|
|
| 223 |
],
|
| 224 |
headers=[c.name for c in fields(AutoEvalColumn) if c.never_hidden] + shown_columns.value,
|
| 225 |
datatype=TYPES,
|
| 226 |
-
max_rows=None,
|
| 227 |
elem_id="leaderboard-table",
|
| 228 |
interactive=False,
|
| 229 |
visible=True,
|
|
|
|
| 230 |
)
|
| 231 |
|
| 232 |
# Dummy leaderboard for handling the case when the user uses backspace key
|
|
@@ -234,7 +233,6 @@ with demo:
|
|
| 234 |
value=original_df[COLS],
|
| 235 |
headers=COLS,
|
| 236 |
datatype=TYPES,
|
| 237 |
-
max_rows=None,
|
| 238 |
visible=False,
|
| 239 |
)
|
| 240 |
search_bar.submit(
|
|
@@ -358,7 +356,7 @@ with demo:
|
|
| 358 |
value=finished_eval_queue_df,
|
| 359 |
headers=EVAL_COLS,
|
| 360 |
datatype=EVAL_TYPES,
|
| 361 |
-
|
| 362 |
)
|
| 363 |
with gr.Accordion(
|
| 364 |
f"🔄 Running Evaluation Queue ({len(running_eval_queue_df)})",
|
|
@@ -369,7 +367,7 @@ with demo:
|
|
| 369 |
value=running_eval_queue_df,
|
| 370 |
headers=EVAL_COLS,
|
| 371 |
datatype=EVAL_TYPES,
|
| 372 |
-
|
| 373 |
)
|
| 374 |
|
| 375 |
with gr.Accordion(
|
|
@@ -381,7 +379,7 @@ with demo:
|
|
| 381 |
value=pending_eval_queue_df,
|
| 382 |
headers=EVAL_COLS,
|
| 383 |
datatype=EVAL_TYPES,
|
| 384 |
-
|
| 385 |
)
|
| 386 |
with gr.Row():
|
| 387 |
gr.Markdown("# ✉️✨ Submit your model here!", elem_classes="markdown-text")
|
|
@@ -442,15 +440,15 @@ with demo:
|
|
| 442 |
show_copy_button=True,
|
| 443 |
)
|
| 444 |
|
| 445 |
-
dummy = gr.Textbox(visible=False)
|
| 446 |
-
demo.load(
|
| 447 |
-
|
| 448 |
-
|
| 449 |
-
|
| 450 |
-
|
| 451 |
-
)
|
| 452 |
|
| 453 |
scheduler = BackgroundScheduler()
|
| 454 |
scheduler.add_job(restart_space, "interval", seconds=1800)
|
| 455 |
scheduler.start()
|
| 456 |
-
demo.queue(
|
|
|
|
| 73 |
|
| 74 |
|
| 75 |
# Basics
|
| 76 |
+
#def change_tab(query_param: str):
|
| 77 |
+
# query_param = query_param.replace("'", '"')
|
| 78 |
+
# query_param = json.loads(query_param)
|
| 79 |
+
# if isinstance(query_param, dict) and "tab" in query_param and query_param["tab"] == "evaluation":
|
| 80 |
+
# return gr.Tabs.update(selected=1)
|
| 81 |
+
# else:
|
| 82 |
+
# return gr.Tabs.update(selected=0)
|
|
|
|
| 83 |
|
| 84 |
|
| 85 |
# Searching and filtering
|
|
|
|
| 191 |
value=False, label="Show gated/private/deleted models", interactive=True
|
| 192 |
)
|
| 193 |
with gr.Column(min_width=320):
|
| 194 |
+
#with gr.Box(elem_id="box-filter"):
|
| 195 |
+
filter_columns_type = gr.CheckboxGroup(
|
| 196 |
+
label="Model types",
|
| 197 |
+
choices=[t.to_str() for t in ModelType],
|
| 198 |
+
value=[t.to_str() for t in ModelType],
|
| 199 |
+
interactive=True,
|
| 200 |
+
elem_id="filter-columns-type",
|
| 201 |
+
)
|
| 202 |
+
filter_columns_precision = gr.CheckboxGroup(
|
| 203 |
+
label="Precision",
|
| 204 |
+
choices=["torch.float16", "torch.bfloat16", "torch.float32", "8bit", "4bit", "GPTQ"],
|
| 205 |
+
value=["torch.float16", "torch.bfloat16", "torch.float32", "8bit", "4bit", "GPTQ"],
|
| 206 |
+
interactive=True,
|
| 207 |
+
elem_id="filter-columns-precision",
|
| 208 |
+
)
|
| 209 |
+
filter_columns_size = gr.CheckboxGroup(
|
| 210 |
+
label="Model sizes (in billions of parameters)",
|
| 211 |
+
choices=list(NUMERIC_INTERVALS.keys()),
|
| 212 |
+
value=list(NUMERIC_INTERVALS.keys()),
|
| 213 |
+
interactive=True,
|
| 214 |
+
elem_id="filter-columns-size",
|
| 215 |
+
)
|
| 216 |
|
| 217 |
leaderboard_table = gr.components.Dataframe(
|
| 218 |
value=leaderboard_df[
|
|
|
|
| 222 |
],
|
| 223 |
headers=[c.name for c in fields(AutoEvalColumn) if c.never_hidden] + shown_columns.value,
|
| 224 |
datatype=TYPES,
|
|
|
|
| 225 |
elem_id="leaderboard-table",
|
| 226 |
interactive=False,
|
| 227 |
visible=True,
|
| 228 |
+
column_widths=["2%", "33%"]
|
| 229 |
)
|
| 230 |
|
| 231 |
# Dummy leaderboard for handling the case when the user uses backspace key
|
|
|
|
| 233 |
value=original_df[COLS],
|
| 234 |
headers=COLS,
|
| 235 |
datatype=TYPES,
|
|
|
|
| 236 |
visible=False,
|
| 237 |
)
|
| 238 |
search_bar.submit(
|
|
|
|
| 356 |
value=finished_eval_queue_df,
|
| 357 |
headers=EVAL_COLS,
|
| 358 |
datatype=EVAL_TYPES,
|
| 359 |
+
row_count=5,
|
| 360 |
)
|
| 361 |
with gr.Accordion(
|
| 362 |
f"🔄 Running Evaluation Queue ({len(running_eval_queue_df)})",
|
|
|
|
| 367 |
value=running_eval_queue_df,
|
| 368 |
headers=EVAL_COLS,
|
| 369 |
datatype=EVAL_TYPES,
|
| 370 |
+
row_count=5,
|
| 371 |
)
|
| 372 |
|
| 373 |
with gr.Accordion(
|
|
|
|
| 379 |
value=pending_eval_queue_df,
|
| 380 |
headers=EVAL_COLS,
|
| 381 |
datatype=EVAL_TYPES,
|
| 382 |
+
row_count=5,
|
| 383 |
)
|
| 384 |
with gr.Row():
|
| 385 |
gr.Markdown("# ✉️✨ Submit your model here!", elem_classes="markdown-text")
|
|
|
|
| 440 |
show_copy_button=True,
|
| 441 |
)
|
| 442 |
|
| 443 |
+
#dummy = gr.Textbox(visible=False)
|
| 444 |
+
#demo.load(
|
| 445 |
+
# change_tab,
|
| 446 |
+
# dummy,
|
| 447 |
+
# tabs,
|
| 448 |
+
# js=get_window_url_params,
|
| 449 |
+
#)
|
| 450 |
|
| 451 |
scheduler = BackgroundScheduler()
|
| 452 |
scheduler.add_job(restart_space, "interval", seconds=1800)
|
| 453 |
scheduler.start()
|
| 454 |
+
demo.queue().launch()
|
requirements.txt
CHANGED
|
@@ -3,7 +3,7 @@ aiofiles==23.1.0
|
|
| 3 |
aiohttp==3.8.4
|
| 4 |
aiosignal==1.3.1
|
| 5 |
altair==4.2.2
|
| 6 |
-
anyio==3.
|
| 7 |
APScheduler==3.10.1
|
| 8 |
async-timeout==4.0.2
|
| 9 |
attrs==23.1.0
|
|
@@ -20,8 +20,8 @@ filelock==3.11.0
|
|
| 20 |
fonttools==4.39.3
|
| 21 |
frozenlist==1.3.3
|
| 22 |
fsspec==2023.5.0
|
| 23 |
-
gradio==3.
|
| 24 |
-
gradio-client==0.
|
| 25 |
h11==0.14.0
|
| 26 |
httpcore==0.17.0
|
| 27 |
httpx==0.24.0
|
|
|
|
| 3 |
aiohttp==3.8.4
|
| 4 |
aiosignal==1.3.1
|
| 5 |
altair==4.2.2
|
| 6 |
+
anyio==3.7.1
|
| 7 |
APScheduler==3.10.1
|
| 8 |
async-timeout==4.0.2
|
| 9 |
attrs==23.1.0
|
|
|
|
| 20 |
fonttools==4.39.3
|
| 21 |
frozenlist==1.3.3
|
| 22 |
fsspec==2023.5.0
|
| 23 |
+
gradio==4.3.0
|
| 24 |
+
gradio-client==0.7.0
|
| 25 |
h11==0.14.0
|
| 26 |
httpcore==0.17.0
|
| 27 |
httpx==0.24.0
|
src/leaderboard/read_evals.py
CHANGED
|
@@ -24,6 +24,7 @@ class EvalResult:
|
|
| 24 |
precision: str = ""
|
| 25 |
model_type: ModelType = ModelType.Unknown
|
| 26 |
weight_type: str = "Original"
|
|
|
|
| 27 |
license: str = "?"
|
| 28 |
likes: int = 0
|
| 29 |
num_params: int = 0
|
|
|
|
| 24 |
precision: str = ""
|
| 25 |
model_type: ModelType = ModelType.Unknown
|
| 26 |
weight_type: str = "Original"
|
| 27 |
+
architecture: str = "Unknown"
|
| 28 |
license: str = "?"
|
| 29 |
likes: int = 0
|
| 30 |
num_params: int = 0
|
src/submission/check_validity.py
CHANGED
|
@@ -67,6 +67,8 @@ def get_model_size(model_info: ModelInfo, precision: str):
|
|
| 67 |
model_size = size_factor * model_size
|
| 68 |
return model_size
|
| 69 |
|
|
|
|
|
|
|
| 70 |
|
| 71 |
def user_submission_permission(submission_name, users_to_submission_dates, rate_limit_period, rate_limit_quota):
|
| 72 |
org_or_user, _ = submission_name.split("/")
|
|
|
|
| 67 |
model_size = size_factor * model_size
|
| 68 |
return model_size
|
| 69 |
|
| 70 |
+
def get_model_arch(model_info: ModelInfo):
|
| 71 |
+
return model_info.config.get("architectures", "Unknown")
|
| 72 |
|
| 73 |
def user_submission_permission(submission_name, users_to_submission_dates, rate_limit_period, rate_limit_quota):
|
| 74 |
org_or_user, _ = submission_name.split("/")
|