Spaces:
Running
Running
jasonshaoshun
commited on
Commit
·
85b6004
1
Parent(s):
0810300
debug
Browse files- app.py +5 -4
- src/display/utils.py +1 -1
app.py
CHANGED
@@ -405,20 +405,21 @@ def init_leaderboard_mib_subgraph(dataframe, track):
|
|
405 |
# Extract unique benchmarks and models from column names
|
406 |
for col in result_columns:
|
407 |
print(f"col is {col}")
|
408 |
-
benchmark, model = col.split('
|
409 |
benchmarks.add(benchmark)
|
410 |
-
models.add(model)
|
|
|
411 |
|
412 |
# Create selection groups
|
413 |
benchmark_selections = {
|
414 |
# For each benchmark, store which columns should be shown
|
415 |
-
benchmark: [col for col in result_columns if col.startswith(f"{benchmark}
|
416 |
for benchmark in benchmarks
|
417 |
}
|
418 |
|
419 |
model_selections = {
|
420 |
# For each model, store which columns should be shown
|
421 |
-
model: [col for col in result_columns if col.endswith(f"
|
422 |
for model in models
|
423 |
}
|
424 |
|
|
|
405 |
# Extract unique benchmarks and models from column names
|
406 |
for col in result_columns:
|
407 |
print(f"col is {col}")
|
408 |
+
benchmark, model = col.split('(')
|
409 |
benchmarks.add(benchmark)
|
410 |
+
models.add(model[:-1])
|
411 |
+
print(f"benchmark is {benchmark} and model is {model}")
|
412 |
|
413 |
# Create selection groups
|
414 |
benchmark_selections = {
|
415 |
# For each benchmark, store which columns should be shown
|
416 |
+
benchmark: [col for col in result_columns if col.startswith(f"{benchmark}(")]
|
417 |
for benchmark in benchmarks
|
418 |
}
|
419 |
|
420 |
model_selections = {
|
421 |
# For each model, store which columns should be shown
|
422 |
+
model: [col for col in result_columns if col.endswith(f"({model})")]
|
423 |
for model in models
|
424 |
}
|
425 |
|
src/display/utils.py
CHANGED
@@ -131,7 +131,7 @@ auto_eval_column_dict_mib_subgraph.append(["Model", ColumnContent, ColumnContent
|
|
131 |
# For each task and model combination
|
132 |
for task in TasksMib_Subgraph:
|
133 |
for model in task.value.models:
|
134 |
-
col_name = f"{task.value.benchmark}
|
135 |
auto_eval_column_dict_mib_subgraph.append([
|
136 |
col_name,
|
137 |
ColumnContent,
|
|
|
131 |
# For each task and model combination
|
132 |
for task in TasksMib_Subgraph:
|
133 |
for model in task.value.models:
|
134 |
+
col_name = f"{task.value.benchmark}({model})"
|
135 |
auto_eval_column_dict_mib_subgraph.append([
|
136 |
col_name,
|
137 |
ColumnContent,
|