Spaces:
Running
Running
jasonshaoshun
commited on
Commit
·
531005f
1
Parent(s):
36438b0
debug
Browse files- app.py +6 -2
- src/display/utils.py +14 -7
app.py
CHANGED
@@ -125,7 +125,11 @@ def init_leaderboard_mib_causalgraph(dataframe, track):
|
|
125 |
if dataframe is None or dataframe.empty:
|
126 |
raise ValueError("Leaderboard DataFrame is empty or None.")
|
127 |
|
128 |
-
#
|
|
|
|
|
|
|
|
|
129 |
# dataframe = dataframe.loc[dataframe["Track"] == track]
|
130 |
|
131 |
print(f"init_leaderboard_mib: dataframe head after loc is {dataframe.head()}\n")
|
@@ -138,7 +142,7 @@ def init_leaderboard_mib_causalgraph(dataframe, track):
|
|
138 |
cant_deselect=[c.name for c in fields(AutoEvalColumn_mib_causalgraph) if c.never_hidden],
|
139 |
label="Select Columns to Display:",
|
140 |
),
|
141 |
-
search_columns=["Method"],
|
142 |
hide_columns=[c.name for c in fields(AutoEvalColumn_mib_causalgraph) if c.hidden],
|
143 |
bool_checkboxgroup_label="Hide models",
|
144 |
interactive=False,
|
|
|
125 |
if dataframe is None or dataframe.empty:
|
126 |
raise ValueError("Leaderboard DataFrame is empty or None.")
|
127 |
|
128 |
+
# Print all columns for debugging
|
129 |
+
print("Available columns in dataframe:", dataframe.columns.tolist())
|
130 |
+
print("Expected columns from AutoEvalColumn_mib_causalgraph:", [c.name for c in fields(AutoEvalColumn_mib_causalgraph) if not c.hidden])
|
131 |
+
|
132 |
+
# Remove this line since we don't need track filtering for causalgraph
|
133 |
# dataframe = dataframe.loc[dataframe["Track"] == track]
|
134 |
|
135 |
print(f"init_leaderboard_mib: dataframe head after loc is {dataframe.head()}\n")
|
|
|
142 |
cant_deselect=[c.name for c in fields(AutoEvalColumn_mib_causalgraph) if c.never_hidden],
|
143 |
label="Select Columns to Display:",
|
144 |
),
|
145 |
+
search_columns=["Method"],
|
146 |
hide_columns=[c.name for c in fields(AutoEvalColumn_mib_causalgraph) if c.hidden],
|
147 |
bool_checkboxgroup_label="Hide models",
|
148 |
interactive=False,
|
src/display/utils.py
CHANGED
@@ -169,28 +169,35 @@ auto_eval_column_dict_mib_causalgraph = []
|
|
169 |
# Method name column
|
170 |
auto_eval_column_dict_mib_causalgraph.append(["method", ColumnContent, ColumnContent("Method", "markdown", True, never_hidden=True)])
|
171 |
|
|
|
|
|
|
|
172 |
# For each model-task-intervention-counterfactual combination
|
173 |
for task in TasksMib_Causalgraph:
|
174 |
for model in task.value.models:
|
175 |
for layer in task.value.layers:
|
176 |
for intervention in task.value.interventions:
|
177 |
for counterfactual in task.value.counterfactuals:
|
178 |
-
# Match exact
|
179 |
col_name = f"{model}_layer{layer}_{intervention}_{counterfactual}"
|
|
|
180 |
auto_eval_column_dict_mib_causalgraph.append([
|
181 |
-
|
182 |
ColumnContent,
|
183 |
ColumnContent(col_name, "number", True)
|
184 |
])
|
185 |
|
186 |
-
|
187 |
-
|
188 |
-
|
189 |
# Create the dataclass
|
190 |
-
AutoEvalColumn_mib_causalgraph = make_dataclass(
|
|
|
|
|
|
|
|
|
191 |
|
192 |
-
# Column selection for display
|
193 |
COLS_MIB_CAUSALGRAPH = [c.name for c in fields(AutoEvalColumn_mib_causalgraph) if not c.hidden]
|
|
|
|
|
194 |
BENCHMARK_COLS_MIB_CAUSALGRAPH = [f"{model}_{task.value.benchmark}_{intervention}".lower()
|
195 |
for task in TasksMib_Causalgraph
|
196 |
for model in task.value.models
|
|
|
169 |
# Method name column
|
170 |
auto_eval_column_dict_mib_causalgraph.append(["method", ColumnContent, ColumnContent("Method", "markdown", True, never_hidden=True)])
|
171 |
|
172 |
+
# Add eval_name column
|
173 |
+
auto_eval_column_dict_mib_causalgraph.append(["eval_name", ColumnContent, ColumnContent("eval_name", "str", True)])
|
174 |
+
|
175 |
# For each model-task-intervention-counterfactual combination
|
176 |
for task in TasksMib_Causalgraph:
|
177 |
for model in task.value.models:
|
178 |
for layer in task.value.layers:
|
179 |
for intervention in task.value.interventions:
|
180 |
for counterfactual in task.value.counterfactuals:
|
181 |
+
# Match exact format from the actual data
|
182 |
col_name = f"{model}_layer{layer}_{intervention}_{counterfactual}"
|
183 |
+
field_name = col_name.lower() # Use lowercase for field name
|
184 |
auto_eval_column_dict_mib_causalgraph.append([
|
185 |
+
field_name,
|
186 |
ColumnContent,
|
187 |
ColumnContent(col_name, "number", True)
|
188 |
])
|
189 |
|
|
|
|
|
|
|
190 |
# Create the dataclass
|
191 |
+
AutoEvalColumn_mib_causalgraph = make_dataclass(
|
192 |
+
"AutoEvalColumn_mib_causalgraph",
|
193 |
+
auto_eval_column_dict_mib_causalgraph,
|
194 |
+
frozen=True
|
195 |
+
)
|
196 |
|
197 |
+
# Column selection for display
|
198 |
COLS_MIB_CAUSALGRAPH = [c.name for c in fields(AutoEvalColumn_mib_causalgraph) if not c.hidden]
|
199 |
+
|
200 |
+
|
201 |
BENCHMARK_COLS_MIB_CAUSALGRAPH = [f"{model}_{task.value.benchmark}_{intervention}".lower()
|
202 |
for task in TasksMib_Causalgraph
|
203 |
for model in task.value.models
|