Spaces:
Sleeping
Sleeping
Update run_eval.py
Browse files- run_eval.py +8 -6
run_eval.py
CHANGED
@@ -44,7 +44,7 @@ for cfg in CONFIGS:
|
|
44 |
adapter_type = cfg.get("adapter_type", "LoRA")
|
45 |
tasks = cfg["tasks"]
|
46 |
|
47 |
-
print(f"\
|
48 |
tokenizer = AutoTokenizer.from_pretrained(base_model_id, use_fast=True)
|
49 |
|
50 |
# Try causal first, fallback to encoder
|
@@ -110,14 +110,16 @@ with tempfile.TemporaryDirectory() as tmp:
|
|
110 |
df_existing = pd.read_parquet(current_path)
|
111 |
df_combined = pd.concat([df_existing, df_new], ignore_index=True)
|
112 |
|
113 |
-
df_combined = (
|
114 |
-
|
115 |
-
.sort_values("run_date")
|
116 |
-
.drop_duplicates(subset=["model_id", "task", "metric"], keep="last")
|
117 |
-
)
|
118 |
|
119 |
df_combined["value"] = pd.to_numeric(df_combined["value"], errors="coerce")
|
120 |
|
|
|
|
|
|
|
|
|
|
|
121 |
out = Path("peft_bench.parquet")
|
122 |
df_combined.to_parquet(out, index=False)
|
123 |
|
|
|
44 |
adapter_type = cfg.get("adapter_type", "LoRA")
|
45 |
tasks = cfg["tasks"]
|
46 |
|
47 |
+
print(f"\nLoading base model: {base_model_id}")
|
48 |
tokenizer = AutoTokenizer.from_pretrained(base_model_id, use_fast=True)
|
49 |
|
50 |
# Try causal first, fallback to encoder
|
|
|
110 |
df_existing = pd.read_parquet(current_path)
|
111 |
df_combined = pd.concat([df_existing, df_new], ignore_index=True)
|
112 |
|
113 |
+
df_combined = df_combined.sort_values("run_date")
|
114 |
+
|
|
|
|
|
|
|
115 |
|
116 |
df_combined["value"] = pd.to_numeric(df_combined["value"], errors="coerce")
|
117 |
|
118 |
+
print("Existing rows:", len(df_existing))
|
119 |
+
print("New rows:", len(df_new))
|
120 |
+
print("Combined rows (pre-dedup):", len(df_existing) + len(df_new))
|
121 |
+
print("Final rows (after dedup):", len(df_combined))
|
122 |
+
|
123 |
out = Path("peft_bench.parquet")
|
124 |
df_combined.to_parquet(out, index=False)
|
125 |
|