AustingDong
commited on
Commit
·
2bbc5f6
1
Parent(s):
8b5e432
Update evaluate.py
Browse files- evaluate/evaluate.py +14 -1
evaluate/evaluate.py
CHANGED
@@ -10,8 +10,20 @@ def set_seed(model_seed = 42):
|
|
10 |
np.random.seed(model_seed)
|
11 |
torch.cuda.manual_seed(model_seed) if torch.cuda.is_available() else None
|
12 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
def evaluate(model_type, num_eval = 10):
|
14 |
for eval_idx in range(num_eval):
|
|
|
15 |
set_seed(np.random.randint(0, 1000))
|
16 |
model_utils, vl_gpt, tokenizer = None, None, None
|
17 |
|
@@ -61,6 +73,7 @@ if __name__ == '__main__':
|
|
61 |
|
62 |
# models = ["ChartGemma", "Janus-Pro-1B", "Janus-Pro-7B", "LLaVA-1.5-7B"]
|
63 |
# models = ["ChartGemma", "Janus-Pro-1B"]
|
64 |
-
models = ["Janus-Pro-7B", "LLaVA-1.5-7B"]
|
|
|
65 |
for model_type in models:
|
66 |
evaluate(model_type=model_type, num_eval=10)
|
|
|
10 |
np.random.seed(model_seed)
|
11 |
torch.cuda.manual_seed(model_seed) if torch.cuda.is_available() else None
|
12 |
|
13 |
+
def clean():
|
14 |
+
# Empty CUDA cache
|
15 |
+
if torch.cuda.is_available():
|
16 |
+
torch.cuda.empty_cache()
|
17 |
+
torch.cuda.ipc_collect() # Frees inter-process CUDA memory
|
18 |
+
|
19 |
+
# Empty MacOS Metal backend (if using Apple Silicon)
|
20 |
+
if torch.backends.mps.is_available():
|
21 |
+
torch.mps.empty_cache()
|
22 |
+
|
23 |
+
|
24 |
def evaluate(model_type, num_eval = 10):
|
25 |
for eval_idx in range(num_eval):
|
26 |
+
clean()
|
27 |
set_seed(np.random.randint(0, 1000))
|
28 |
model_utils, vl_gpt, tokenizer = None, None, None
|
29 |
|
|
|
73 |
|
74 |
# models = ["ChartGemma", "Janus-Pro-1B", "Janus-Pro-7B", "LLaVA-1.5-7B"]
|
75 |
# models = ["ChartGemma", "Janus-Pro-1B"]
|
76 |
+
# models = ["Janus-Pro-7B", "LLaVA-1.5-7B"]
|
77 |
+
models = ["LLaVA-1.5-7B"]
|
78 |
for model_type in models:
|
79 |
evaluate(model_type=model_type, num_eval=10)
|