Spaces:
Running
Running
ycy
commited on
Commit
·
360820c
1
Parent(s):
589c39f
- app.py +1 -1
- src/about.py +5 -38
- src/display/utils.py +10 -10
- src/leaderboard/read_evals.py +9 -9
app.py
CHANGED
@@ -86,7 +86,7 @@ def init_leaderboard(dataframe):
|
|
86 |
label="Select the number of parameters (B)",
|
87 |
),
|
88 |
ColumnFilter(
|
89 |
-
AutoEvalColumn.still_on_hub.name, type="boolean", label="
|
90 |
),
|
91 |
],
|
92 |
bool_checkboxgroup_label="Hide models",
|
|
|
86 |
label="Select the number of parameters (B)",
|
87 |
),
|
88 |
ColumnFilter(
|
89 |
+
AutoEvalColumn.still_on_hub.name, type="boolean", label="Available on the hub", default=False
|
90 |
),
|
91 |
],
|
92 |
bool_checkboxgroup_label="Hide models",
|
src/about.py
CHANGED
@@ -27,50 +27,17 @@ NUM_FEWSHOT = 0 # Change with your few shot
|
|
27 |
#TODO title
|
28 |
TITLE = """<h1 align="center" id="space-title">VLM Detailed Caption Leaderborad</h1>"""
|
29 |
|
30 |
-
|
31 |
-
|
32 |
-
#TODO interoduction
|
33 |
INTRODUCTION_TEXT = """
|
34 |
<div style="display: flex; flex-wrap: wrap; gap: 10px;">
|
35 |
-
|
36 |
-
|
37 |
-
<img src="https://img.shields.io/github/v/release/yanchenyang123/DeepSeek-V3?style=flat-square" alt="release">
|
38 |
-
</a>
|
39 |
-
<!-- Docker badge -->
|
40 |
-
<a href="https://hub.docker.com/r/your-docker-image" target="_blank">
|
41 |
-
<img src="https://img.shields.io/badge/docker-v1.3.1-blue?style=flat-square" alt="docker">
|
42 |
-
</a>
|
43 |
-
<!-- Stars badge -->
|
44 |
-
<a href="https://github.com/yanchenyang123/DeepSeek-V3/stargazers" target="_blank">
|
45 |
-
<img src="https://img.shields.io/github/stars/yanchenyang123/DeepSeek-V3?style=flat-square" alt="stars">
|
46 |
-
</a>
|
47 |
-
<!-- Issues badge -->
|
48 |
<a href="https://github.com/yanchenyang123/DeepSeek-V3/issues" target="_blank">
|
49 |
-
<img src="https://img.shields.io/
|
50 |
-
</a>
|
51 |
-
<!-- Contributors badge -->
|
52 |
-
<a href="https://github.com/yanchenyang123/DeepSeek-V3/graphs/contributors" target="_blank">
|
53 |
-
<img src="https://img.shields.io/github/contributors/yanchenyang123/DeepSeek-V3?color=yellowgreen&style=flat-square" alt="contributors">
|
54 |
-
</a>
|
55 |
-
<!-- Forks badge -->
|
56 |
-
<a href="https://github.com/yanchenyang123/DeepSeek-V3/network/members" target="_blank">
|
57 |
-
<img src="https://img.shields.io/github/forks/yanchenyang123/DeepSeek-V3?color=cyan&style=flat-square" alt="forks">
|
58 |
-
</a>
|
59 |
-
<!-- License badge -->
|
60 |
-
<a href="https://github.com/yanchenyang123/DeepSeek-V3/blob/main/LICENSE" target="_blank">
|
61 |
-
<img src="https://img.shields.io/github/license/yanchenyang123/DeepSeek-V3?color=black&style=flat-square" alt="license">
|
62 |
-
</a>
|
63 |
-
<!-- WeChat badge -->
|
64 |
-
<a href="https://your-wechat-link.com" target="_blank">
|
65 |
-
<img src="https://img.shields.io/badge/WeChat-微信-green?style=flat-square" alt="WeChat">
|
66 |
-
</a>
|
67 |
-
<!-- Hugging Face Spaces badge -->
|
68 |
-
<a href="https://huggingface.co/spaces/your-space" target="_blank">
|
69 |
-
<img src="https://img.shields.io/badge/Open%20in%20Spaces-%F0%9F%A4%AF-orange?style=flat-square" alt="Open in Spaces">
|
70 |
</a>
|
|
|
71 |
</div>
|
72 |
"""
|
73 |
-
|
74 |
#TODO About
|
75 |
LLM_BENCHMARKS_TEXT = f"""
|
76 |
<h2 style="color: #2c3e50;"> Why Leaderboard? </h2>
|
|
|
27 |
#TODO title
|
28 |
TITLE = """<h1 align="center" id="space-title">VLM Detailed Caption Leaderborad</h1>"""
|
29 |
|
30 |
+
# introduction text
|
|
|
|
|
31 |
INTRODUCTION_TEXT = """
|
32 |
<div style="display: flex; flex-wrap: wrap; gap: 10px;">
|
33 |
+
|
34 |
+
<!-- Paper badge with custom icon -->
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
35 |
<a href="https://github.com/yanchenyang123/DeepSeek-V3/issues" target="_blank">
|
36 |
+
<img src="https://img.shields.io/badge/Paper-Available-blue?style=flat-square&logo=book&logoColor=white" alt="Paper">
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
37 |
</a>
|
38 |
+
|
39 |
</div>
|
40 |
"""
|
|
|
41 |
#TODO About
|
42 |
LLM_BENCHMARKS_TEXT = f"""
|
43 |
<h2 style="color: #2c3e50;"> Why Leaderboard? </h2>
|
src/display/utils.py
CHANGED
@@ -24,22 +24,22 @@ class ColumnContent:
|
|
24 |
auto_eval_column_dict = []
|
25 |
#TODO
|
26 |
# Init
|
27 |
-
auto_eval_column_dict.append(["model_type_symbol", ColumnContent, ColumnContent("T", "str", True, never_hidden=True)])
|
28 |
auto_eval_column_dict.append(["model", ColumnContent, ColumnContent("Model", "markdown", True, never_hidden=True)])
|
29 |
# #Scores
|
30 |
# auto_eval_column_dict.append(["average", ColumnContent, ColumnContent("Average ⬆️", "number", True)])
|
31 |
for task in Tasks:
|
32 |
-
auto_eval_column_dict.append([task.name, ColumnContent, ColumnContent(task.value.col_name, "float", True)])
|
33 |
# Model information
|
34 |
-
auto_eval_column_dict.append(["model_type", ColumnContent, ColumnContent("Type", "str", False)])
|
35 |
-
auto_eval_column_dict.append(["architecture", ColumnContent, ColumnContent("Architecture", "str", False)])
|
36 |
-
auto_eval_column_dict.append(["weight_type", ColumnContent, ColumnContent("Weight type", "str", False, True)])
|
37 |
-
auto_eval_column_dict.append(["precision", ColumnContent, ColumnContent("Precision", "str", False)])
|
38 |
-
auto_eval_column_dict.append(["license", ColumnContent, ColumnContent("Hub License", "str", False)])
|
39 |
-
auto_eval_column_dict.append(["params", ColumnContent, ColumnContent("#Params (B)", "float", False)])
|
40 |
-
auto_eval_column_dict.append(["likes", ColumnContent, ColumnContent("Hub ❤️", "number", False)])
|
41 |
auto_eval_column_dict.append(["still_on_hub", ColumnContent, ColumnContent("Available on the hub", "bool", False)])
|
42 |
-
auto_eval_column_dict.append(["revision", ColumnContent, ColumnContent("Model sha", "str", False, False)])
|
43 |
|
44 |
# We use make dataclass to dynamically fill the scores from Tasks
|
45 |
AutoEvalColumn = make_dataclass("AutoEvalColumn", auto_eval_column_dict, frozen=True)
|
|
|
24 |
auto_eval_column_dict = []
|
25 |
#TODO
|
26 |
# Init
|
27 |
+
#auto_eval_column_dict.append(["model_type_symbol", ColumnContent, ColumnContent("T", "str", True, never_hidden=True)])
|
28 |
auto_eval_column_dict.append(["model", ColumnContent, ColumnContent("Model", "markdown", True, never_hidden=True)])
|
29 |
# #Scores
|
30 |
# auto_eval_column_dict.append(["average", ColumnContent, ColumnContent("Average ⬆️", "number", True)])
|
31 |
for task in Tasks:
|
32 |
+
auto_eval_column_dict.append([task.name, ColumnContent, ColumnContent(task.value.col_name, "float", True , never_hidden= True)])
|
33 |
# Model information
|
34 |
+
#auto_eval_column_dict.append(["model_type", ColumnContent, ColumnContent("Type", "str", False)])
|
35 |
+
#auto_eval_column_dict.append(["architecture", ColumnContent, ColumnContent("Architecture", "str", False)])
|
36 |
+
#auto_eval_column_dict.append(["weight_type", ColumnContent, ColumnContent("Weight type", "str", False, True)])
|
37 |
+
#auto_eval_column_dict.append(["precision", ColumnContent, ColumnContent("Precision", "str", False)])
|
38 |
+
#auto_eval_column_dict.append(["license", ColumnContent, ColumnContent("Hub License", "str", False)])
|
39 |
+
#auto_eval_column_dict.append(["params", ColumnContent, ColumnContent("#Params (B)", "float", False)])
|
40 |
+
#auto_eval_column_dict.append(["likes", ColumnContent, ColumnContent("Hub ❤️", "number", False)])
|
41 |
auto_eval_column_dict.append(["still_on_hub", ColumnContent, ColumnContent("Available on the hub", "bool", False)])
|
42 |
+
#auto_eval_column_dict.append(["revision", ColumnContent, ColumnContent("Model sha", "str", False, False)])
|
43 |
|
44 |
# We use make dataclass to dynamically fill the scores from Tasks
|
45 |
AutoEvalColumn = make_dataclass("AutoEvalColumn", auto_eval_column_dict, frozen=True)
|
src/leaderboard/read_evals.py
CHANGED
@@ -115,17 +115,17 @@ class EvalResult:
|
|
115 |
#average = next(iter(self.results.values()))
|
116 |
data_dict = {
|
117 |
"eval_name": self.eval_name, # not a column, just a save name,
|
118 |
-
AutoEvalColumn.precision.name: self.precision.value.name, #results/precision
|
119 |
-
AutoEvalColumn.model_type.name: self.model_type.value.name, #requests/model_type
|
120 |
-
AutoEvalColumn.model_type_symbol.name: self.model_type.value.symbol,
|
121 |
-
AutoEvalColumn.weight_type.name: self.weight_type.value.name,
|
122 |
-
AutoEvalColumn.architecture.name: self.architecture,
|
123 |
AutoEvalColumn.model.name: make_clickable_model(self.full_model),
|
124 |
-
AutoEvalColumn.revision.name: self.revision,
|
125 |
#AutoEvalColumn.average.name: average,
|
126 |
-
AutoEvalColumn.license.name: self.license,
|
127 |
-
AutoEvalColumn.likes.name: self.likes,
|
128 |
-
AutoEvalColumn.params.name: self.num_params,
|
129 |
AutoEvalColumn.still_on_hub.name: self.still_on_hub,
|
130 |
}
|
131 |
|
|
|
115 |
#average = next(iter(self.results.values()))
|
116 |
data_dict = {
|
117 |
"eval_name": self.eval_name, # not a column, just a save name,
|
118 |
+
#AutoEvalColumn.precision.name: self.precision.value.name, #results/precision
|
119 |
+
#AutoEvalColumn.model_type.name: self.model_type.value.name, #requests/model_type
|
120 |
+
#AutoEvalColumn.model_type_symbol.name: self.model_type.value.symbol,
|
121 |
+
#AutoEvalColumn.weight_type.name: self.weight_type.value.name,
|
122 |
+
#AutoEvalColumn.architecture.name: self.architecture,
|
123 |
AutoEvalColumn.model.name: make_clickable_model(self.full_model),
|
124 |
+
#AutoEvalColumn.revision.name: self.revision,
|
125 |
#AutoEvalColumn.average.name: average,
|
126 |
+
#AutoEvalColumn.license.name: self.license,
|
127 |
+
#AutoEvalColumn.likes.name: self.likes,
|
128 |
+
#AutoEvalColumn.params.name: self.num_params,
|
129 |
AutoEvalColumn.still_on_hub.name: self.still_on_hub,
|
130 |
}
|
131 |
|