ycy commited on
Commit
360820c
·
1 Parent(s): 589c39f
Files changed (4) hide show
  1. app.py +1 -1
  2. src/about.py +5 -38
  3. src/display/utils.py +10 -10
  4. src/leaderboard/read_evals.py +9 -9
app.py CHANGED
@@ -86,7 +86,7 @@ def init_leaderboard(dataframe):
86
  label="Select the number of parameters (B)",
87
  ),
88
  ColumnFilter(
89
- AutoEvalColumn.still_on_hub.name, type="boolean", label="Closed-source", default=True
90
  ),
91
  ],
92
  bool_checkboxgroup_label="Hide models",
 
86
  label="Select the number of parameters (B)",
87
  ),
88
  ColumnFilter(
89
+ AutoEvalColumn.still_on_hub.name, type="boolean", label="Available on the hub", default=False
90
  ),
91
  ],
92
  bool_checkboxgroup_label="Hide models",
src/about.py CHANGED
@@ -27,50 +27,17 @@ NUM_FEWSHOT = 0 # Change with your few shot
27
  #TODO title
28
  TITLE = """<h1 align="center" id="space-title">VLM Detailed Caption Leaderborad</h1>"""
29
 
30
-
31
-
32
- #TODO interoduction
33
  INTRODUCTION_TEXT = """
34
  <div style="display: flex; flex-wrap: wrap; gap: 10px;">
35
- <!-- Release badge -->
36
- <a href="https://github.com/yanchenyang123/DeepSeek-V3/releases" target="_blank">
37
- <img src="https://img.shields.io/github/v/release/yanchenyang123/DeepSeek-V3?style=flat-square" alt="release">
38
- </a>
39
- <!-- Docker badge -->
40
- <a href="https://hub.docker.com/r/your-docker-image" target="_blank">
41
- <img src="https://img.shields.io/badge/docker-v1.3.1-blue?style=flat-square" alt="docker">
42
- </a>
43
- <!-- Stars badge -->
44
- <a href="https://github.com/yanchenyang123/DeepSeek-V3/stargazers" target="_blank">
45
- <img src="https://img.shields.io/github/stars/yanchenyang123/DeepSeek-V3?style=flat-square" alt="stars">
46
- </a>
47
- <!-- Issues badge -->
48
  <a href="https://github.com/yanchenyang123/DeepSeek-V3/issues" target="_blank">
49
- <img src="https://img.shields.io/github/issues/yanchenyang123/DeepSeek-V3?color=pink&style=flat-square" alt="issues">
50
- </a>
51
- <!-- Contributors badge -->
52
- <a href="https://github.com/yanchenyang123/DeepSeek-V3/graphs/contributors" target="_blank">
53
- <img src="https://img.shields.io/github/contributors/yanchenyang123/DeepSeek-V3?color=yellowgreen&style=flat-square" alt="contributors">
54
- </a>
55
- <!-- Forks badge -->
56
- <a href="https://github.com/yanchenyang123/DeepSeek-V3/network/members" target="_blank">
57
- <img src="https://img.shields.io/github/forks/yanchenyang123/DeepSeek-V3?color=cyan&style=flat-square" alt="forks">
58
- </a>
59
- <!-- License badge -->
60
- <a href="https://github.com/yanchenyang123/DeepSeek-V3/blob/main/LICENSE" target="_blank">
61
- <img src="https://img.shields.io/github/license/yanchenyang123/DeepSeek-V3?color=black&style=flat-square" alt="license">
62
- </a>
63
- <!-- WeChat badge -->
64
- <a href="https://your-wechat-link.com" target="_blank">
65
- <img src="https://img.shields.io/badge/WeChat-微信-green?style=flat-square" alt="WeChat">
66
- </a>
67
- <!-- Hugging Face Spaces badge -->
68
- <a href="https://huggingface.co/spaces/your-space" target="_blank">
69
- <img src="https://img.shields.io/badge/Open%20in%20Spaces-%F0%9F%A4%AF-orange?style=flat-square" alt="Open in Spaces">
70
  </a>
 
71
  </div>
72
  """
73
-
74
  #TODO About
75
  LLM_BENCHMARKS_TEXT = f"""
76
  <h2 style="color: #2c3e50;"> Why Leaderboard? </h2>
 
27
  #TODO title
28
  TITLE = """<h1 align="center" id="space-title">VLM Detailed Caption Leaderborad</h1>"""
29
 
30
+ # introduction text
 
 
31
  INTRODUCTION_TEXT = """
32
  <div style="display: flex; flex-wrap: wrap; gap: 10px;">
33
+
34
+ <!-- Paper badge with custom icon -->
 
 
 
 
 
 
 
 
 
 
 
35
  <a href="https://github.com/yanchenyang123/DeepSeek-V3/issues" target="_blank">
36
+ <img src="https://img.shields.io/badge/Paper-Available-blue?style=flat-square&logo=book&logoColor=white" alt="Paper">
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
37
  </a>
38
+
39
  </div>
40
  """
 
41
  #TODO About
42
  LLM_BENCHMARKS_TEXT = f"""
43
  <h2 style="color: #2c3e50;"> Why Leaderboard? </h2>
src/display/utils.py CHANGED
@@ -24,22 +24,22 @@ class ColumnContent:
24
  auto_eval_column_dict = []
25
  #TODO
26
  # Init
27
- auto_eval_column_dict.append(["model_type_symbol", ColumnContent, ColumnContent("T", "str", True, never_hidden=True)])
28
  auto_eval_column_dict.append(["model", ColumnContent, ColumnContent("Model", "markdown", True, never_hidden=True)])
29
  # #Scores
30
  # auto_eval_column_dict.append(["average", ColumnContent, ColumnContent("Average ⬆️", "number", True)])
31
  for task in Tasks:
32
- auto_eval_column_dict.append([task.name, ColumnContent, ColumnContent(task.value.col_name, "float", True)])
33
  # Model information
34
- auto_eval_column_dict.append(["model_type", ColumnContent, ColumnContent("Type", "str", False)])
35
- auto_eval_column_dict.append(["architecture", ColumnContent, ColumnContent("Architecture", "str", False)])
36
- auto_eval_column_dict.append(["weight_type", ColumnContent, ColumnContent("Weight type", "str", False, True)])
37
- auto_eval_column_dict.append(["precision", ColumnContent, ColumnContent("Precision", "str", False)])
38
- auto_eval_column_dict.append(["license", ColumnContent, ColumnContent("Hub License", "str", False)])
39
- auto_eval_column_dict.append(["params", ColumnContent, ColumnContent("#Params (B)", "float", False)])
40
- auto_eval_column_dict.append(["likes", ColumnContent, ColumnContent("Hub ❤️", "number", False)])
41
  auto_eval_column_dict.append(["still_on_hub", ColumnContent, ColumnContent("Available on the hub", "bool", False)])
42
- auto_eval_column_dict.append(["revision", ColumnContent, ColumnContent("Model sha", "str", False, False)])
43
 
44
  # We use make dataclass to dynamically fill the scores from Tasks
45
  AutoEvalColumn = make_dataclass("AutoEvalColumn", auto_eval_column_dict, frozen=True)
 
24
  auto_eval_column_dict = []
25
  #TODO
26
  # Init
27
+ #auto_eval_column_dict.append(["model_type_symbol", ColumnContent, ColumnContent("T", "str", True, never_hidden=True)])
28
  auto_eval_column_dict.append(["model", ColumnContent, ColumnContent("Model", "markdown", True, never_hidden=True)])
29
  # #Scores
30
  # auto_eval_column_dict.append(["average", ColumnContent, ColumnContent("Average ⬆️", "number", True)])
31
  for task in Tasks:
32
+ auto_eval_column_dict.append([task.name, ColumnContent, ColumnContent(task.value.col_name, "float", True , never_hidden= True)])
33
  # Model information
34
+ #auto_eval_column_dict.append(["model_type", ColumnContent, ColumnContent("Type", "str", False)])
35
+ #auto_eval_column_dict.append(["architecture", ColumnContent, ColumnContent("Architecture", "str", False)])
36
+ #auto_eval_column_dict.append(["weight_type", ColumnContent, ColumnContent("Weight type", "str", False, True)])
37
+ #auto_eval_column_dict.append(["precision", ColumnContent, ColumnContent("Precision", "str", False)])
38
+ #auto_eval_column_dict.append(["license", ColumnContent, ColumnContent("Hub License", "str", False)])
39
+ #auto_eval_column_dict.append(["params", ColumnContent, ColumnContent("#Params (B)", "float", False)])
40
+ #auto_eval_column_dict.append(["likes", ColumnContent, ColumnContent("Hub ❤️", "number", False)])
41
  auto_eval_column_dict.append(["still_on_hub", ColumnContent, ColumnContent("Available on the hub", "bool", False)])
42
+ #auto_eval_column_dict.append(["revision", ColumnContent, ColumnContent("Model sha", "str", False, False)])
43
 
44
  # We use make dataclass to dynamically fill the scores from Tasks
45
  AutoEvalColumn = make_dataclass("AutoEvalColumn", auto_eval_column_dict, frozen=True)
src/leaderboard/read_evals.py CHANGED
@@ -115,17 +115,17 @@ class EvalResult:
115
  #average = next(iter(self.results.values()))
116
  data_dict = {
117
  "eval_name": self.eval_name, # not a column, just a save name,
118
- AutoEvalColumn.precision.name: self.precision.value.name, #results/precision
119
- AutoEvalColumn.model_type.name: self.model_type.value.name, #requests/model_type
120
- AutoEvalColumn.model_type_symbol.name: self.model_type.value.symbol,
121
- AutoEvalColumn.weight_type.name: self.weight_type.value.name,
122
- AutoEvalColumn.architecture.name: self.architecture,
123
  AutoEvalColumn.model.name: make_clickable_model(self.full_model),
124
- AutoEvalColumn.revision.name: self.revision,
125
  #AutoEvalColumn.average.name: average,
126
- AutoEvalColumn.license.name: self.license,
127
- AutoEvalColumn.likes.name: self.likes,
128
- AutoEvalColumn.params.name: self.num_params,
129
  AutoEvalColumn.still_on_hub.name: self.still_on_hub,
130
  }
131
 
 
115
  #average = next(iter(self.results.values()))
116
  data_dict = {
117
  "eval_name": self.eval_name, # not a column, just a save name,
118
+ #AutoEvalColumn.precision.name: self.precision.value.name, #results/precision
119
+ #AutoEvalColumn.model_type.name: self.model_type.value.name, #requests/model_type
120
+ #AutoEvalColumn.model_type_symbol.name: self.model_type.value.symbol,
121
+ #AutoEvalColumn.weight_type.name: self.weight_type.value.name,
122
+ #AutoEvalColumn.architecture.name: self.architecture,
123
  AutoEvalColumn.model.name: make_clickable_model(self.full_model),
124
+ #AutoEvalColumn.revision.name: self.revision,
125
  #AutoEvalColumn.average.name: average,
126
+ #AutoEvalColumn.license.name: self.license,
127
+ #AutoEvalColumn.likes.name: self.likes,
128
+ #AutoEvalColumn.params.name: self.num_params,
129
  AutoEvalColumn.still_on_hub.name: self.still_on_hub,
130
  }
131