Text Generation
Transformers
Safetensors
English
Russian
mistral
conversational
text-generation-inference
Files changed (1) hide show
  1. README.md +139 -26
README.md CHANGED
@@ -1,13 +1,17 @@
1
  ---
2
- library_name: transformers
3
- base_model: HuggingFaceH4/zephyr-7b-beta
4
- license: mit
5
  language:
6
  - ru
7
  - en
 
 
8
  tags:
9
  - python
10
  - code
 
 
 
 
 
11
  pipeline_tag: conversational
12
  model-index:
13
  - name: zephyr-python-ru-merged
@@ -18,76 +22,172 @@ model-index:
18
  name: gsm8k
19
  type: gsm8k
20
  metrics:
21
- - name: Grade School Math 8K (5-Shot)
22
- type: Grade School Math 8K (5-Shot)
23
  value: 32.52
 
24
  source:
25
- name: Open LLM Leaderboard
26
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
 
27
  - task:
28
  type: text-generation
29
  dataset:
30
  name: ai2_arc
31
  type: ai2_arc
32
  metrics:
33
- - name: AI2 Reasoning Challenge (25-Shot)
34
- type: AI2 Reasoning Challenge (25-Shot)
35
  value: 56.06
 
36
  source:
37
- name: Open LLM Leaderboard
38
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
 
39
  - task:
40
  type: text-generation
41
  dataset:
42
  name: HellaSwag
43
  type: HellaSwag
44
  metrics:
45
- - name: HellaSwag (10-Shot)
46
- type: HellaSwag (10-Shot)
47
  value: 82.06
 
48
  source:
49
- name: Open LLM Leaderboard
50
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
 
51
  - task:
52
  type: text-generation
53
  dataset:
54
  name: MMLU
55
  type: MMLU
56
  metrics:
57
- - name: MMLU (5-Shot)
58
- type: MMLU (5-Shot)
59
  value: 60.02
 
60
  source:
61
- name: Open LLM Leaderboard
62
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
 
63
  - task:
64
  type: text-generation
65
  dataset:
66
  name: truthfulqa
67
  type: truthfulqa
68
  metrics:
69
- - name: truthfulqa:mc (0-Shot)
70
- type: truthfulqa:mc (0-Shot)
71
  value: 52.81
 
72
  source:
73
- name: Open LLM Leaderboard
74
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
 
75
  - task:
76
  type: text-generation
77
  dataset:
78
  name: winogrande
79
  type: winogrande
80
  metrics:
81
- - name: winogrande (5-Shot)
82
- type: winogrande (5-Shot)
83
  value: 76.95
 
84
  source:
85
- name: Open LLM Leaderboard
86
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
87
- datasets:
88
- - MexIvanov/Vezora-Tested-22k-Python-Alpaca-ru
89
- - MexIvanov/CodeExercise-Python-27k-ru
90
- - zelkame/ru-stackoverflow-py
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
91
  ---
92
 
93
  # Model Card for Model ID
@@ -188,4 +288,17 @@ The following `bitsandbytes` quantization config was used during training:
188
  ### Framework versions
189
 
190
 
191
- - PEFT 0.6.2
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
 
 
 
2
  language:
3
  - ru
4
  - en
5
+ license: mit
6
+ library_name: transformers
7
  tags:
8
  - python
9
  - code
10
+ datasets:
11
+ - MexIvanov/Vezora-Tested-22k-Python-Alpaca-ru
12
+ - MexIvanov/CodeExercise-Python-27k-ru
13
+ - zelkame/ru-stackoverflow-py
14
+ base_model: HuggingFaceH4/zephyr-7b-beta
15
  pipeline_tag: conversational
16
  model-index:
17
  - name: zephyr-python-ru-merged
 
22
  name: gsm8k
23
  type: gsm8k
24
  metrics:
25
+ - type: Grade School Math 8K (5-Shot)
 
26
  value: 32.52
27
+ name: Grade School Math 8K (5-Shot)
28
  source:
 
29
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
30
+ name: Open LLM Leaderboard
31
  - task:
32
  type: text-generation
33
  dataset:
34
  name: ai2_arc
35
  type: ai2_arc
36
  metrics:
37
+ - type: AI2 Reasoning Challenge (25-Shot)
 
38
  value: 56.06
39
+ name: AI2 Reasoning Challenge (25-Shot)
40
  source:
 
41
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
42
+ name: Open LLM Leaderboard
43
  - task:
44
  type: text-generation
45
  dataset:
46
  name: HellaSwag
47
  type: HellaSwag
48
  metrics:
49
+ - type: HellaSwag (10-Shot)
 
50
  value: 82.06
51
+ name: HellaSwag (10-Shot)
52
  source:
 
53
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
54
+ name: Open LLM Leaderboard
55
  - task:
56
  type: text-generation
57
  dataset:
58
  name: MMLU
59
  type: MMLU
60
  metrics:
61
+ - type: MMLU (5-Shot)
 
62
  value: 60.02
63
+ name: MMLU (5-Shot)
64
  source:
 
65
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
66
+ name: Open LLM Leaderboard
67
  - task:
68
  type: text-generation
69
  dataset:
70
  name: truthfulqa
71
  type: truthfulqa
72
  metrics:
73
+ - type: truthfulqa:mc (0-Shot)
 
74
  value: 52.81
75
+ name: truthfulqa:mc (0-Shot)
76
  source:
 
77
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
78
+ name: Open LLM Leaderboard
79
  - task:
80
  type: text-generation
81
  dataset:
82
  name: winogrande
83
  type: winogrande
84
  metrics:
85
+ - type: winogrande (5-Shot)
 
86
  value: 76.95
87
+ name: winogrande (5-Shot)
88
  source:
 
89
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard
90
+ name: Open LLM Leaderboard
91
+ - task:
92
+ type: text-generation
93
+ name: Text Generation
94
+ dataset:
95
+ name: AI2 Reasoning Challenge (25-Shot)
96
+ type: ai2_arc
97
+ config: ARC-Challenge
98
+ split: test
99
+ args:
100
+ num_few_shot: 25
101
+ metrics:
102
+ - type: acc_norm
103
+ value: 56.06
104
+ name: normalized accuracy
105
+ source:
106
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MexIvanov/zephyr-python-ru-merged
107
+ name: Open LLM Leaderboard
108
+ - task:
109
+ type: text-generation
110
+ name: Text Generation
111
+ dataset:
112
+ name: HellaSwag (10-Shot)
113
+ type: hellaswag
114
+ split: validation
115
+ args:
116
+ num_few_shot: 10
117
+ metrics:
118
+ - type: acc_norm
119
+ value: 82.06
120
+ name: normalized accuracy
121
+ source:
122
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MexIvanov/zephyr-python-ru-merged
123
+ name: Open LLM Leaderboard
124
+ - task:
125
+ type: text-generation
126
+ name: Text Generation
127
+ dataset:
128
+ name: MMLU (5-Shot)
129
+ type: cais/mmlu
130
+ config: all
131
+ split: test
132
+ args:
133
+ num_few_shot: 5
134
+ metrics:
135
+ - type: acc
136
+ value: 60.2
137
+ name: accuracy
138
+ source:
139
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MexIvanov/zephyr-python-ru-merged
140
+ name: Open LLM Leaderboard
141
+ - task:
142
+ type: text-generation
143
+ name: Text Generation
144
+ dataset:
145
+ name: TruthfulQA (0-shot)
146
+ type: truthful_qa
147
+ config: multiple_choice
148
+ split: validation
149
+ args:
150
+ num_few_shot: 0
151
+ metrics:
152
+ - type: mc2
153
+ value: 52.81
154
+ source:
155
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MexIvanov/zephyr-python-ru-merged
156
+ name: Open LLM Leaderboard
157
+ - task:
158
+ type: text-generation
159
+ name: Text Generation
160
+ dataset:
161
+ name: Winogrande (5-shot)
162
+ type: winogrande
163
+ config: winogrande_xl
164
+ split: validation
165
+ args:
166
+ num_few_shot: 5
167
+ metrics:
168
+ - type: acc
169
+ value: 76.95
170
+ name: accuracy
171
+ source:
172
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MexIvanov/zephyr-python-ru-merged
173
+ name: Open LLM Leaderboard
174
+ - task:
175
+ type: text-generation
176
+ name: Text Generation
177
+ dataset:
178
+ name: GSM8k (5-shot)
179
+ type: gsm8k
180
+ config: main
181
+ split: test
182
+ args:
183
+ num_few_shot: 5
184
+ metrics:
185
+ - type: acc
186
+ value: 32.52
187
+ name: accuracy
188
+ source:
189
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=MexIvanov/zephyr-python-ru-merged
190
+ name: Open LLM Leaderboard
191
  ---
192
 
193
  # Model Card for Model ID
 
288
  ### Framework versions
289
 
290
 
291
+ - PEFT 0.6.2
292
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
293
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_MexIvanov__zephyr-python-ru-merged)
294
+
295
+ | Metric |Value|
296
+ |---------------------------------|----:|
297
+ |Avg. |60.10|
298
+ |AI2 Reasoning Challenge (25-Shot)|56.06|
299
+ |HellaSwag (10-Shot) |82.06|
300
+ |MMLU (5-Shot) |60.20|
301
+ |TruthfulQA (0-shot) |52.81|
302
+ |Winogrande (5-shot) |76.95|
303
+ |GSM8k (5-shot) |32.52|
304
+