nsarrazin HF Staff commited on
Commit
a954f65
·
unverified ·
1 Parent(s): 4785c57

feat(huggingchat): Model update 24/09/2024 (#1485)

Browse files
Files changed (1) hide show
  1. chart/env/prod.yaml +53 -71
chart/env/prod.yaml CHANGED
@@ -53,9 +53,8 @@ envVars:
53
  "tools": true,
54
  "preprompt": "",
55
  "parameters": {
56
- "temperature": 0.6,
57
- "top_p": 0.9,
58
  "stop": ["<|endoftext|>", "<|eot_id|>"],
 
59
  "max_new_tokens": 1024,
60
  "truncate": 7167
61
  },
@@ -115,42 +114,10 @@ envVars:
115
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/qwen-logo.png",
116
  "preprompt": "You are Qwen, created by Alibaba Cloud. You are a helpful assistant.",
117
  "parameters": {
118
- "temperature": 0.6,
119
  "stop": ["<|endoftext|>", "<|im_end|>"],
120
- "truncate": 28672,
121
- "max_new_tokens": 3072,
122
- },
123
- "promptExamples": [
124
- {
125
- "title": "Write an email from bullet list",
126
- "prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)"
127
- },
128
- {
129
- "title": "Code a snake game",
130
- "prompt": "Code a basic snake game in python, give explanations for each step."
131
- },
132
- {
133
- "title": "Assist in a task",
134
- "prompt": "How do I make a delicious lemon cheesecake?"
135
- }
136
- ]
137
- },
138
- {
139
- "name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
140
- "description": "A high-quality sparse mixture of experts model with open weights.",
141
- "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
142
- "websiteUrl": "https://mistral.ai/news/mixtral-of-experts/",
143
- "modelUrl": "https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1",
144
- "tokenizer": "mistralai/Mixtral-8x7B-Instruct-v0.1",
145
- "preprompt": "",
146
- "parameters": {
147
  "temperature": 0.6,
148
- "top_p": 0.95,
149
- "repetition_penalty": 1.2,
150
- "top_k": 50,
151
- "truncate": 24576,
152
- "max_new_tokens": 8192,
153
- "stop": ["</s>"]
154
  },
155
  "promptExamples": [
156
  {
@@ -168,12 +135,12 @@ envVars:
168
  ]
169
  },
170
  {
171
- "name": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
172
- "description": "Nous Hermes' strong flagship model trained on the Mixtral 8x7B.",
173
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png",
174
  "websiteUrl": "https://nousresearch.com/",
175
- "modelUrl": "https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
176
- "tokenizer": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
177
  "promptExamples": [
178
  {
179
  "title": "Write an email from bullet list",
@@ -189,32 +156,26 @@ envVars:
189
  }
190
  ],
191
  "parameters": {
192
- "temperature": 0.7,
193
- "top_p": 0.95,
194
- "repetition_penalty": 1,
195
- "top_k": 50,
196
- "truncate": 24576,
197
- "max_new_tokens": 2048,
198
- "stop": ["<|im_end|>"]
199
  }
200
  },
201
  {
202
- "name": "mistralai/Mistral-7B-Instruct-v0.3",
203
- "displayName": "mistralai/Mistral-7B-Instruct-v0.3",
 
204
  "description": "A small model with good capabilities in language understanding and commonsense reasoning.",
205
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
206
- "websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/",
207
- "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3",
208
- "tokenizer": "mistralai/Mistral-7B-Instruct-v0.3",
209
  "preprompt": "",
210
  "parameters": {
211
- "temperature": 0.3,
212
- "top_p": 0.95,
213
- "repetition_penalty": 1.2,
214
- "top_k": 50,
215
- "truncate": 3072,
216
- "max_new_tokens": 1024,
217
- "stop": ["</s>"]
218
  },
219
  "promptExamples": [
220
  {
@@ -232,18 +193,18 @@ envVars:
232
  ]
233
  },
234
  {
235
- "name": "microsoft/Phi-3-mini-4k-instruct",
236
- "tokenizer": "microsoft/Phi-3-mini-4k-instruct",
237
  "description": "One of the best small models (3.8B parameters), super fast for simple tasks.",
238
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/microsoft-logo.png",
239
- "modelUrl": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct",
240
- "websiteUrl": "https://azure.microsoft.com/en-us/blog/introducing-phi-3-redefining-whats-possible-with-slms/",
241
  "preprompt": "",
242
  "parameters": {
243
  "stop": ["<|end|>", "<|endoftext|>", "<|assistant|>"],
244
- "temperature": 0.7,
245
- "max_new_tokens": 1024,
246
- "truncate": 3071
247
  },
248
  "promptExamples": [
249
  {
@@ -268,7 +229,6 @@ envVars:
268
  },
269
  "parameters": {
270
  "temperature": 0.6,
271
- "top_p": 0.9,
272
  "stop": ["<|endoftext|>", "<|eot_id|>"]
273
  },
274
  "unlisted": true
@@ -290,10 +250,32 @@ envVars:
290
  { "name": "meta-llama/Llama-2-70b-chat-hf" },
291
  { "name": "codellama/CodeLlama-70b-Instruct-hf" },
292
  { "name": "openchat/openchat-3.5-0106" },
293
- { "name": "meta-llama/Meta-Llama-3-70B-Instruct"},
294
- { "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8"},
295
- { "name": "CohereForAI/c4ai-command-r-plus", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"},
296
- { "name": "01-ai/Yi-1.5-34B-Chat", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"}
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
297
  ]
298
  PUBLIC_ORIGIN: "https://huggingface.co"
299
  PUBLIC_SHARE_PREFIX: "https://hf.co/chat"
 
53
  "tools": true,
54
  "preprompt": "",
55
  "parameters": {
 
 
56
  "stop": ["<|endoftext|>", "<|eot_id|>"],
57
+ "temperature": 0.6,
58
  "max_new_tokens": 1024,
59
  "truncate": 7167
60
  },
 
114
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/qwen-logo.png",
115
  "preprompt": "You are Qwen, created by Alibaba Cloud. You are a helpful assistant.",
116
  "parameters": {
 
117
  "stop": ["<|endoftext|>", "<|im_end|>"],
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
118
  "temperature": 0.6,
119
+ "truncate": 28672,
120
+ "max_new_tokens": 3072
 
 
 
 
121
  },
122
  "promptExamples": [
123
  {
 
135
  ]
136
  },
137
  {
138
+ "name": "NousResearch/Hermes-3-Llama-3.1-8B",
139
+ "description": "Nous Research's latest Hermes 3 release in 8B size. Follows instruction closely.",
140
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png",
141
  "websiteUrl": "https://nousresearch.com/",
142
+ "modelUrl": "https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B",
143
+ "tokenizer": "NousResearch/Hermes-3-Llama-3.1-8B",
144
  "promptExamples": [
145
  {
146
  "title": "Write an email from bullet list",
 
156
  }
157
  ],
158
  "parameters": {
159
+ "stop": ["<|im_end|>"],
160
+ "temperature": 0.6,
161
+ "truncate": 14336,
162
+ "max_new_tokens": 1536
 
 
 
163
  }
164
  },
165
  {
166
+ "name": "mistralai/Mistral-Nemo-Instruct-2407",
167
+ "tokenizer": "mistralai/Mistral-Nemo-Instruct-2407",
168
+ "displayName": "mistralai/Mistral-Nemo-Instruct-2407",
169
  "description": "A small model with good capabilities in language understanding and commonsense reasoning.",
170
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
171
+ "websiteUrl": "https://mistral.ai/news/mistral-nemo/",
172
+ "modelUrl": "https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407",
 
173
  "preprompt": "",
174
  "parameters": {
175
+ "stop": ["</s>"],
176
+ "temperature": 0.6,
177
+ "truncate": 14336,
178
+ "max_new_tokens": 1536
 
 
 
179
  },
180
  "promptExamples": [
181
  {
 
193
  ]
194
  },
195
  {
196
+ "name": "microsoft/Phi-3.5-mini-instruct",
197
+ "tokenizer": "microsoft/Phi-3.5-mini-instruct",
198
  "description": "One of the best small models (3.8B parameters), super fast for simple tasks.",
199
  "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/microsoft-logo.png",
200
+ "modelUrl": "https://huggingface.co/microsoft/Phi-3.5-mini-instruct",
201
+ "websiteUrl": "https://techcommunity.microsoft.com/t5/ai-azure-ai-services-blog/discover-the-new-multi-lingual-high-quality-phi-3-5-slms/ba-p/4225280/",
202
  "preprompt": "",
203
  "parameters": {
204
  "stop": ["<|end|>", "<|endoftext|>", "<|assistant|>"],
205
+ "temperature": 0.6,
206
+ "truncate": 28672,
207
+ "max_new_tokens": 3072
208
  },
209
  "promptExamples": [
210
  {
 
229
  },
230
  "parameters": {
231
  "temperature": 0.6,
 
232
  "stop": ["<|endoftext|>", "<|eot_id|>"]
233
  },
234
  "unlisted": true
 
250
  { "name": "meta-llama/Llama-2-70b-chat-hf" },
251
  { "name": "codellama/CodeLlama-70b-Instruct-hf" },
252
  { "name": "openchat/openchat-3.5-0106" },
253
+ { "name": "meta-llama/Meta-Llama-3-70B-Instruct" },
254
+ { "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8" },
255
+ {
256
+ "name": "CohereForAI/c4ai-command-r-plus",
257
+ "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
258
+ },
259
+ {
260
+ "name": "01-ai/Yi-1.5-34B-Chat",
261
+ "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
262
+ },
263
+ {
264
+ "name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
265
+ "transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
266
+ },
267
+ {
268
+ "name": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
269
+ "transferTo": "NousResearch/Hermes-3-Llama-3.1-8B"
270
+ },
271
+ {
272
+ "name": "mistralai/Mistral-7B-Instruct-v0.3",
273
+ "transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
274
+ },
275
+ {
276
+ "name": "microsoft/Phi-3-mini-4k-instruct",
277
+ "transferTo": "microsoft/Phi-3.5-mini-instruct"
278
+ }
279
  ]
280
  PUBLIC_ORIGIN: "https://huggingface.co"
281
  PUBLIC_SHARE_PREFIX: "https://hf.co/chat"