@@ -53,9 +53,8 @@ envVars:
53
53
"tools": true,
54
54
"preprompt": "",
55
55
"parameters": {
56
- "temperature": 0.6,
57
- "top_p": 0.9,
58
56
"stop": ["<|endoftext|>", "<|eot_id|>"],
57
+ "temperature": 0.6,
59
58
"max_new_tokens": 1024,
60
59
"truncate": 7167
61
60
},
@@ -115,42 +114,10 @@ envVars:
115
114
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/qwen-logo.png",
116
115
"preprompt": "You are Qwen, created by Alibaba Cloud. You are a helpful assistant.",
117
116
"parameters": {
118
- "temperature": 0.6,
119
117
"stop": ["<|endoftext|>", "<|im_end|>"],
120
- "truncate": 28672,
121
- "max_new_tokens": 3072,
122
- },
123
- "promptExamples": [
124
- {
125
- "title": "Write an email from bullet list",
126
- "prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)"
127
- },
128
- {
129
- "title": "Code a snake game",
130
- "prompt": "Code a basic snake game in python, give explanations for each step."
131
- },
132
- {
133
- "title": "Assist in a task",
134
- "prompt": "How do I make a delicious lemon cheesecake?"
135
- }
136
- ]
137
- },
138
- {
139
- "name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
140
- "description": "A high-quality sparse mixture of experts model with open weights.",
141
- "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
142
- "websiteUrl": "https://mistral.ai/news/mixtral-of-experts/",
143
- "modelUrl": "https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1",
144
- "tokenizer": "mistralai/Mixtral-8x7B-Instruct-v0.1",
145
- "preprompt": "",
146
- "parameters": {
147
118
"temperature": 0.6,
148
- "top_p": 0.95,
149
- "repetition_penalty": 1.2,
150
- "top_k": 50,
151
- "truncate": 24576,
152
- "max_new_tokens": 8192,
153
- "stop": ["</s>"]
119
+ "truncate": 28672,
120
+ "max_new_tokens": 3072
154
121
},
155
122
"promptExamples": [
156
123
{
@@ -168,12 +135,12 @@ envVars:
168
135
]
169
136
},
170
137
{
171
- "name": "NousResearch/Nous- Hermes-2-Mixtral-8x7B-DPO ",
172
- "description": "Nous Hermes' strong flagship model trained on the Mixtral 8x7B .",
138
+ "name": "NousResearch/Hermes-3-Llama-3.1-8B ",
139
+ "description": "Nous Research's latest Hermes 3 release in 8B size. Follows instruction closely .",
173
140
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png",
174
141
"websiteUrl": "https://nousresearch.com/",
175
- "modelUrl": "https://huggingface.co/NousResearch/Nous- Hermes-2-Mixtral-8x7B-DPO ",
176
- "tokenizer": "NousResearch/Nous- Hermes-2-Mixtral-8x7B-DPO ",
142
+ "modelUrl": "https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B ",
143
+ "tokenizer": "NousResearch/Hermes-3-Llama-3.1-8B ",
177
144
"promptExamples": [
178
145
{
179
146
"title": "Write an email from bullet list",
@@ -189,32 +156,26 @@ envVars:
189
156
}
190
157
],
191
158
"parameters": {
192
- "temperature": 0.7,
193
- "top_p": 0.95,
194
- "repetition_penalty": 1,
195
- "top_k": 50,
196
- "truncate": 24576,
197
- "max_new_tokens": 2048,
198
- "stop": ["<|im_end|>"]
159
+ "stop": ["<|im_end|>"],
160
+ "temperature": 0.6,
161
+ "truncate": 14336,
162
+ "max_new_tokens": 1536
199
163
}
200
164
},
201
165
{
202
- "name": "mistralai/Mistral-7B-Instruct-v0.3",
203
- "displayName": "mistralai/Mistral-7B-Instruct-v0.3",
166
+ "name": "mistralai/Mistral-Nemo-Instruct-2407",
167
+ "tokenizer": "mistralai/Mistral-Nemo-Instruct-2407",
168
+ "displayName": "mistralai/Mistral-Nemo-Instruct-2407",
204
169
"description": "A small model with good capabilities in language understanding and commonsense reasoning.",
205
170
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
206
- "websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/",
207
- "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3",
208
- "tokenizer": "mistralai/Mistral-7B-Instruct-v0.3",
171
+ "websiteUrl": "https://mistral.ai/news/mistral-nemo/",
172
+ "modelUrl": "https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407",
209
173
"preprompt": "",
210
174
"parameters": {
211
- "temperature": 0.3,
212
- "top_p": 0.95,
213
- "repetition_penalty": 1.2,
214
- "top_k": 50,
215
- "truncate": 3072,
216
- "max_new_tokens": 1024,
217
- "stop": ["</s>"]
175
+ "stop": ["</s>"],
176
+ "temperature": 0.6,
177
+ "truncate": 14336,
178
+ "max_new_tokens": 1536
218
179
},
219
180
"promptExamples": [
220
181
{
@@ -232,18 +193,18 @@ envVars:
232
193
]
233
194
},
234
195
{
235
- "name": "microsoft/Phi-3-mini-4k -instruct",
236
- "tokenizer": "microsoft/Phi-3-mini-4k -instruct",
196
+ "name": "microsoft/Phi-3.5 -mini-instruct",
197
+ "tokenizer": "microsoft/Phi-3.5 -mini-instruct",
237
198
"description": "One of the best small models (3.8B parameters), super fast for simple tasks.",
238
199
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/microsoft-logo.png",
239
- "modelUrl": "https://huggingface.co/microsoft/Phi-3-mini-4k -instruct",
240
- "websiteUrl": "https://azure .microsoft.com/en-us/ blog/introducing-phi-3-redefining-whats-possible-with- slms/",
200
+ "modelUrl": "https://huggingface.co/microsoft/Phi-3.5 -mini-instruct",
201
+ "websiteUrl": "https://techcommunity .microsoft.com/t5/ai-azure-ai-services- blog/discover-the-new-multi-lingual-high-quality-phi-3-5- slms/ba-p/4225280 /",
241
202
"preprompt": "",
242
203
"parameters": {
243
204
"stop": ["<|end|>", "<|endoftext|>", "<|assistant|>"],
244
- "temperature": 0.7 ,
245
- "max_new_tokens ": 1024 ,
246
- "truncate ": 3071
205
+ "temperature": 0.6 ,
206
+ "truncate ": 28672 ,
207
+ "max_new_tokens ": 3072
247
208
},
248
209
"promptExamples": [
249
210
{
@@ -268,7 +229,6 @@ envVars:
268
229
},
269
230
"parameters": {
270
231
"temperature": 0.6,
271
- "top_p": 0.9,
272
232
"stop": ["<|endoftext|>", "<|eot_id|>"]
273
233
},
274
234
"unlisted": true
@@ -290,10 +250,32 @@ envVars:
290
250
{ "name": "meta-llama/Llama-2-70b-chat-hf" },
291
251
{ "name": "codellama/CodeLlama-70b-Instruct-hf" },
292
252
{ "name": "openchat/openchat-3.5-0106" },
293
- { "name": "meta-llama/Meta-Llama-3-70B-Instruct"},
294
- { "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8"},
295
- { "name": "CohereForAI/c4ai-command-r-plus", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"},
296
- { "name": "01-ai/Yi-1.5-34B-Chat", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"}
253
+ { "name": "meta-llama/Meta-Llama-3-70B-Instruct" },
254
+ { "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8" },
255
+ {
256
+ "name": "CohereForAI/c4ai-command-r-plus",
257
+ "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
258
+ },
259
+ {
260
+ "name": "01-ai/Yi-1.5-34B-Chat",
261
+ "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
262
+ },
263
+ {
264
+ "name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
265
+ "transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
266
+ },
267
+ {
268
+ "name": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
269
+ "transferTo": "NousResearch/Hermes-3-Llama-3.1-8B"
270
+ },
271
+ {
272
+ "name": "mistralai/Mistral-7B-Instruct-v0.3",
273
+ "transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
274
+ },
275
+ {
276
+ "name": "microsoft/Phi-3-mini-4k-instruct",
277
+ "transferTo": "microsoft/Phi-3.5-mini-instruct"
278
+ }
297
279
]
298
280
PUBLIC_ORIGIN : " https://huggingface.co"
299
281
PUBLIC_SHARE_PREFIX : " https://hf.co/chat"
0 commit comments