Skip to content

Commit c42d5f7

Browse files
authored
feat(huggingchat): Model update 24/09/2024 (#1485)
1 parent 7423bf0 commit c42d5f7

File tree

1 file changed

+53
-71
lines changed

1 file changed

+53
-71
lines changed

chart/env/prod.yaml

Lines changed: 53 additions & 71 deletions
Original file line numberDiff line numberDiff line change
@@ -53,9 +53,8 @@ envVars:
5353
"tools": true,
5454
"preprompt": "",
5555
"parameters": {
56-
"temperature": 0.6,
57-
"top_p": 0.9,
5856
"stop": ["<|endoftext|>", "<|eot_id|>"],
57+
"temperature": 0.6,
5958
"max_new_tokens": 1024,
6059
"truncate": 7167
6160
},
@@ -115,42 +114,10 @@ envVars:
115114
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/qwen-logo.png",
116115
"preprompt": "You are Qwen, created by Alibaba Cloud. You are a helpful assistant.",
117116
"parameters": {
118-
"temperature": 0.6,
119117
"stop": ["<|endoftext|>", "<|im_end|>"],
120-
"truncate": 28672,
121-
"max_new_tokens": 3072,
122-
},
123-
"promptExamples": [
124-
{
125-
"title": "Write an email from bullet list",
126-
"prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)"
127-
},
128-
{
129-
"title": "Code a snake game",
130-
"prompt": "Code a basic snake game in python, give explanations for each step."
131-
},
132-
{
133-
"title": "Assist in a task",
134-
"prompt": "How do I make a delicious lemon cheesecake?"
135-
}
136-
]
137-
},
138-
{
139-
"name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
140-
"description": "A high-quality sparse mixture of experts model with open weights.",
141-
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
142-
"websiteUrl": "https://mistral.ai/news/mixtral-of-experts/",
143-
"modelUrl": "https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1",
144-
"tokenizer": "mistralai/Mixtral-8x7B-Instruct-v0.1",
145-
"preprompt": "",
146-
"parameters": {
147118
"temperature": 0.6,
148-
"top_p": 0.95,
149-
"repetition_penalty": 1.2,
150-
"top_k": 50,
151-
"truncate": 24576,
152-
"max_new_tokens": 8192,
153-
"stop": ["</s>"]
119+
"truncate": 28672,
120+
"max_new_tokens": 3072
154121
},
155122
"promptExamples": [
156123
{
@@ -168,12 +135,12 @@ envVars:
168135
]
169136
},
170137
{
171-
"name": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
172-
"description": "Nous Hermes' strong flagship model trained on the Mixtral 8x7B.",
138+
"name": "NousResearch/Hermes-3-Llama-3.1-8B",
139+
"description": "Nous Research's latest Hermes 3 release in 8B size. Follows instruction closely.",
173140
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png",
174141
"websiteUrl": "https://nousresearch.com/",
175-
"modelUrl": "https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
176-
"tokenizer": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
142+
"modelUrl": "https://huggingface.co/NousResearch/Hermes-3-Llama-3.1-8B",
143+
"tokenizer": "NousResearch/Hermes-3-Llama-3.1-8B",
177144
"promptExamples": [
178145
{
179146
"title": "Write an email from bullet list",
@@ -189,32 +156,26 @@ envVars:
189156
}
190157
],
191158
"parameters": {
192-
"temperature": 0.7,
193-
"top_p": 0.95,
194-
"repetition_penalty": 1,
195-
"top_k": 50,
196-
"truncate": 24576,
197-
"max_new_tokens": 2048,
198-
"stop": ["<|im_end|>"]
159+
"stop": ["<|im_end|>"],
160+
"temperature": 0.6,
161+
"truncate": 14336,
162+
"max_new_tokens": 1536
199163
}
200164
},
201165
{
202-
"name": "mistralai/Mistral-7B-Instruct-v0.3",
203-
"displayName": "mistralai/Mistral-7B-Instruct-v0.3",
166+
"name": "mistralai/Mistral-Nemo-Instruct-2407",
167+
"tokenizer": "mistralai/Mistral-Nemo-Instruct-2407",
168+
"displayName": "mistralai/Mistral-Nemo-Instruct-2407",
204169
"description": "A small model with good capabilities in language understanding and commonsense reasoning.",
205170
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
206-
"websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/",
207-
"modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.3",
208-
"tokenizer": "mistralai/Mistral-7B-Instruct-v0.3",
171+
"websiteUrl": "https://mistral.ai/news/mistral-nemo/",
172+
"modelUrl": "https://huggingface.co/mistralai/Mistral-Nemo-Instruct-2407",
209173
"preprompt": "",
210174
"parameters": {
211-
"temperature": 0.3,
212-
"top_p": 0.95,
213-
"repetition_penalty": 1.2,
214-
"top_k": 50,
215-
"truncate": 3072,
216-
"max_new_tokens": 1024,
217-
"stop": ["</s>"]
175+
"stop": ["</s>"],
176+
"temperature": 0.6,
177+
"truncate": 14336,
178+
"max_new_tokens": 1536
218179
},
219180
"promptExamples": [
220181
{
@@ -232,18 +193,18 @@ envVars:
232193
]
233194
},
234195
{
235-
"name": "microsoft/Phi-3-mini-4k-instruct",
236-
"tokenizer": "microsoft/Phi-3-mini-4k-instruct",
196+
"name": "microsoft/Phi-3.5-mini-instruct",
197+
"tokenizer": "microsoft/Phi-3.5-mini-instruct",
237198
"description": "One of the best small models (3.8B parameters), super fast for simple tasks.",
238199
"logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/microsoft-logo.png",
239-
"modelUrl": "https://huggingface.co/microsoft/Phi-3-mini-4k-instruct",
240-
"websiteUrl": "https://azure.microsoft.com/en-us/blog/introducing-phi-3-redefining-whats-possible-with-slms/",
200+
"modelUrl": "https://huggingface.co/microsoft/Phi-3.5-mini-instruct",
201+
"websiteUrl": "https://techcommunity.microsoft.com/t5/ai-azure-ai-services-blog/discover-the-new-multi-lingual-high-quality-phi-3-5-slms/ba-p/4225280/",
241202
"preprompt": "",
242203
"parameters": {
243204
"stop": ["<|end|>", "<|endoftext|>", "<|assistant|>"],
244-
"temperature": 0.7,
245-
"max_new_tokens": 1024,
246-
"truncate": 3071
205+
"temperature": 0.6,
206+
"truncate": 28672,
207+
"max_new_tokens": 3072
247208
},
248209
"promptExamples": [
249210
{
@@ -268,7 +229,6 @@ envVars:
268229
},
269230
"parameters": {
270231
"temperature": 0.6,
271-
"top_p": 0.9,
272232
"stop": ["<|endoftext|>", "<|eot_id|>"]
273233
},
274234
"unlisted": true
@@ -290,10 +250,32 @@ envVars:
290250
{ "name": "meta-llama/Llama-2-70b-chat-hf" },
291251
{ "name": "codellama/CodeLlama-70b-Instruct-hf" },
292252
{ "name": "openchat/openchat-3.5-0106" },
293-
{ "name": "meta-llama/Meta-Llama-3-70B-Instruct"},
294-
{ "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8"},
295-
{ "name": "CohereForAI/c4ai-command-r-plus", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"},
296-
{ "name": "01-ai/Yi-1.5-34B-Chat", "transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"}
253+
{ "name": "meta-llama/Meta-Llama-3-70B-Instruct" },
254+
{ "name": "meta-llama/Meta-Llama-3.1-405B-Instruct-FP8" },
255+
{
256+
"name": "CohereForAI/c4ai-command-r-plus",
257+
"transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
258+
},
259+
{
260+
"name": "01-ai/Yi-1.5-34B-Chat",
261+
"transferTo": "CohereForAI/c4ai-command-r-plus-08-2024"
262+
},
263+
{
264+
"name": "mistralai/Mixtral-8x7B-Instruct-v0.1",
265+
"transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
266+
},
267+
{
268+
"name": "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
269+
"transferTo": "NousResearch/Hermes-3-Llama-3.1-8B"
270+
},
271+
{
272+
"name": "mistralai/Mistral-7B-Instruct-v0.3",
273+
"transferTo": "mistralai/Mistral-Nemo-Instruct-2407"
274+
},
275+
{
276+
"name": "microsoft/Phi-3-mini-4k-instruct",
277+
"transferTo": "microsoft/Phi-3.5-mini-instruct"
278+
}
297279
]
298280
PUBLIC_ORIGIN: "https://huggingface.co"
299281
PUBLIC_SHARE_PREFIX: "https://hf.co/chat"

0 commit comments

Comments
 (0)