@@ -4,7 +4,9 @@ MODELS=`[
4
4
{
5
5
"name" : "mistralai/Mixtral-8x7B-Instruct-v0.1",
6
6
"description" : "The latest MoE model from Mistral AI! 8x7B and outperforms Llama 2 70B in most benchmarks.",
7
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
7
8
"websiteUrl" : "https://mistral.ai/news/mixtral-of-experts/",
9
+ "modelUrl": "https://huggingface.co/mistralai/Mixtral-8x7B-Instruct-v0.1",
8
10
"preprompt" : "",
9
11
"chatPromptTemplate": "<s> {{#each messages}}{{#ifUser}}[INST]{{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}} {{content}} [/INST]{{/ifUser}}{{#ifAssistant}} {{content}}</s> {{/ifAssistant}}{{/each}}",
10
12
"parameters" : {
@@ -29,10 +31,39 @@ MODELS=`[
29
31
}
30
32
]
31
33
},
32
- {
34
+ {
35
+ "name" : "google/gemma-7b-it",
36
+ "description": "Gemma 7B belongs to a family of lightweight models built by Google, based on the same research and technology used to create the Gemini models.",
37
+ "websiteUrl" : "https://blog.google/technology/developers/gemma-open-models/",
38
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/google-logo.png",
39
+ "modelUrl": "https://huggingface.co/google/gemma-7b-it",
40
+ "preprompt": "",
41
+ "chatPromptTemplate" : "{{#each messages}}{{#ifUser}}<start_of_turn>user\n{{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}}<end_of_turn>\n<start_of_turn>model\n{{/ifUser}}{{#ifAssistant}}{{content}}<end_of_turn>\n{{/ifAssistant}}{{/each}}",
42
+ "promptExamples": [
43
+ {
44
+ "title": "Write an email from bullet list",
45
+ "prompt": "As a restaurant owner, write a professional email to the supplier to get these products every week: \n\n- Wine (x10)\n- Eggs (x24)\n- Bread (x12)"
46
+ }, {
47
+ "title": "Code a snake game",
48
+ "prompt": "Code a basic snake game in python, give explanations for each step."
49
+ }, {
50
+ "title": "Assist in a task",
51
+ "prompt": "How do I make a delicious lemon cheesecake?"
52
+ }
53
+ ],
54
+ "parameters": {
55
+ "do_sample": true,
56
+ "truncate": 7168,
57
+ "max_new_tokens": 1024,
58
+ "stop" : ["<end_of_turn>"]
59
+ }
60
+ },
61
+ {
33
62
"name": "meta-llama/Llama-2-70b-chat-hf",
34
63
"description": "The latest and biggest model from Meta, fine-tuned for chat.",
64
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
35
65
"websiteUrl": "https://ai.meta.com/llama/",
66
+ "modelUrl": "https://huggingface.co/meta-llama/Llama-2-70b-chat-hf",
36
67
"preprompt": " ",
37
68
"chatPromptTemplate" : "<s>[INST] <<SYS>>\n{{preprompt}}\n<</SYS>>\n\n{{#each messages}}{{#ifUser}}{{content}} [/INST] {{/ifUser}}{{#ifAssistant}}{{content}} </s><s>[INST] {{/ifAssistant}}{{/each}}",
38
69
"promptExamples": [
@@ -60,7 +91,9 @@ MODELS=`[
60
91
{
61
92
"name" : "NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
62
93
"description" : "Nous Hermes 2 Mixtral 8x7B DPO is the new flagship Nous Research model trained over the Mixtral 8x7B MoE LLM.",
94
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/nous-logo.png",
63
95
"websiteUrl" : "https://nousresearch.com/",
96
+ "modelUrl": "https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO",
64
97
"chatPromptTemplate" : "{{#if @root.preprompt}}<|im_start|>system\n{{@root.preprompt}}<|im_end|>\n{{/if}}{{#each messages}}{{#ifUser}}<|im_start|>user\n{{content}}<|im_end|>\n<|im_start|>assistant\n{{/ifUser}}{{#ifAssistant}}{{content}}<|im_end|>\n{{/ifAssistant}}{{/each}}",
65
98
"promptExamples": [
66
99
{
@@ -88,7 +121,9 @@ MODELS=`[
88
121
"name": "codellama/CodeLlama-70b-Instruct-hf",
89
122
"displayName": "codellama/CodeLlama-70b-Instruct-hf",
90
123
"description": "Code Llama, a state of the art code model from Meta. Now in 70B!",
124
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/meta-logo.png",
91
125
"websiteUrl": "https://ai.meta.com/blog/code-llama-large-language-model-coding/",
126
+ "modelUrl": "https://huggingface.co/codellama/CodeLlama-70b-Instruct-hf",
92
127
"preprompt": "",
93
128
"chatPromptTemplate" : "<s>{{#if @root.preprompt}}Source: system\n\n {{@root.preprompt}} <step> {{/if}}{{#each messages}}{{#ifUser}}Source: user\n\n {{content}} <step> {{/ifUser}}{{#ifAssistant}}Source: assistant\n\n {{content}} <step> {{/ifAssistant}}{{/each}}Source: assistant\nDestination: user\n\n ",
94
129
"promptExamples": [
@@ -117,7 +152,9 @@ MODELS=`[
117
152
"name": "mistralai/Mistral-7B-Instruct-v0.1",
118
153
"displayName": "mistralai/Mistral-7B-Instruct-v0.1",
119
154
"description": "Mistral 7B is a new Apache 2.0 model, released by Mistral AI that outperforms Llama2 13B in benchmarks.",
155
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
120
156
"websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/",
157
+ "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.1",
121
158
"preprompt": "",
122
159
"chatPromptTemplate" : "<s>{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}</s>{{/ifAssistant}}{{/each}}",
123
160
"parameters": {
@@ -147,7 +184,9 @@ MODELS=`[
147
184
"name": "mistralai/Mistral-7B-Instruct-v0.2",
148
185
"displayName": "mistralai/Mistral-7B-Instruct-v0.2",
149
186
"description": "Mistral 7B is a new Apache 2.0 model, released by Mistral AI that outperforms Llama2 13B in benchmarks.",
187
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/mistral-logo.png",
150
188
"websiteUrl": "https://mistral.ai/news/announcing-mistral-7b/",
189
+ "modelUrl": "https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2",
151
190
"preprompt": "",
152
191
"chatPromptTemplate" : "<s>{{#each messages}}{{#ifUser}}[INST] {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}} [/INST]{{/ifUser}}{{#ifAssistant}}{{content}}</s>{{/ifAssistant}}{{/each}}",
153
192
"parameters": {
@@ -176,7 +215,9 @@ MODELS=`[
176
215
"name": "openchat/openchat-3.5-0106",
177
216
"displayName": "openchat/openchat-3.5-0106",
178
217
"description": "OpenChat 3.5 is the #1 model on MT-Bench, with only 7B parameters.",
218
+ "logoUrl": "https://huggingface.co/datasets/huggingchat/models-logo/resolve/main/openchat-logo.png",
179
219
"websiteUrl": "https://huggingface.co/openchat/openchat-3.5-0106",
220
+ "modelUrl": "https://huggingface.co/openchat/openchat-3.5-0106",
180
221
"preprompt": "",
181
222
"chatPromptTemplate" : "<s>{{#each messages}}{{#ifUser}}GPT4 Correct User: {{#if @first}}{{#if @root.preprompt}}{{@root.preprompt}}\n{{/if}}{{/if}}{{content}}<|end_of_turn|>GPT4 Correct Assistant:{{/ifUser}}{{#ifAssistant}}{{content}}<|end_of_turn|>{{/ifAssistant}}{{/each}}",
182
223
"parameters": {
0 commit comments