Skip to content

Commit ae64e16

Browse files
committed
update modelmap to include gcp models with overlap for aws and docker. review updates
1 parent 6598200 commit ae64e16

File tree

5 files changed

+44
-35
lines changed

5 files changed

+44
-35
lines changed

src/api/modelmapper.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@ def load_model_map():
1313

1414
def get_model(provider, model):
1515
provider = provider.lower()
16-
model = model.lower()
16+
model = model.lower().removesuffix(":latest")
1717

1818
available_models = _model_map.get(provider, {})
1919
return available_models.get(model, model)

src/api/routers/test_vertex.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -209,7 +209,7 @@ def test_get_chat_completion_model_name_known_chat_model():
209209
# But for now, test as is
210210
result = vertex.get_chat_completion_model_name(model_alias)
211211
# Should remove 'publishers/' and 'models/' from the string
212-
assert result == "google/gemini-2.0-flash-lite-001"
212+
assert result == "gemini-2.0-flash-lite-001"
213213

214214
def test_get_chat_completion_model_name_unknown_model():
215215
model_alias = "some-other-model"

src/api/routers/vertex.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,8 +14,6 @@
1414
from api.modelmapper import get_model
1515

1616
known_chat_models = [
17-
"publishers/meta-llama/models/llama-3-1-8b-instruct",
18-
"publishers/meta-llama/models/llama-3-8b-instruct",
1917
"publishers/mistral-ai/models/mistral-7b-instruct-v0.3",
2018
"publishers/mistral-ai/models/mistral-nemo-instruct-2407",
2119
"publishers/mistral-ai/models/mistral-nemo@2407",

src/api/test_modelmapper.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,6 @@
55
@patch("api.modelmapper._model_map", {
66
"provider1": {
77
"model1": "mapped_model1",
8-
"model1:latest": "mapped_model1",
98
"model2": "mapped_model2"
109
}
1110
})

src/data/modelmap.json

Lines changed: 42 additions & 30 deletions
Original file line numberDiff line numberDiff line change
@@ -1,54 +1,66 @@
11
{
22
"aws": {
3-
"ai/mistral": "mistral.mistral-7b-instruct-v0:2",
4-
"ai/mistral:latest": "mistral.mistral-7b-instruct-v0:2",
5-
"ai/mistral:7B-Q4_0": "mistral.mistral-7b-instruct-v0:2",
6-
"ai/mistral:7B-F16": "mistral.mistral-7b-instruct-v0:2",
7-
"ai/mistral:7B-Q4_K_M": "mistral.mistral-7b-instruct-v0:2",
3+
"ai/claude3-haiku": "anthropic.claude-3-haiku-20240307-v1:0",
84
"ai/deepseek-r1-distill-llama": "deepseek.r1-v1:0",
9-
"ai/deepseek-r1-distill-llama:latest": "deepseek.r1-v1:0",
105
"ai/deepseek-r1-distill-llama:70B-Q4_0": "deepseek.r1-v1:0",
11-
"ai/deepseek-r1-distill-llama:8B-Q4_0": "deepseek.r1-v1:0",
126
"ai/deepseek-r1-distill-llama:70B-Q4_K_M": "deepseek.r1-v1:0",
137
"ai/deepseek-r1-distill-llama:8B-F16": "deepseek.r1-v1:0",
8+
"ai/deepseek-r1-distill-llama:8B-Q4_0": "deepseek.r1-v1:0",
149
"ai/deepseek-r1-distill-llama:8B-Q4_K_M": "deepseek.r1-v1:0",
10+
"ai/llama3.1": "meta.llama3-1-8b-instruct-v1:0",
11+
"ai/llama3.1:8B-F16": "meta.llama3-1-8b-instruct-v1:0",
12+
"ai/llama3.1:8B-Q4_K_M": "meta.llama3-1-8b-instruct-v1:0",
1513
"ai/llama3.2": "meta.llama3-2-1b-instruct-v1:0",
16-
"ai/llama3.2:latest": "meta.llama3-2-1b-instruct-v1:0",
17-
"ai/llama3.2:3B-Q4_0": "meta.llama3-2-3b-instruct-v1:0",
14+
"ai/llama3.2:1B-F16": "meta.llama3-2-1b-instruct-v1:0",
1815
"ai/llama3.2:1B-Q4_0": "meta.llama3-2-1b-instruct-v1:0",
16+
"ai/llama3.2:1B-Q8_0": "meta.llama3-2-1b-instruct-v1:0",
1917
"ai/llama3.2:3B-F16": "meta.llama3-2-3b-instruct-v1:0",
18+
"ai/llama3.2:3B-Q4_0": "meta.llama3-2-3b-instruct-v1:0",
2019
"ai/llama3.2:3B-Q4_K_M": "meta.llama3-2-3b-instruct-v1:0",
21-
"ai/llama3.2:1B-F16": "meta.llama3-2-1b-instruct-v1:0",
22-
"ai/llama3.2:1B-Q8_0": "meta.llama3-2-1b-instruct-v1:0",
2320
"ai/llama3.3": "meta.llama3-3-70b-instruct-v1:0",
24-
"ai/llama3.3:latest": "meta.llama3-3-70b-instruct-v1:0",
2521
"ai/llama3.3:70B-Q4_0": "meta.llama3-3-70b-instruct-v1:0",
2622
"ai/llama3.3:70B-Q4_K_M": "meta.llama3-3-70b-instruct-v1:0",
27-
"ai/llama3.1": "meta.llama3-1-8b-instruct-v1:0",
28-
"ai/llama3.1:latest": "meta.llama3-1-8b-instruct-v1:0",
29-
"ai/llama3.1:8B-F16": "meta.llama3-1-8b-instruct-v1:0",
30-
"ai/llama3.1:8B-Q4_K_M": "meta.llama3-1-8b-instruct-v1:0",
31-
"ai/claude3-haiku": "anthropic.claude-3-haiku-20240307-v1:0",
32-
"ai/claude3-haiku:latest": "anthropic.claude-3-haiku-20240307-v1:0",
23+
"ai/mistral": "mistral.mistral-7b-instruct-v0:2",
24+
"ai/mistral:7B-F16": "mistral.mistral-7b-instruct-v0:2",
25+
"ai/mistral:7B-Q4_0": "mistral.mistral-7b-instruct-v0:2",
26+
"ai/mistral:7B-Q4_K_M": "mistral.mistral-7b-instruct-v0:2",
3327
"ai/smollm2": "us.amazon.nova-micro-v1:0",
34-
"ai/smollm2:latest": "us.amazon.nova-micro-v1:0",
28+
"nova-lite": "use.amazon.nova-lite-v1:0",
29+
"nova-micro": "us.amazon.nova-micro-v1:0",
30+
"nova-premier-v1": "amazon.nova-premier-v1:0",
31+
"nova-pro": "us.amazon.nova-pro-v1:0",
3532
"default": "us.amazon.nova-micro-v1:0"
3633
},
3734
"gcp": {
35+
"ai/claude3-haiku": "publishers/anthropic/models/claude-3-haiku@20240307",
36+
"ai/deepseek-r1-distill-llama": "publishers/deepseek-ai/models/deepseek-ai/deepseek-r1-distill-llama-70b",
37+
"ai/deepseek-r1-distill-llama:70B-Q4_0": "publishers/deepseek-ai/models/deepseek-ai/deepseek-r1-distill-llama-70b",
38+
"ai/deepseek-r1-distill-llama:70B-Q4_K_M": "publishers/deepseek-ai/models/deepseek-ai/deepseek-r1-distill-llama-70b",
39+
"ai/deepseek-r1-distill-llama:8B-F16": "publishers/deepseek-ai/models/deepseek-ai/deepseek-ai/deepseek-r1-distill-llama-8b",
40+
"ai/deepseek-r1-distill-llama:8B-Q4_0": "publishers/deepseek-ai/models/deepseek-ai/deepseek-r1-distill-llama-8b",
41+
"ai/deepseek-r1-distill-llama:8B-Q4_K_M": "publishers/deepseek-ai/models/deepseek-ai/deepseek-r1-distill-llama-8b",
42+
"ai/llama3.1": "publishers/meta/models/llama-3-1-8b-instruct",
43+
"ai/llama3.1:8B-F16": "publishers/meta/models/llama-3-1-8b-instruct",
44+
"ai/llama3.1:8B-Q4_K_M": "publishers/meta/models/llama-3-1-8b-instruct",
45+
"ai/llama3.2": "publishers/meta/models/llama-3-2-90b-vision",
46+
"ai/llama3.2:1B-F16": "publishers/meta/models/llama-3-2-90b-vision",
47+
"ai/llama3.2:1B-Q4_0": "publishers/meta/models/llama-3-2-90b-vision",
48+
"ai/llama3.2:1B-Q8_0": "publishers/meta/models/llama-3-2-90b-vision",
49+
"ai/llama3.2:3B-F16": "publishers/meta/models/llama-3-2-90b-vision",
50+
"ai/llama3.2:3B-Q4_0": "publishers/meta/models/llama-3-2-90b-vision",
51+
"ai/llama3.2:3B-Q4_K_M": "publishers/meta/models/llama-3-2-90b-vision",
52+
"ai/llama3.3": "publishers/meta/models/llama-3-3-70b-instruct",
53+
"ai/llama3.3:70B-Q4_0": "publishers/meta/models/llama-3-3-70b-instruct",
54+
"ai/llama3.3:70B-Q4_K_M": "publishers/meta/models/llama-3-3-70b-instruct",
55+
"ai/mistral": "publishers/mistralai/models/mistral-nemo",
56+
"ai/mistral:7B-F16": "publishers/mistralai/models/mistral-nemo",
57+
"ai/mistral:7B-Q4_0": "publishers/mistralai/models/mistral-nemo",
58+
"ai/mistral:7B-Q4_K_M": "publishers/mistralai/models/mistral-nemo",
3859
"ai/smollm2": "publishers/google/models/gemini-2.0-flash-lite-001",
39-
"ai/smollm2:latest": "publishers/google/models/gemini-2.0-flash-lite-001",
40-
"gemini-2.5-pro": "publishers/google/models/gemini-2.5-pro-preview-05-06",
41-
"gemini-2.5-pro:latest": "publishers/google/models/gemini-2.5-pro-preview-05-06",
42-
"gemini-2.5-flash": "publishers/google/models/gemini-2.5-flash-preview-05-20",
43-
"gemini-2.5-flash:latest": "publishers/google/models/gemini-2.5-flash-preview-05-20",
4460
"gemini-2.0-flash": "publishers/google/models/gemini-2.0-flash-001",
45-
"gemini-2.0-flash:latest": "publishers/google/models/gemini-2.0-flash-001",
4661
"gemini-2.0-flash-lite": "publishers/google/models/gemini-2.0-flash-lite-001",
47-
"gemini-2.0-flash-lite:latest": "publishers/google/models/gemini-2.0-flash-lite-001",
48-
"claude-3-haiku": "publishers/anthropic/models/claude-3-haiku@20240307",
49-
"claude-3-haiku:latest": "publishers/anthropic/models/claude-3-haiku@20240307",
50-
"ai/mistral": "publishers/mistralai/models/mistral-nemo",
51-
"a/mistral:latest": "publishers/mistralai/models/mistral-nemo",
62+
"gemini-2.5-flash": "publishers/google/models/gemini-2.5-flash-preview-05-20",
63+
"gemini-2.5-pro": "publishers/google/models/gemini-2.5-pro-preview-05-06",
5264
"default": "publishers/google/models/gemini-2.0-flash-lite-001"
5365
}
5466
}

0 commit comments

Comments
 (0)