@@ -483,12 +483,12 @@ def test_list_containers(self, mock_get_container_config):
483
483
],
484
484
"inference" : [
485
485
{
486
- "name" : "dsmc:/ /odsc-vllm-serving " ,
487
- "version" : "0.4.1.3 " ,
488
- "display_name" : "VLLM :0.4.1 " ,
489
- "family" : "odsc-vllm -serving" ,
490
- "platforms" : ["NVIDIA_GPU " ],
491
- "model_formats" : ["SAFETENSORS " ],
486
+ "name" : "iad.ocir.io/ociodscdev /odsc-llama-cpp-python-aio-linux_arm64_v8 " ,
487
+ "version" : "0.2.75.5 " ,
488
+ "display_name" : "LLAMA-CPP :0.2.75 " ,
489
+ "family" : "odsc-llama-cpp -serving" ,
490
+ "platforms" : ["ARM_CPU " ],
491
+ "model_formats" : ["GGUF " ],
492
492
},
493
493
{
494
494
"name" : "dsmc://odsc-text-generation-inference" ,
@@ -499,12 +499,12 @@ def test_list_containers(self, mock_get_container_config):
499
499
"model_formats" : ["SAFETENSORS" ],
500
500
},
501
501
{
502
- "name" : "iad.ocir.io/ociodscdev /odsc-llama-cpp-python-aio-linux_arm64_v8 " ,
503
- "version" : "0.2.75.5 " ,
504
- "display_name" : "LLAMA-CPP :0.2.75 " ,
505
- "family" : "odsc-llama-cpp -serving" ,
506
- "platforms" : ["ARM_CPU " ],
507
- "model_formats" : ["GGUF " ],
502
+ "name" : "dsmc:/ /odsc-vllm-serving " ,
503
+ "version" : "0.4.1.3 " ,
504
+ "display_name" : "VLLM :0.4.1 " ,
505
+ "family" : "odsc-vllm -serving" ,
506
+ "platforms" : ["NVIDIA_GPU " ],
507
+ "model_formats" : ["SAFETENSORS " ],
508
508
},
509
509
],
510
510
"finetune" : [
0 commit comments