Skip to content

Error "Model {modelName} is enabled only for api versions 2024-12-01-preview and later" when using o3-mini model with azure-ai-inference #40759

@nischays

Description

@nischays
  • Package Name: azure-ai-inference
  • Package Version: 1.0.0b9
  • Python Version: 3.11

Describe the bug
When attempting to use the o3-mini model with ChatCompletionsClient, an error occurs:

Traceback (most recent call last):
  File "/workspaces/testaiinference/main.py", line 13, in <module>
    response = client.complete(
               ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/azure/ai/inference/_patch.py", line 738, in complete
    raise HttpResponseError(response=response)
Error: (BadRequest) Model {modelName} is enabled only for api versions 2024-12-01-preview and later
Code: BadRequest
Message: Model {modelName} is enabled only for api versions 2024-12-01-preview and later

When calling the API using the api version 2024-12-01-preview, the error is presented:

Traceback (most recent call last):
  File "/workspaces/testaiinference/main.py", line 13, in <module>
    response = client.complete(
               ^^^^^^^^^^^^^^^^
  File "/usr/local/lib/python3.11/site-packages/azure/ai/inference/_patch.py", line 737, in complete
    map_error(status_code=response.status_code, response=response, error_map=error_map)
  File "/usr/local/lib/python3.11/site-packages/azure/core/exceptions.py", line 163, in map_error
    raise error
Error: (404) Resource not found
Code: 404
Message: Resource not found

To Reproduce
Steps to reproduce the behavior:

  1. Set the environment variables AZURE_INFERENCE_ENDPOINT and AZURE_INFERENCE_KEY correctly.
  2. Run the following code with and without providing api_version:
import os
from azure.ai.inference import ChatCompletionsClient
from azure.core.credentials import AzureKeyCredential
from azure.ai.inference.models import SystemMessage, UserMessage

try:
    client = ChatCompletionsClient(
        endpoint=os.getenv("AZURE_INFERENCE_ENDPOINT"),
        credential=AzureKeyCredential(os.getenv("AZURE_INFERENCE_KEY")),
        #api_version="2024-12-01-preview",
    )

    response = client.complete(
        messages=[
            SystemMessage(content="You are a helpful assistant."),
            UserMessage(content="What is the capital of France?"),
        ],
        model="o3-mini",
    )

    print("azure-ai-inference response:", response.choices[0].message.content)

except Exception as e:
    print(f"Error: {e}")

Expected behavior
I expected the API call to succeed and return a chat completion response from the o3-mini model.
Instead, it fails due to an API version mismatch.

Screenshots
N/A (Error is text-based and printed to console)

Additional context

  • Other models (e.g., gpt-4o) work without issues using the same setup.
  • The same model works with the AzureOpenAI library

Metadata

Metadata

Assignees

Labels

AI Model InferenceIssues related to the client library for Azure AI Model Inference (\sdk\ai\azure-ai-inference)ClientThis issue points to a problem in the data-plane of the library.Service AttentionWorkflow: This issue is responsible by Azure service team.customer-reportedIssues that are reported by GitHub users external to the Azure organization.needs-team-attentionWorkflow: This issue needs attention from Azure service team or SDK teamquestionThe issue doesn't require a change to the product in order to be resolved. Most issues start as that

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions