diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index e88637a67876..b2425b469acc 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -740,13 +740,73 @@ "supports_tool_choice": true, "supports_reasoning": true }, + "o3-pro": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] + }, + "o3-pro-2025-06-10": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] + }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -756,15 +816,28 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-2025-04-16": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -774,7 +847,20 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-mini": { "max_tokens": 100000, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index e88637a67876..b2425b469acc 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -740,13 +740,73 @@ "supports_tool_choice": true, "supports_reasoning": true }, + "o3-pro": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] + }, + "o3-pro-2025-06-10": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] + }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -756,15 +816,28 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-2025-04-16": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -774,7 +847,20 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-mini": { "max_tokens": 100000,