From 359543394b3efc51690e0c58373de07aea58b62d Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Tue, 10 Jun 2025 16:24:21 -0700 Subject: [PATCH 1/3] build(model_prices_and_context_window.json): add o3-pro pricing --- ...odel_prices_and_context_window_backup.json | 19 +++++++++++++++++++ model_prices_and_context_window.json | 19 +++++++++++++++++++ 2 files changed, 38 insertions(+) diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index e88637a67876..325f37376fe7 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -740,6 +740,25 @@ "supports_tool_choice": true, "supports_reasoning": true }, + "o3-pro": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true + }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index e88637a67876..325f37376fe7 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -740,6 +740,25 @@ "supports_tool_choice": true, "supports_reasoning": true }, + "o3-pro": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true + }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, From cc777b99f299113bf18bd3108b288abe310611d1 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Tue, 10 Jun 2025 16:31:20 -0700 Subject: [PATCH 2/3] build(model_prices_and_context_window.json): add updated o3 model pricing --- ...odel_prices_and_context_window_backup.json | 55 ++++++++++++++++--- model_prices_and_context_window.json | 55 ++++++++++++++++--- 2 files changed, 92 insertions(+), 18 deletions(-) diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 325f37376fe7..d32460a21c9d 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -757,15 +757,26 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -775,15 +786,28 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-2025-04-16": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -793,7 +817,20 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-mini": { "max_tokens": 100000, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 325f37376fe7..d32460a21c9d 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -757,15 +757,26 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -775,15 +786,28 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-2025-04-16": { "max_tokens": 100000, "max_input_tokens": 200000, "max_output_tokens": 100000, - "input_cost_per_token": 1e-05, - "output_cost_per_token": 4e-05, - "cache_read_input_token_cost": 2.5e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, + "cache_read_input_token_cost": 0.5e-06, "litellm_provider": "openai", "mode": "chat", "supports_function_calling": true, @@ -793,7 +817,20 @@ "supports_prompt_caching": true, "supports_response_schema": true, "supports_reasoning": true, - "supports_tool_choice": true + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/chat/completions", + "/v1/completions", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] }, "o3-mini": { "max_tokens": 100000, From cbb7e53406bd15fc77e374e95c38edab4e79f93b Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Tue, 10 Jun 2025 16:32:46 -0700 Subject: [PATCH 3/3] build(model_prices_and_context_window.json): add new o3-pro model version --- ...odel_prices_and_context_window_backup.json | 30 +++++++++++++++++++ model_prices_and_context_window.json | 30 +++++++++++++++++++ 2 files changed, 60 insertions(+) diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index d32460a21c9d..b2425b469acc 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -770,6 +770,36 @@ "text" ] }, + "o3-pro-2025-06-10": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] + }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index d32460a21c9d..b2425b469acc 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -770,6 +770,36 @@ "text" ] }, + "o3-pro-2025-06-10": { + "max_tokens": 100000, + "max_input_tokens": 200000, + "max_output_tokens": 100000, + "input_cost_per_token": 20e-06, + "input_cost_per_token_batches": 10e-06, + "output_cost_per_token_batches": 40e-06, + "output_cost_per_token": 80e-06, + "litellm_provider": "openai", + "mode": "chat", + "supports_function_calling": true, + "supports_parallel_function_calling": false, + "supports_vision": true, + "supports_pdf_input": true, + "supports_prompt_caching": true, + "supports_response_schema": true, + "supports_reasoning": true, + "supports_tool_choice": true, + "supported_endpoints": [ + "/v1/responses", + "/v1/batch" + ], + "supported_modalities": [ + "text", + "image" + ], + "supported_output_modalities": [ + "text" + ] + }, "o3": { "max_tokens": 100000, "max_input_tokens": 200000,