|
|
@@ -1224,6 +1224,102 @@
|
|
|
"supports_system_messages": true,
|
|
|
"supports_tool_choice": true
|
|
|
},
|
|
|
+ "azure/eu/gpt-5-2025-08-07": {
|
|
|
+ "cache_read_input_token_cost": 1.375e-07,
|
|
|
+ "input_cost_per_token": 1.375e-06,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1.1e-05,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "azure/eu/gpt-5-mini-2025-08-07": {
|
|
|
+ "cache_read_input_token_cost": 2.75e-08,
|
|
|
+ "input_cost_per_token": 2.75e-07,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 2.2e-06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "azure/eu/gpt-5-nano-2025-08-07": {
|
|
|
+ "cache_read_input_token_cost": 5.5e-09,
|
|
|
+ "input_cost_per_token": 5.5e-08,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 4.4e-07,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
"azure/eu/o1-2024-12-17": {
|
|
|
"cache_read_input_token_cost": 8.25e-06,
|
|
|
"input_cost_per_token": 1.65e-05,
|
|
|
@@ -2738,14 +2834,14 @@
|
|
|
},
|
|
|
"azure/o3-2025-04-16": {
|
|
|
"deprecation_date": "2026-04-16",
|
|
|
- "cache_read_input_token_cost": 2.5e-06,
|
|
|
- "input_cost_per_token": 1e-05,
|
|
|
+ "cache_read_input_token_cost": 5e-07,
|
|
|
+ "input_cost_per_token": 2e-06,
|
|
|
"litellm_provider": "azure",
|
|
|
"max_input_tokens": 200000,
|
|
|
"max_output_tokens": 100000,
|
|
|
"max_tokens": 100000,
|
|
|
"mode": "chat",
|
|
|
- "output_cost_per_token": 4e-05,
|
|
|
+ "output_cost_per_token": 8e-06,
|
|
|
"supported_endpoints": [
|
|
|
"/v1/chat/completions",
|
|
|
"/v1/batch",
|
|
|
@@ -3004,6 +3100,107 @@
|
|
|
"litellm_provider": "azure",
|
|
|
"mode": "audio_speech"
|
|
|
},
|
|
|
+ "azure/us/gpt-4.1-2025-04-14": {
|
|
|
+ "deprecation_date": "2026-11-04",
|
|
|
+ "cache_read_input_token_cost": 5.5e-07,
|
|
|
+ "input_cost_per_token": 2.2e-06,
|
|
|
+ "input_cost_per_token_batches": 1.1e-06,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 1047576,
|
|
|
+ "max_output_tokens": 32768,
|
|
|
+ "max_tokens": 32768,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 8.8e-06,
|
|
|
+ "output_cost_per_token_batches": 4.4e-06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true,
|
|
|
+ "supports_web_search": false
|
|
|
+ },
|
|
|
+ "azure/us/gpt-4.1-mini-2025-04-14": {
|
|
|
+ "deprecation_date": "2026-11-04",
|
|
|
+ "cache_read_input_token_cost": 1.1e-07,
|
|
|
+ "input_cost_per_token": 4.4e-07,
|
|
|
+ "input_cost_per_token_batches": 2.2e-07,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 1047576,
|
|
|
+ "max_output_tokens": 32768,
|
|
|
+ "max_tokens": 32768,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1.76e-06,
|
|
|
+ "output_cost_per_token_batches": 8.8e-07,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true,
|
|
|
+ "supports_web_search": false
|
|
|
+ },
|
|
|
+ "azure/us/gpt-4.1-nano-2025-04-14": {
|
|
|
+ "deprecation_date": "2026-11-04",
|
|
|
+ "cache_read_input_token_cost": 2.5e-08,
|
|
|
+ "input_cost_per_token": 1.1e-07,
|
|
|
+ "input_cost_per_token_batches": 6e-08,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 1047576,
|
|
|
+ "max_output_tokens": 32768,
|
|
|
+ "max_tokens": 32768,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 4.4e-07,
|
|
|
+ "output_cost_per_token_batches": 2.2e-07,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
"azure/us/gpt-4o-2024-08-06": {
|
|
|
"deprecation_date": "2026-02-27",
|
|
|
"cache_read_input_token_cost": 1.375e-06,
|
|
|
@@ -3118,6 +3315,102 @@
|
|
|
"supports_system_messages": true,
|
|
|
"supports_tool_choice": true
|
|
|
},
|
|
|
+ "azure/us/gpt-5-2025-08-07": {
|
|
|
+ "cache_read_input_token_cost": 1.375e-07,
|
|
|
+ "input_cost_per_token": 1.375e-06,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1.1e-05,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "azure/us/gpt-5-mini-2025-08-07": {
|
|
|
+ "cache_read_input_token_cost": 2.75e-08,
|
|
|
+ "input_cost_per_token": 2.75e-07,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 2.2e-06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "azure/us/gpt-5-nano-2025-08-07": {
|
|
|
+ "cache_read_input_token_cost": 5.5e-09,
|
|
|
+ "input_cost_per_token": 5.5e-08,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 4.4e-07,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
"azure/us/o1-2024-12-17": {
|
|
|
"cache_read_input_token_cost": 8.25e-06,
|
|
|
"input_cost_per_token": 1.65e-05,
|
|
|
@@ -3163,6 +3456,36 @@
|
|
|
"supports_prompt_caching": true,
|
|
|
"supports_vision": false
|
|
|
},
|
|
|
+ "azure/us/o3-2025-04-16": {
|
|
|
+ "deprecation_date": "2026-04-16",
|
|
|
+ "cache_read_input_token_cost": 5.5e-07,
|
|
|
+ "input_cost_per_token": 2.2e-06,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 200000,
|
|
|
+ "max_output_tokens": 100000,
|
|
|
+ "max_tokens": 100000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 8.8e-06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/batch",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_parallel_function_calling": false,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
"azure/us/o3-mini-2025-01-31": {
|
|
|
"cache_read_input_token_cost": 6.05e-07,
|
|
|
"input_cost_per_token": 1.21e-06,
|
|
|
@@ -3179,6 +3502,23 @@
|
|
|
"supports_tool_choice": true,
|
|
|
"supports_vision": false
|
|
|
},
|
|
|
+ "azure/us/o4-mini-2025-04-16": {
|
|
|
+ "cache_read_input_token_cost": 3.1e-07,
|
|
|
+ "input_cost_per_token": 1.21e-06,
|
|
|
+ "litellm_provider": "azure",
|
|
|
+ "max_input_tokens": 200000,
|
|
|
+ "max_output_tokens": 100000,
|
|
|
+ "max_tokens": 100000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 4.84e-06,
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_parallel_function_calling": false,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
"azure/whisper-1": {
|
|
|
"input_cost_per_second": 0.0001,
|
|
|
"litellm_provider": "azure",
|
|
|
@@ -8515,10 +8855,50 @@
|
|
|
"/v1/images/generations"
|
|
|
]
|
|
|
},
|
|
|
+ "fal_ai/fal-ai/flux-pro/v1.1": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.04,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
"fal_ai/fal-ai/flux-pro/v1.1-ultra": {
|
|
|
"litellm_provider": "fal_ai",
|
|
|
"mode": "image_generation",
|
|
|
- "output_cost_per_image": 0.0398,
|
|
|
+ "output_cost_per_image": 0.06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
+ "fal_ai/fal-ai/flux/schnell": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.003,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
+ "fal_ai/fal-ai/bytedance/seedream/v3/text-to-image": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.03,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
+ "fal_ai/fal-ai/bytedance/dreamina/v3.1/text-to-image": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.03,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
+ "fal_ai/fal-ai/ideogram/v3": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.06,
|
|
|
"supported_endpoints": [
|
|
|
"/v1/images/generations"
|
|
|
]
|
|
|
@@ -8531,6 +8911,22 @@
|
|
|
"/v1/images/generations"
|
|
|
]
|
|
|
},
|
|
|
+ "fal_ai/fal-ai/imagen4/preview/fast": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.02,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
+ "fal_ai/fal-ai/imagen4/preview/ultra": {
|
|
|
+ "litellm_provider": "fal_ai",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "output_cost_per_image": 0.06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/images/generations"
|
|
|
+ ]
|
|
|
+ },
|
|
|
"fal_ai/fal-ai/recraft/v3/text-to-image": {
|
|
|
"litellm_provider": "fal_ai",
|
|
|
"mode": "image_generation",
|
|
|
@@ -9963,6 +10359,7 @@
|
|
|
"supports_function_calling": false,
|
|
|
"supports_parallel_function_calling": true,
|
|
|
"supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
"supports_response_schema": false,
|
|
|
"supports_system_messages": true,
|
|
|
"supports_tool_choice": true,
|
|
|
@@ -11568,6 +11965,7 @@
|
|
|
"supports_audio_output": true,
|
|
|
"supports_function_calling": true,
|
|
|
"supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
"supports_response_schema": true,
|
|
|
"supports_system_messages": true,
|
|
|
"supports_tool_choice": true,
|
|
|
@@ -11670,6 +12068,7 @@
|
|
|
"litellm_provider": "vertex_ai-language-models",
|
|
|
"max_audio_length_hours": 8.4,
|
|
|
"max_audio_per_prompt": 1,
|
|
|
+ "supports_reasoning": false,
|
|
|
"max_images_per_prompt": 3000,
|
|
|
"max_input_tokens": 32768,
|
|
|
"max_output_tokens": 32768,
|
|
|
@@ -13845,8 +14244,115 @@
|
|
|
"supports_reasoning": true,
|
|
|
"supports_response_schema": true,
|
|
|
"supports_system_messages": true,
|
|
|
- "supports_tool_choice": true,
|
|
|
- "supports_service_tier": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_service_tier": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "gpt-5.1": {
|
|
|
+ "cache_read_input_token_cost": 1.25e-07,
|
|
|
+ "cache_read_input_token_cost_priority": 2.5e-07,
|
|
|
+ "input_cost_per_token": 1.25e-06,
|
|
|
+ "input_cost_per_token_priority": 2.5e-06,
|
|
|
+ "litellm_provider": "openai",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1e-05,
|
|
|
+ "output_cost_per_token_priority": 2e-05,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_service_tier": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "gpt-5.1-2025-11-13": {
|
|
|
+ "cache_read_input_token_cost": 1.25e-07,
|
|
|
+ "cache_read_input_token_cost_priority": 2.5e-07,
|
|
|
+ "input_cost_per_token": 1.25e-06,
|
|
|
+ "input_cost_per_token_priority": 2.5e-06,
|
|
|
+ "litellm_provider": "openai",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1e-05,
|
|
|
+ "output_cost_per_token_priority": 2e-05,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_service_tier": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "gpt-5.1-chat-latest": {
|
|
|
+ "cache_read_input_token_cost": 1.25e-07,
|
|
|
+ "cache_read_input_token_cost_priority": 2.5e-07,
|
|
|
+ "input_cost_per_token": 1.25e-06,
|
|
|
+ "input_cost_per_token_priority": 2.5e-06,
|
|
|
+ "litellm_provider": "openai",
|
|
|
+ "max_input_tokens": 128000,
|
|
|
+ "max_output_tokens": 16384,
|
|
|
+ "max_tokens": 16384,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1e-05,
|
|
|
+ "output_cost_per_token_priority": 2e-05,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/chat/completions",
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": false,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": false,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": true,
|
|
|
+ "supports_tool_choice": false,
|
|
|
"supports_vision": true
|
|
|
},
|
|
|
"gpt-5-pro": {
|
|
|
@@ -14048,6 +14554,72 @@
|
|
|
"supports_tool_choice": true,
|
|
|
"supports_vision": true
|
|
|
},
|
|
|
+ "gpt-5.1-codex": {
|
|
|
+ "cache_read_input_token_cost": 1.25e-07,
|
|
|
+ "cache_read_input_token_cost_priority": 2.5e-07,
|
|
|
+ "input_cost_per_token": 1.25e-06,
|
|
|
+ "input_cost_per_token_priority": 2.5e-06,
|
|
|
+ "litellm_provider": "openai",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "responses",
|
|
|
+ "output_cost_per_token": 1e-05,
|
|
|
+ "output_cost_per_token_priority": 2e-05,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": false,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
+ "gpt-5.1-codex-mini": {
|
|
|
+ "cache_read_input_token_cost": 2.5e-08,
|
|
|
+ "cache_read_input_token_cost_priority": 4.5e-08,
|
|
|
+ "input_cost_per_token": 2.5e-07,
|
|
|
+ "input_cost_per_token_priority": 4.5e-07,
|
|
|
+ "litellm_provider": "openai",
|
|
|
+ "max_input_tokens": 272000,
|
|
|
+ "max_output_tokens": 128000,
|
|
|
+ "max_tokens": 128000,
|
|
|
+ "mode": "responses",
|
|
|
+ "output_cost_per_token": 2e-06,
|
|
|
+ "output_cost_per_token_priority": 3.6e-06,
|
|
|
+ "supported_endpoints": [
|
|
|
+ "/v1/responses"
|
|
|
+ ],
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "text"
|
|
|
+ ],
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_native_streaming": true,
|
|
|
+ "supports_parallel_function_calling": true,
|
|
|
+ "supports_pdf_input": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_system_messages": false,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_vision": true
|
|
|
+ },
|
|
|
"gpt-5-mini": {
|
|
|
"cache_read_input_token_cost": 2.5e-08,
|
|
|
"cache_read_input_token_cost_flex": 1.25e-08,
|
|
|
@@ -16199,6 +16771,21 @@
|
|
|
"supports_response_schema": true,
|
|
|
"supports_tool_choice": true
|
|
|
},
|
|
|
+ "mistral/magistral-medium-2509": {
|
|
|
+ "input_cost_per_token": 2e-06,
|
|
|
+ "litellm_provider": "mistral",
|
|
|
+ "max_input_tokens": 40000,
|
|
|
+ "max_output_tokens": 40000,
|
|
|
+ "max_tokens": 40000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 5e-06,
|
|
|
+ "source": "https://mistral.ai/news/magistral",
|
|
|
+ "supports_assistant_prefill": true,
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_response_schema": true,
|
|
|
+ "supports_tool_choice": true
|
|
|
+ },
|
|
|
"mistral/mistral-ocr-latest": {
|
|
|
"litellm_provider": "mistral",
|
|
|
"ocr_cost_per_page": 1e-3,
|
|
|
@@ -16624,6 +17211,20 @@
|
|
|
"source": "https://platform.moonshot.ai/docs/pricing",
|
|
|
"supports_vision": true
|
|
|
},
|
|
|
+ "moonshot/kimi-k2-thinking": {
|
|
|
+ "cache_read_input_token_cost": 1.5e-7,
|
|
|
+ "input_cost_per_token": 6e-7,
|
|
|
+ "litellm_provider": "moonshot",
|
|
|
+ "max_input_tokens": 262144,
|
|
|
+ "max_output_tokens": 262144,
|
|
|
+ "max_tokens": 262144,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 2.5e-6,
|
|
|
+ "source": "https://platform.moonshot.ai/docs/pricing/chat#generation-model-kimi-k2",
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_web_search": true
|
|
|
+ },
|
|
|
"moonshot/moonshot-v1-128k": {
|
|
|
"input_cost_per_token": 2e-06,
|
|
|
"litellm_provider": "moonshot",
|
|
|
@@ -18280,6 +18881,21 @@
|
|
|
"supports_reasoning": true,
|
|
|
"supports_tool_choice": true
|
|
|
},
|
|
|
+ "openrouter/deepseek/deepseek-v3.2-exp": {
|
|
|
+ "input_cost_per_token": 2e-07,
|
|
|
+ "input_cost_per_token_cache_hit": 2e-08,
|
|
|
+ "litellm_provider": "openrouter",
|
|
|
+ "max_input_tokens": 163840,
|
|
|
+ "max_output_tokens": 163840,
|
|
|
+ "max_tokens": 8192,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 4e-07,
|
|
|
+ "supports_assistant_prefill": true,
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_prompt_caching": true,
|
|
|
+ "supports_reasoning": false,
|
|
|
+ "supports_tool_choice": true
|
|
|
+ },
|
|
|
"openrouter/deepseek/deepseek-coder": {
|
|
|
"input_cost_per_token": 1.4e-07,
|
|
|
"litellm_provider": "openrouter",
|
|
|
@@ -18523,6 +19139,19 @@
|
|
|
"output_cost_per_token": 1e-06,
|
|
|
"supports_tool_choice": true
|
|
|
},
|
|
|
+ "openrouter/minimax/minimax-m2": {
|
|
|
+ "input_cost_per_token": 2.55e-7,
|
|
|
+ "litellm_provider": "openrouter",
|
|
|
+ "max_input_tokens": 204800,
|
|
|
+ "max_output_tokens": 204800,
|
|
|
+ "max_tokens": 32768,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1.02e-6,
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_prompt_caching": false,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_tool_choice": true
|
|
|
+ },
|
|
|
"openrouter/mistralai/mistral-7b-instruct": {
|
|
|
"input_cost_per_token": 1.3e-07,
|
|
|
"litellm_provider": "openrouter",
|
|
|
@@ -18994,15 +19623,16 @@
|
|
|
"supports_vision": true
|
|
|
},
|
|
|
"openrouter/qwen/qwen3-coder": {
|
|
|
- "input_cost_per_token": 1e-06,
|
|
|
+ "input_cost_per_token": 2.2e-7,
|
|
|
"litellm_provider": "openrouter",
|
|
|
- "max_input_tokens": 1000000,
|
|
|
- "max_output_tokens": 1000000,
|
|
|
- "max_tokens": 1000000,
|
|
|
+ "max_input_tokens": 262100,
|
|
|
+ "max_output_tokens": 262100,
|
|
|
+ "max_tokens": 262100,
|
|
|
"mode": "chat",
|
|
|
- "output_cost_per_token": 5e-06,
|
|
|
+ "output_cost_per_token": 9.5e-7,
|
|
|
"source": "https://openrouter.ai/qwen/qwen3-coder",
|
|
|
- "supports_tool_choice": true
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_function_calling": true
|
|
|
},
|
|
|
"openrouter/switchpoint/router": {
|
|
|
"input_cost_per_token": 8.5e-07,
|
|
|
@@ -19051,6 +19681,32 @@
|
|
|
"supports_tool_choice": true,
|
|
|
"supports_web_search": false
|
|
|
},
|
|
|
+ "openrouter/z-ai/glm-4.6": {
|
|
|
+ "input_cost_per_token": 4.0e-7,
|
|
|
+ "litellm_provider": "openrouter",
|
|
|
+ "max_input_tokens": 202800,
|
|
|
+ "max_output_tokens": 131000,
|
|
|
+ "max_tokens": 202800,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1.75e-6,
|
|
|
+ "source": "https://openrouter.ai/z-ai/glm-4.6",
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_tool_choice": true
|
|
|
+ },
|
|
|
+ "openrouter/z-ai/glm-4.6:exacto": {
|
|
|
+ "input_cost_per_token": 4.5e-7,
|
|
|
+ "litellm_provider": "openrouter",
|
|
|
+ "max_input_tokens": 202800,
|
|
|
+ "max_output_tokens": 131000,
|
|
|
+ "max_tokens": 202800,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 1.9e-6,
|
|
|
+ "source": "https://openrouter.ai/z-ai/glm-4.6:exacto",
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_reasoning": true,
|
|
|
+ "supports_tool_choice": true
|
|
|
+ },
|
|
|
"ovhcloud/DeepSeek-R1-Distill-Llama-70B": {
|
|
|
"input_cost_per_token": 6.7e-07,
|
|
|
"litellm_provider": "ovhcloud",
|
|
|
@@ -23148,6 +23804,19 @@
|
|
|
"supports_function_calling": true,
|
|
|
"supports_tool_choice": true
|
|
|
},
|
|
|
+ "vertex_ai/moonshotai/kimi-k2-thinking-maas": {
|
|
|
+ "input_cost_per_token": 6e-07,
|
|
|
+ "litellm_provider": "vertex_ai-moonshot_models",
|
|
|
+ "max_input_tokens": 256000,
|
|
|
+ "max_output_tokens": 256000,
|
|
|
+ "max_tokens": 256000,
|
|
|
+ "mode": "chat",
|
|
|
+ "output_cost_per_token": 2.5e-06,
|
|
|
+ "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing#partner-models",
|
|
|
+ "supports_function_calling": true,
|
|
|
+ "supports_tool_choice": true,
|
|
|
+ "supports_web_search": true
|
|
|
+ },
|
|
|
"vertex_ai/mistral-medium-3": {
|
|
|
"input_cost_per_token": 4e-07,
|
|
|
"litellm_provider": "vertex_ai-mistral_models",
|
|
|
@@ -23484,6 +24153,22 @@
|
|
|
"mode": "embedding",
|
|
|
"output_cost_per_token": 0.0
|
|
|
},
|
|
|
+ "voyage/voyage-3.5": {
|
|
|
+ "input_cost_per_token": 6e-08,
|
|
|
+ "litellm_provider": "voyage",
|
|
|
+ "max_input_tokens": 32000,
|
|
|
+ "max_tokens": 32000,
|
|
|
+ "mode": "embedding",
|
|
|
+ "output_cost_per_token": 0.0
|
|
|
+ },
|
|
|
+ "voyage/voyage-3.5-lite": {
|
|
|
+ "input_cost_per_token": 2e-08,
|
|
|
+ "litellm_provider": "voyage",
|
|
|
+ "max_input_tokens": 32000,
|
|
|
+ "max_tokens": 32000,
|
|
|
+ "mode": "embedding",
|
|
|
+ "output_cost_per_token": 0.0
|
|
|
+ },
|
|
|
"voyage/voyage-code-2": {
|
|
|
"input_cost_per_token": 1.2e-07,
|
|
|
"litellm_provider": "voyage",
|
|
|
@@ -24030,7 +24715,6 @@
|
|
|
"supports_parallel_function_calling": false,
|
|
|
"supports_vision": false
|
|
|
},
|
|
|
-
|
|
|
"whisper-1": {
|
|
|
"input_cost_per_second": 0.0001,
|
|
|
"litellm_provider": "openai",
|
|
|
@@ -24040,30 +24724,6 @@
|
|
|
"/v1/audio/transcriptions"
|
|
|
]
|
|
|
},
|
|
|
- "vertex_ai/qwen/qwen3-next-80b-a3b-instruct-maas": {
|
|
|
- "input_cost_per_token": 1.5e-07,
|
|
|
- "litellm_provider": "vertex_ai-qwen_models",
|
|
|
- "max_input_tokens": 262144,
|
|
|
- "max_output_tokens": 262144,
|
|
|
- "max_tokens": 262144,
|
|
|
- "mode": "chat",
|
|
|
- "output_cost_per_token": 1.2e-06,
|
|
|
- "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
|
|
- "supports_function_calling": true,
|
|
|
- "supports_tool_choice": true
|
|
|
- },
|
|
|
- "vertex_ai/qwen/qwen3-next-80b-a3b-thinking-maas": {
|
|
|
- "input_cost_per_token": 1.5e-07,
|
|
|
- "litellm_provider": "vertex_ai-qwen_models",
|
|
|
- "max_input_tokens": 262144,
|
|
|
- "max_output_tokens": 262144,
|
|
|
- "max_tokens": 262144,
|
|
|
- "mode": "chat",
|
|
|
- "output_cost_per_token": 1.2e-06,
|
|
|
- "source": "https://cloud.google.com/vertex-ai/generative-ai/pricing",
|
|
|
- "supports_function_calling": true,
|
|
|
- "supports_tool_choice": true
|
|
|
- },
|
|
|
"xai/grok-2": {
|
|
|
"input_cost_per_token": 2e-06,
|
|
|
"litellm_provider": "xai",
|
|
|
@@ -24537,5 +25197,116 @@
|
|
|
"1024x1792",
|
|
|
"1792x1024"
|
|
|
]
|
|
|
+ },
|
|
|
+ "runwayml/gen4_turbo": {
|
|
|
+ "litellm_provider": "runwayml",
|
|
|
+ "mode": "video_generation",
|
|
|
+ "output_cost_per_video_per_second": 0.05,
|
|
|
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "video"
|
|
|
+ ],
|
|
|
+ "supported_resolutions": [
|
|
|
+ "1280x720",
|
|
|
+ "720x1280"
|
|
|
+ ],
|
|
|
+ "metadata": {
|
|
|
+ "comment": "5 credits per second @ $0.01 per credit = $0.05 per second"
|
|
|
+ }
|
|
|
+ },
|
|
|
+ "runwayml/gen4_aleph": {
|
|
|
+ "litellm_provider": "runwayml",
|
|
|
+ "mode": "video_generation",
|
|
|
+ "output_cost_per_video_per_second": 0.15,
|
|
|
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "video"
|
|
|
+ ],
|
|
|
+ "supported_resolutions": [
|
|
|
+ "1280x720",
|
|
|
+ "720x1280"
|
|
|
+ ],
|
|
|
+ "metadata": {
|
|
|
+ "comment": "15 credits per second @ $0.01 per credit = $0.15 per second"
|
|
|
+ }
|
|
|
+ },
|
|
|
+ "runwayml/gen3a_turbo": {
|
|
|
+ "litellm_provider": "runwayml",
|
|
|
+ "mode": "video_generation",
|
|
|
+ "output_cost_per_video_per_second": 0.05,
|
|
|
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "video"
|
|
|
+ ],
|
|
|
+ "supported_resolutions": [
|
|
|
+ "1280x720",
|
|
|
+ "720x1280"
|
|
|
+ ],
|
|
|
+ "metadata": {
|
|
|
+ "comment": "5 credits per second @ $0.01 per credit = $0.05 per second"
|
|
|
+ }
|
|
|
+ },
|
|
|
+ "runwayml/gen4_image": {
|
|
|
+ "litellm_provider": "runwayml",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "input_cost_per_image": 0.05,
|
|
|
+ "output_cost_per_image": 0.05,
|
|
|
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_resolutions": [
|
|
|
+ "1280x720",
|
|
|
+ "1920x1080"
|
|
|
+ ],
|
|
|
+ "metadata": {
|
|
|
+ "comment": "5 credits per 720p image or 8 credits per 1080p image @ $0.01 per credit. Using 5 credits ($0.05) as base cost"
|
|
|
+ }
|
|
|
+ },
|
|
|
+ "runwayml/gen4_image_turbo": {
|
|
|
+ "litellm_provider": "runwayml",
|
|
|
+ "mode": "image_generation",
|
|
|
+ "input_cost_per_image": 0.02,
|
|
|
+ "output_cost_per_image": 0.02,
|
|
|
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
|
+ "supported_modalities": [
|
|
|
+ "text",
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_output_modalities": [
|
|
|
+ "image"
|
|
|
+ ],
|
|
|
+ "supported_resolutions": [
|
|
|
+ "1280x720",
|
|
|
+ "1920x1080"
|
|
|
+ ],
|
|
|
+ "metadata": {
|
|
|
+ "comment": "2 credits per image (any resolution) @ $0.01 per credit = $0.02 per image"
|
|
|
+ }
|
|
|
+ },
|
|
|
+ "runwayml/eleven_multilingual_v2": {
|
|
|
+ "litellm_provider": "runwayml",
|
|
|
+ "mode": "audio_speech",
|
|
|
+ "input_cost_per_character": 3e-07,
|
|
|
+ "source": "https://docs.dev.runwayml.com/guides/pricing/",
|
|
|
+ "metadata": {
|
|
|
+ "comment": "Estimated cost based on standard TTS pricing. RunwayML uses ElevenLabs models."
|
|
|
+ }
|
|
|
}
|
|
|
}
|