From 25b7bdc8473a79a2d26f03686b0cf593035c7171 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" <41898282+github-actions[bot]@users.noreply.github.com> Date: Wed, 14 Jan 2026 13:52:54 +0000 Subject: [PATCH] =?UTF-8?q?chore:=20=E5=90=8C=E6=AD=A5=E6=A8=A1=E5=9E=8B?= =?UTF-8?q?=E4=BB=B7=E6=A0=BC=E6=95=B0=E6=8D=AE=20(3f47198bec7bed0d3abaaed?= =?UTF-8?q?772d3c2a3b9f52d8157470238d4ea9ef7862bd585)?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- model_prices_and_context_window.json | 450 ++++++++++++------------- model_prices_and_context_window.sha256 | 2 +- 2 files changed, 226 insertions(+), 226 deletions(-) diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index cf2ee10c..26b30a93 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -2893,7 +2893,7 @@ "input_cost_per_token": 1.25e-06, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "azure", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -2964,7 +2964,7 @@ "input_cost_per_token": 1.25e-06, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "azure", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -2997,7 +2997,7 @@ "input_cost_per_token": 2.5e-07, "input_cost_per_token_priority": 4.5e-07, "litellm_provider": "azure", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -3535,7 +3535,7 @@ "input_cost_per_token": 1.75e-06, "input_cost_per_token_priority": 3.5e-06, "litellm_provider": "azure", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -8624,7 +8624,7 @@ "input_cost_per_token": 1.24999e-06, "input_dbu_cost_per_token": 1.7857e-05, "litellm_provider": "databricks", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "metadata": { @@ -8639,7 +8639,7 @@ "input_cost_per_token": 1.24999e-06, "input_dbu_cost_per_token": 1.7857e-05, "litellm_provider": "databricks", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "metadata": { @@ -8654,7 +8654,7 @@ "input_cost_per_token": 2.4997000000000006e-07, "input_dbu_cost_per_token": 3.571e-06, "litellm_provider": "databricks", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "metadata": { @@ -8669,7 +8669,7 @@ "input_cost_per_token": 4.998e-08, "input_dbu_cost_per_token": 7.14e-07, "litellm_provider": "databricks", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "metadata": { @@ -10103,7 +10103,7 @@ "supports_tool_choice": true }, "deepseek/deepseek-reasoner": { - "cache_read_input_token_cost": 2.8e-08, + "cache_read_input_token_cost": 2.8e-08, "input_cost_per_token": 2.8e-07, "input_cost_per_token_cache_hit": 2.8e-08, "litellm_provider": "deepseek", @@ -17496,7 +17496,7 @@ "input_cost_per_token_flex": 6.25e-07, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -17533,7 +17533,7 @@ "input_cost_per_token": 1.25e-06, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -17569,7 +17569,7 @@ "input_cost_per_token": 1.25e-06, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -17640,7 +17640,7 @@ "input_cost_per_token": 1.75e-06, "input_cost_per_token_priority": 3.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -17677,7 +17677,7 @@ "input_cost_per_token": 1.75e-06, "input_cost_per_token_priority": 3.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -17745,7 +17745,7 @@ "gpt-5.2-pro": { "input_cost_per_token": 2.1e-05, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -17776,7 +17776,7 @@ "gpt-5.2-pro-2025-12-11": { "input_cost_per_token": 2.1e-05, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -17808,7 +17808,7 @@ "input_cost_per_token": 1.5e-05, "input_cost_per_token_batches": 7.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 128000, "max_output_tokens": 272000, "max_tokens": 272000, "mode": "responses", @@ -17841,7 +17841,7 @@ "input_cost_per_token": 1.5e-05, "input_cost_per_token_batches": 7.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 128000, "max_output_tokens": 272000, "max_tokens": 272000, "mode": "responses", @@ -17878,7 +17878,7 @@ "input_cost_per_token_flex": 6.25e-07, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -17977,7 +17977,7 @@ "cache_read_input_token_cost": 1.25e-07, "input_cost_per_token": 1.25e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -18009,7 +18009,7 @@ "input_cost_per_token": 1.25e-06, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -18040,7 +18040,7 @@ "cache_read_input_token_cost": 1.25e-07, "input_cost_per_token": 1.25e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -18072,7 +18072,7 @@ "input_cost_per_token": 2.5e-07, "input_cost_per_token_priority": 4.5e-07, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "responses", @@ -18107,7 +18107,7 @@ "input_cost_per_token_flex": 1.25e-07, "input_cost_per_token_priority": 4.5e-07, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -18146,7 +18146,7 @@ "input_cost_per_token_flex": 1.25e-07, "input_cost_per_token_priority": 4.5e-07, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -18184,7 +18184,7 @@ "input_cost_per_token_flex": 2.5e-08, "input_cost_per_token_priority": 2.5e-06, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -18219,7 +18219,7 @@ "input_cost_per_token": 5e-08, "input_cost_per_token_flex": 2.5e-08, "litellm_provider": "openai", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -23251,7 +23251,7 @@ "cache_read_input_token_cost": 1.25e-07, "input_cost_per_token": 1.25e-06, "litellm_provider": "openrouter", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -23270,7 +23270,7 @@ "cache_read_input_token_cost": 1.25e-07, "input_cost_per_token": 1.25e-06, "litellm_provider": "openrouter", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -23289,7 +23289,7 @@ "cache_read_input_token_cost": 2.5e-08, "input_cost_per_token": 2.5e-07, "litellm_provider": "openrouter", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -23308,7 +23308,7 @@ "cache_read_input_token_cost": 5e-09, "input_cost_per_token": 5e-08, "litellm_provider": "openrouter", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -23328,7 +23328,7 @@ "cache_read_input_token_cost": 1.75e-07, "input_cost_per_token": 1.75e-06, "litellm_provider": "openrouter", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -23358,7 +23358,7 @@ "input_cost_per_image": 0, "input_cost_per_token": 2.1e-05, "litellm_provider": "openrouter", - "max_input_tokens": 400000, + "max_input_tokens": 272000, "max_output_tokens": 128000, "max_tokens": 128000, "mode": "chat", @@ -24478,7 +24478,7 @@ }, "replicate/openai/gpt-5": { "input_cost_per_token": 1.25e-06, - "output_cost_per_token": 1.0e-05, + "output_cost_per_token": 1e-05, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24489,7 +24489,7 @@ "supports_response_schema": true }, "replicateopenai/gpt-oss-20b": { - "input_cost_per_token": 9.0e-08, + "input_cost_per_token": 9e-08, "output_cost_per_token": 3.6e-07, "litellm_provider": "replicate", "mode": "chat", @@ -24497,8 +24497,8 @@ "supports_system_messages": true }, "replicate/anthropic/claude-4.5-haiku": { - "input_cost_per_token": 1.0e-06, - "output_cost_per_token": 5.0e-06, + "input_cost_per_token": 1e-06, + "output_cost_per_token": 5e-06, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24510,7 +24510,7 @@ "supports_prompt_caching": true }, "replicate/ibm-granite/granite-3.3-8b-instruct": { - "input_cost_per_token": 3.0e-08, + "input_cost_per_token": 3e-08, "output_cost_per_token": 2.5e-07, "litellm_provider": "replicate", "mode": "chat", @@ -24519,7 +24519,7 @@ }, "replicate/openai/gpt-4o": { "input_cost_per_token": 2.5e-06, - "output_cost_per_token": 1.0e-05, + "output_cost_per_token": 1e-05, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24532,9 +24532,9 @@ "supports_audio_output": true }, "replicate/openai/o4-mini": { - "input_cost_per_token": 1.0e-06, - "output_cost_per_token": 4.0e-06, - "output_cost_per_reasoning_token": 4.0e-06, + "input_cost_per_token": 1e-06, + "output_cost_per_token": 4e-06, + "output_cost_per_reasoning_token": 4e-06, "litellm_provider": "replicate", "mode": "chat", "supports_reasoning": true, @@ -24551,8 +24551,8 @@ }, "replicate/openai/o1": { "input_cost_per_token": 1.5e-05, - "output_cost_per_token": 6.0e-05, - "output_cost_per_reasoning_token": 6.0e-05, + "output_cost_per_token": 6e-05, + "output_cost_per_reasoning_token": 6e-05, "litellm_provider": "replicate", "mode": "chat", "supports_reasoning": true, @@ -24560,7 +24560,7 @@ }, "replicate/openai/gpt-4o-mini": { "input_cost_per_token": 1.5e-07, - "output_cost_per_token": 6.0e-07, + "output_cost_per_token": 6e-07, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24579,7 +24579,7 @@ "supports_system_messages": true }, "replicate/anthropic/claude-4-sonnet": { - "input_cost_per_token": 3.0e-06, + "input_cost_per_token": 3e-06, "output_cost_per_token": 1.5e-05, "litellm_provider": "replicate", "mode": "chat", @@ -24603,7 +24603,7 @@ "supports_system_messages": true }, "replicate/anthropic/claude-3.7-sonnet": { - "input_cost_per_token": 3.0e-06, + "input_cost_per_token": 3e-06, "output_cost_per_token": 1.5e-05, "litellm_provider": "replicate", "mode": "chat", @@ -24616,8 +24616,8 @@ "supports_prompt_caching": true }, "replicate/anthropic/claude-3.5-haiku": { - "input_cost_per_token": 1.0e-06, - "output_cost_per_token": 5.0e-06, + "input_cost_per_token": 1e-06, + "output_cost_per_token": 5e-06, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24642,7 +24642,7 @@ "supports_prompt_caching": true }, "replicate/google/gemini-3-pro": { - "input_cost_per_token": 2.0e-06, + "input_cost_per_token": 2e-06, "output_cost_per_token": 1.2e-05, "litellm_provider": "replicate", "mode": "chat", @@ -24654,7 +24654,7 @@ "supports_response_schema": true }, "replicate/anthropic/claude-4.5-sonnet": { - "input_cost_per_token": 3.0e-06, + "input_cost_per_token": 3e-06, "output_cost_per_token": 1.5e-05, "litellm_provider": "replicate", "mode": "chat", @@ -24667,8 +24667,8 @@ "supports_prompt_caching": true }, "replicate/openai/gpt-4.1": { - "input_cost_per_token": 2.0e-06, - "output_cost_per_token": 8.0e-06, + "input_cost_per_token": 2e-06, + "output_cost_per_token": 8e-06, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24679,15 +24679,15 @@ "supports_response_schema": true }, "replicate/openai/gpt-4.1-nano": { - "input_cost_per_token": 1.0e-07, - "output_cost_per_token": 4.0e-07, + "input_cost_per_token": 1e-07, + "output_cost_per_token": 4e-07, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, "supports_system_messages": true }, "replicate/openai/gpt-4.1-mini": { - "input_cost_per_token": 4.0e-07, + "input_cost_per_token": 4e-07, "output_cost_per_token": 1.6e-06, "litellm_provider": "replicate", "mode": "chat", @@ -24699,8 +24699,8 @@ "supports_response_schema": true }, "replicate/openai/gpt-5-nano": { - "input_cost_per_token": 5.0e-08, - "output_cost_per_token": 4.0e-07, + "input_cost_per_token": 5e-08, + "output_cost_per_token": 4e-07, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24708,7 +24708,7 @@ }, "replicate/openai/gpt-5-mini": { "input_cost_per_token": 2.5e-07, - "output_cost_per_token": 2.0e-06, + "output_cost_per_token": 2e-06, "litellm_provider": "replicate", "mode": "chat", "supports_function_calling": true, @@ -24760,8 +24760,8 @@ }, "replicate/deepseek-ai/deepseek-r1": { "input_cost_per_token": 3.75e-06, - "output_cost_per_token": 1.0e-05, - "output_cost_per_reasoning_token": 1.0e-05, + "output_cost_per_token": 1e-05, + "output_cost_per_reasoning_token": 1e-05, "litellm_provider": "replicate", "mode": "chat", "max_input_tokens": 65536, @@ -32578,8 +32578,8 @@ "novita/deepseek/deepseek-v3.2": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.69e-7, - "output_cost_per_token": 4e-7, + "input_cost_per_token": 2.69e-07, + "output_cost_per_token": 4e-07, "max_input_tokens": 163840, "max_output_tokens": 65536, "max_tokens": 65536, @@ -32588,14 +32588,14 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.345e-7, - "input_cost_per_token_cache_hit": 1.345e-7, + "cache_read_input_token_cost": 1.345e-07, + "input_cost_per_token_cache_hit": 1.345e-07, "supports_reasoning": true }, "novita/minimax/minimax-m2.1": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, "output_cost_per_token": 1.2e-06, "max_input_tokens": 204800, "max_output_tokens": 131072, @@ -32605,13 +32605,13 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 3e-8, - "input_cost_per_token_cache_hit": 3e-8 + "cache_read_input_token_cost": 3e-08, + "input_cost_per_token_cache_hit": 3e-08 }, "novita/zai-org/glm-4.7": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6e-7, + "input_cost_per_token": 6e-07, "output_cost_per_token": 2.2e-06, "max_input_tokens": 204800, "max_output_tokens": 131072, @@ -32621,15 +32621,15 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token_cache_hit": 1.1e-7, + "cache_read_input_token_cost": 1.1e-07, + "input_cost_per_token_cache_hit": 1.1e-07, "supports_reasoning": true }, "novita/xiaomimimo/mimo-v2-flash": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1e-7, - "output_cost_per_token": 3e-7, + "input_cost_per_token": 1e-07, + "output_cost_per_token": 3e-07, "max_input_tokens": 262144, "max_output_tokens": 32000, "max_tokens": 32000, @@ -32638,15 +32638,15 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 2e-8, - "input_cost_per_token_cache_hit": 2e-8, + "cache_read_input_token_cost": 2e-08, + "input_cost_per_token_cache_hit": 2e-08, "supports_reasoning": true }, "novita/zai-org/autoglm-phone-9b-multilingual": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3.5e-8, - "output_cost_per_token": 1.38e-7, + "input_cost_per_token": 3.5e-08, + "output_cost_per_token": 1.38e-07, "max_input_tokens": 65536, "max_output_tokens": 65536, "max_tokens": 65536, @@ -32656,7 +32656,7 @@ "novita/moonshotai/kimi-k2-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6e-7, + "input_cost_per_token": 6e-07, "output_cost_per_token": 2.5e-06, "max_input_tokens": 262144, "max_output_tokens": 262144, @@ -32671,7 +32671,7 @@ "novita/minimax/minimax-m2": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, "output_cost_per_token": 1.2e-06, "max_input_tokens": 204800, "max_output_tokens": 131072, @@ -32680,15 +32680,15 @@ "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_system_messages": true, - "cache_read_input_token_cost": 3e-8, - "input_cost_per_token_cache_hit": 3e-8, + "cache_read_input_token_cost": 3e-08, + "input_cost_per_token_cache_hit": 3e-08, "supports_reasoning": true }, "novita/paddlepaddle/paddleocr-vl": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2e-8, - "output_cost_per_token": 2e-8, + "input_cost_per_token": 2e-08, + "output_cost_per_token": 2e-08, "max_input_tokens": 16384, "max_output_tokens": 16384, "max_tokens": 16384, @@ -32698,8 +32698,8 @@ "novita/deepseek/deepseek-v3.2-exp": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 4.1e-7, + "input_cost_per_token": 2.7e-07, + "output_cost_per_token": 4.1e-07, "max_input_tokens": 163840, "max_output_tokens": 65536, "max_tokens": 65536, @@ -32713,7 +32713,7 @@ "novita/qwen/qwen3-vl-235b-a22b-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 9.8e-7, + "input_cost_per_token": 9.8e-07, "output_cost_per_token": 3.95e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -32725,8 +32725,8 @@ "novita/zai-org/glm-4.6v": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, - "output_cost_per_token": 9e-7, + "input_cost_per_token": 3e-07, + "output_cost_per_token": 9e-07, "max_input_tokens": 131072, "max_output_tokens": 32768, "max_tokens": 32768, @@ -32736,14 +32736,14 @@ "supports_vision": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 5.5e-8, - "input_cost_per_token_cache_hit": 5.5e-8, + "cache_read_input_token_cost": 5.5e-08, + "input_cost_per_token_cache_hit": 5.5e-08, "supports_reasoning": true }, "novita/zai-org/glm-4.6": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5.5e-7, + "input_cost_per_token": 5.5e-07, "output_cost_per_token": 2.2e-06, "max_input_tokens": 204800, "max_output_tokens": 131072, @@ -32753,14 +32753,14 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token_cache_hit": 1.1e-7, + "cache_read_input_token_cost": 1.1e-07, + "input_cost_per_token_cache_hit": 1.1e-07, "supports_reasoning": true }, "novita/kwaipilot/kat-coder-pro": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, "output_cost_per_token": 1.2e-06, "max_input_tokens": 256000, "max_output_tokens": 128000, @@ -32770,13 +32770,13 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 6e-8, - "input_cost_per_token_cache_hit": 6e-8 + "cache_read_input_token_cost": 6e-08, + "input_cost_per_token_cache_hit": 6e-08 }, "novita/qwen/qwen3-next-80b-a3b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.5e-7, + "input_cost_per_token": 1.5e-07, "output_cost_per_token": 1.5e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -32790,7 +32790,7 @@ "novita/qwen/qwen3-next-80b-a3b-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.5e-7, + "input_cost_per_token": 1.5e-07, "output_cost_per_token": 1.5e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -32805,8 +32805,8 @@ "novita/deepseek/deepseek-ocr": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-8, - "output_cost_per_token": 3e-8, + "input_cost_per_token": 3e-08, + "output_cost_per_token": 3e-08, "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, @@ -32818,7 +32818,7 @@ "novita/deepseek/deepseek-v3.1-terminus": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.7e-7, + "input_cost_per_token": 2.7e-07, "output_cost_per_token": 1e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -32828,14 +32828,14 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.35e-7, - "input_cost_per_token_cache_hit": 1.35e-7, + "cache_read_input_token_cost": 1.35e-07, + "input_cost_per_token_cache_hit": 1.35e-07, "supports_reasoning": true }, "novita/qwen/qwen3-vl-235b-a22b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, "output_cost_per_token": 1.5e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -32864,8 +32864,8 @@ "novita/skywork/r1v4-lite": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2e-7, - "output_cost_per_token": 6e-7, + "input_cost_per_token": 2e-07, + "output_cost_per_token": 6e-07, "max_input_tokens": 262144, "max_output_tokens": 65536, "max_tokens": 65536, @@ -32877,7 +32877,7 @@ "novita/deepseek/deepseek-v3.1": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.7e-7, + "input_cost_per_token": 2.7e-07, "output_cost_per_token": 1e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -32887,14 +32887,14 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.35e-7, - "input_cost_per_token_cache_hit": 1.35e-7, + "cache_read_input_token_cost": 1.35e-07, + "input_cost_per_token_cache_hit": 1.35e-07, "supports_reasoning": true }, "novita/moonshotai/kimi-k2-0905": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6e-7, + "input_cost_per_token": 6e-07, "output_cost_per_token": 2.5e-06, "max_input_tokens": 262144, "max_output_tokens": 262144, @@ -32908,7 +32908,7 @@ "novita/qwen/qwen3-coder-480b-a35b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, "output_cost_per_token": 1.3e-06, "max_input_tokens": 262144, "max_output_tokens": 65536, @@ -32922,8 +32922,8 @@ "novita/qwen/qwen3-coder-30b-a3b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-8, - "output_cost_per_token": 2.7e-7, + "input_cost_per_token": 7e-08, + "output_cost_per_token": 2.7e-07, "max_input_tokens": 160000, "max_output_tokens": 32768, "max_tokens": 32768, @@ -32936,8 +32936,8 @@ "novita/openai/gpt-oss-120b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5e-8, - "output_cost_per_token": 2.5e-7, + "input_cost_per_token": 5e-08, + "output_cost_per_token": 2.5e-07, "max_input_tokens": 131072, "max_output_tokens": 32768, "max_tokens": 32768, @@ -32952,7 +32952,7 @@ "novita/moonshotai/kimi-k2-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5.7e-7, + "input_cost_per_token": 5.7e-07, "output_cost_per_token": 2.3e-06, "max_input_tokens": 131072, "max_output_tokens": 131072, @@ -32966,7 +32966,7 @@ "novita/deepseek/deepseek-v3-0324": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.7e-7, + "input_cost_per_token": 2.7e-07, "output_cost_per_token": 1.12e-06, "max_input_tokens": 163840, "max_output_tokens": 163840, @@ -32976,13 +32976,13 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.35e-7, - "input_cost_per_token_cache_hit": 1.35e-7 + "cache_read_input_token_cost": 1.35e-07, + "input_cost_per_token_cache_hit": 1.35e-07 }, "novita/zai-org/glm-4.5": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6e-7, + "input_cost_per_token": 6e-07, "output_cost_per_token": 2.2e-06, "max_input_tokens": 131072, "max_output_tokens": 98304, @@ -32991,14 +32991,14 @@ "supports_parallel_function_calling": true, "supports_tool_choice": true, "supports_system_messages": true, - "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token_cache_hit": 1.1e-7, + "cache_read_input_token_cost": 1.1e-07, + "input_cost_per_token_cache_hit": 1.1e-07, "supports_reasoning": true }, "novita/qwen/qwen3-235b-a22b-thinking-2507": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, "output_cost_per_token": 3e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -33012,8 +33012,8 @@ "novita/meta-llama/llama-3.1-8b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2e-8, - "output_cost_per_token": 5e-8, + "input_cost_per_token": 2e-08, + "output_cost_per_token": 5e-08, "max_input_tokens": 16384, "max_output_tokens": 16384, "max_tokens": 16384, @@ -33022,8 +33022,8 @@ "novita/google/gemma-3-12b-it": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5e-8, - "output_cost_per_token": 1e-7, + "input_cost_per_token": 5e-08, + "output_cost_per_token": 1e-07, "max_input_tokens": 131072, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33035,7 +33035,7 @@ "novita/zai-org/glm-4.5v": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6e-7, + "input_cost_per_token": 6e-07, "output_cost_per_token": 1.8e-06, "max_input_tokens": 65536, "max_output_tokens": 16384, @@ -33046,15 +33046,15 @@ "supports_vision": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 1.1e-7, - "input_cost_per_token_cache_hit": 1.1e-7, + "cache_read_input_token_cost": 1.1e-07, + "input_cost_per_token_cache_hit": 1.1e-07, "supports_reasoning": true }, "novita/openai/gpt-oss-20b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 4e-8, - "output_cost_per_token": 1.5e-7, + "input_cost_per_token": 4e-08, + "output_cost_per_token": 1.5e-07, "max_input_tokens": 131072, "max_output_tokens": 32768, "max_tokens": 32768, @@ -33067,8 +33067,8 @@ "novita/qwen/qwen3-235b-a22b-instruct-2507": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 9e-8, - "output_cost_per_token": 5.8e-7, + "input_cost_per_token": 9e-08, + "output_cost_per_token": 5.8e-07, "max_input_tokens": 131072, "max_output_tokens": 16384, "max_tokens": 16384, @@ -33081,8 +33081,8 @@ "novita/deepseek/deepseek-r1-distill-qwen-14b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.5e-7, - "output_cost_per_token": 1.5e-7, + "input_cost_per_token": 1.5e-07, + "output_cost_per_token": 1.5e-07, "max_input_tokens": 32768, "max_output_tokens": 16384, "max_tokens": 16384, @@ -33094,8 +33094,8 @@ "novita/meta-llama/llama-3.3-70b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.35e-7, - "output_cost_per_token": 4e-7, + "input_cost_per_token": 1.35e-07, + "output_cost_per_token": 4e-07, "max_input_tokens": 131072, "max_output_tokens": 120000, "max_tokens": 120000, @@ -33107,8 +33107,8 @@ "novita/qwen/qwen-2.5-72b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3.8e-7, - "output_cost_per_token": 4e-7, + "input_cost_per_token": 3.8e-07, + "output_cost_per_token": 4e-07, "max_input_tokens": 32000, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33121,8 +33121,8 @@ "novita/mistralai/mistral-nemo": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 4e-8, - "output_cost_per_token": 1.7e-7, + "input_cost_per_token": 4e-08, + "output_cost_per_token": 1.7e-07, "max_input_tokens": 60288, "max_output_tokens": 16000, "max_tokens": 16000, @@ -33133,7 +33133,7 @@ "novita/minimaxai/minimax-m1-80k": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5.5e-7, + "input_cost_per_token": 5.5e-07, "output_cost_per_token": 2.2e-06, "max_input_tokens": 1000000, "max_output_tokens": 40000, @@ -33147,7 +33147,7 @@ "novita/deepseek/deepseek-r1-0528": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-7, + "input_cost_per_token": 7e-07, "output_cost_per_token": 2.5e-06, "max_input_tokens": 163840, "max_output_tokens": 32768, @@ -33157,15 +33157,15 @@ "supports_tool_choice": true, "supports_system_messages": true, "supports_response_schema": true, - "cache_read_input_token_cost": 3.5e-7, - "input_cost_per_token_cache_hit": 3.5e-7, + "cache_read_input_token_cost": 3.5e-07, + "input_cost_per_token_cache_hit": 3.5e-07, "supports_reasoning": true }, "novita/deepseek/deepseek-r1-distill-qwen-32b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-7, - "output_cost_per_token": 3e-7, + "input_cost_per_token": 3e-07, + "output_cost_per_token": 3e-07, "max_input_tokens": 64000, "max_output_tokens": 32000, "max_tokens": 32000, @@ -33177,8 +33177,8 @@ "novita/meta-llama/llama-3-8b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 4e-8, - "output_cost_per_token": 4e-8, + "input_cost_per_token": 4e-08, + "output_cost_per_token": 4e-08, "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33187,8 +33187,8 @@ "novita/microsoft/wizardlm-2-8x22b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6.2e-7, - "output_cost_per_token": 6.2e-7, + "input_cost_per_token": 6.2e-07, + "output_cost_per_token": 6.2e-07, "max_input_tokens": 65535, "max_output_tokens": 8000, "max_tokens": 8000, @@ -33197,8 +33197,8 @@ "novita/deepseek/deepseek-r1-0528-qwen3-8b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 6e-8, - "output_cost_per_token": 9e-8, + "input_cost_per_token": 6e-08, + "output_cost_per_token": 9e-08, "max_input_tokens": 128000, "max_output_tokens": 32000, "max_tokens": 32000, @@ -33208,8 +33208,8 @@ "novita/deepseek/deepseek-r1-distill-llama-70b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 8e-7, - "output_cost_per_token": 8e-7, + "input_cost_per_token": 8e-07, + "output_cost_per_token": 8e-07, "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33221,8 +33221,8 @@ "novita/meta-llama/llama-3-70b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5.1e-7, - "output_cost_per_token": 7.4e-7, + "input_cost_per_token": 5.1e-07, + "output_cost_per_token": 7.4e-07, "max_input_tokens": 8192, "max_output_tokens": 8000, "max_tokens": 8000, @@ -33233,8 +33233,8 @@ "novita/qwen/qwen3-235b-a22b-fp8": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2e-7, - "output_cost_per_token": 8e-7, + "input_cost_per_token": 2e-07, + "output_cost_per_token": 8e-07, "max_input_tokens": 40960, "max_output_tokens": 20000, "max_tokens": 20000, @@ -33244,8 +33244,8 @@ "novita/meta-llama/llama-4-maverick-17b-128e-instruct-fp8": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.7e-7, - "output_cost_per_token": 8.5e-7, + "input_cost_per_token": 2.7e-07, + "output_cost_per_token": 8.5e-07, "max_input_tokens": 1048576, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33255,8 +33255,8 @@ "novita/meta-llama/llama-4-scout-17b-16e-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.8e-7, - "output_cost_per_token": 5.9e-7, + "input_cost_per_token": 1.8e-07, + "output_cost_per_token": 5.9e-07, "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, @@ -33266,8 +33266,8 @@ "novita/nousresearch/hermes-2-pro-llama-3-8b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.4e-7, - "output_cost_per_token": 1.4e-7, + "input_cost_per_token": 1.4e-07, + "output_cost_per_token": 1.4e-07, "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33278,8 +33278,8 @@ "novita/qwen/qwen2.5-vl-72b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 8e-7, - "output_cost_per_token": 8e-7, + "input_cost_per_token": 8e-07, + "output_cost_per_token": 8e-07, "max_input_tokens": 32768, "max_output_tokens": 32768, "max_tokens": 32768, @@ -33302,8 +33302,8 @@ "novita/baidu/ernie-4.5-21B-a3b-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-8, - "output_cost_per_token": 2.8e-7, + "input_cost_per_token": 7e-08, + "output_cost_per_token": 2.8e-07, "max_input_tokens": 131072, "max_output_tokens": 65536, "max_tokens": 65536, @@ -33313,8 +33313,8 @@ "novita/sao10k/l3-8b-lunaris": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5e-8, - "output_cost_per_token": 5e-8, + "input_cost_per_token": 5e-08, + "output_cost_per_token": 5e-08, "max_input_tokens": 8192, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33325,8 +33325,8 @@ "novita/baichuan/baichuan-m2-32b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-8, - "output_cost_per_token": 7e-8, + "input_cost_per_token": 7e-08, + "output_cost_per_token": 7e-08, "max_input_tokens": 131072, "max_output_tokens": 131072, "max_tokens": 131072, @@ -33335,7 +33335,7 @@ "novita/baidu/ernie-4.5-vl-424b-a47b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 4.2e-7, + "input_cost_per_token": 4.2e-07, "output_cost_per_token": 1.25e-06, "max_input_tokens": 123000, "max_output_tokens": 16000, @@ -33347,7 +33347,7 @@ "novita/baidu/ernie-4.5-300b-a47b-paddle": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.8e-7, + "input_cost_per_token": 2.8e-07, "output_cost_per_token": 1.1e-06, "max_input_tokens": 123000, "max_output_tokens": 12000, @@ -33359,7 +33359,7 @@ "novita/deepseek/deepseek-prover-v2-671b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-7, + "input_cost_per_token": 7e-07, "output_cost_per_token": 2.5e-06, "max_input_tokens": 160000, "max_output_tokens": 160000, @@ -33369,8 +33369,8 @@ "novita/qwen/qwen3-32b-fp8": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1e-7, - "output_cost_per_token": 4.5e-7, + "input_cost_per_token": 1e-07, + "output_cost_per_token": 4.5e-07, "max_input_tokens": 40960, "max_output_tokens": 20000, "max_tokens": 20000, @@ -33380,8 +33380,8 @@ "novita/qwen/qwen3-30b-a3b-fp8": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 9e-8, - "output_cost_per_token": 4.5e-7, + "input_cost_per_token": 9e-08, + "output_cost_per_token": 4.5e-07, "max_input_tokens": 40960, "max_output_tokens": 20000, "max_tokens": 20000, @@ -33391,8 +33391,8 @@ "novita/google/gemma-3-27b-it": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.19e-7, - "output_cost_per_token": 2e-7, + "input_cost_per_token": 1.19e-07, + "output_cost_per_token": 2e-07, "max_input_tokens": 98304, "max_output_tokens": 16384, "max_tokens": 16384, @@ -33402,7 +33402,7 @@ "novita/deepseek/deepseek-v3-turbo": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 4e-7, + "input_cost_per_token": 4e-07, "output_cost_per_token": 1.3e-06, "max_input_tokens": 64000, "max_output_tokens": 16000, @@ -33415,7 +33415,7 @@ "novita/deepseek/deepseek-r1-turbo": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-7, + "input_cost_per_token": 7e-07, "output_cost_per_token": 2.5e-06, "max_input_tokens": 64000, "max_output_tokens": 16000, @@ -33429,8 +33429,8 @@ "novita/Sao10K/L3-8B-Stheno-v3.2": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 5e-8, - "output_cost_per_token": 5e-8, + "input_cost_per_token": 5e-08, + "output_cost_per_token": 5e-08, "max_input_tokens": 8192, "max_output_tokens": 32000, "max_tokens": 32000, @@ -33442,8 +33442,8 @@ "novita/gryphe/mythomax-l2-13b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 9e-8, - "output_cost_per_token": 9e-8, + "input_cost_per_token": 9e-08, + "output_cost_per_token": 9e-08, "max_input_tokens": 4096, "max_output_tokens": 3200, "max_tokens": 3200, @@ -33452,8 +33452,8 @@ "novita/baidu/ernie-4.5-vl-28b-a3b-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3.9e-7, - "output_cost_per_token": 3.9e-7, + "input_cost_per_token": 3.9e-07, + "output_cost_per_token": 3.9e-07, "max_input_tokens": 131072, "max_output_tokens": 65536, "max_tokens": 65536, @@ -33468,8 +33468,8 @@ "novita/qwen/qwen3-vl-8b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 8e-8, - "output_cost_per_token": 5e-7, + "input_cost_per_token": 8e-08, + "output_cost_per_token": 5e-07, "max_input_tokens": 131072, "max_output_tokens": 32768, "max_tokens": 32768, @@ -33483,8 +33483,8 @@ "novita/zai-org/glm-4.5-air": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.3e-7, - "output_cost_per_token": 8.5e-7, + "input_cost_per_token": 1.3e-07, + "output_cost_per_token": 8.5e-07, "max_input_tokens": 131072, "max_output_tokens": 98304, "max_tokens": 98304, @@ -33497,8 +33497,8 @@ "novita/qwen/qwen3-vl-30b-a3b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2e-7, - "output_cost_per_token": 7e-7, + "input_cost_per_token": 2e-07, + "output_cost_per_token": 7e-07, "max_input_tokens": 131072, "max_output_tokens": 32768, "max_tokens": 32768, @@ -33512,7 +33512,7 @@ "novita/qwen/qwen3-vl-30b-a3b-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2e-7, + "input_cost_per_token": 2e-07, "output_cost_per_token": 1e-06, "max_input_tokens": 131072, "max_output_tokens": 32768, @@ -33527,8 +33527,8 @@ "novita/qwen/qwen3-omni-30b-a3b-thinking": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 9.7e-7, + "input_cost_per_token": 2.5e-07, + "output_cost_per_token": 9.7e-07, "max_input_tokens": 65536, "max_output_tokens": 16384, "max_tokens": 16384, @@ -33544,8 +33544,8 @@ "novita/qwen/qwen3-omni-30b-a3b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 9.7e-7, + "input_cost_per_token": 2.5e-07, + "output_cost_per_token": 9.7e-07, "max_input_tokens": 65536, "max_output_tokens": 16384, "max_tokens": 16384, @@ -33561,8 +33561,8 @@ "novita/qwen/qwen-mt-plus": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 2.5e-7, - "output_cost_per_token": 7.5e-7, + "input_cost_per_token": 2.5e-07, + "output_cost_per_token": 7.5e-07, "max_input_tokens": 16384, "max_output_tokens": 8192, "max_tokens": 8192, @@ -33571,8 +33571,8 @@ "novita/baidu/ernie-4.5-vl-28b-a3b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 1.4e-7, - "output_cost_per_token": 5.6e-7, + "input_cost_per_token": 1.4e-07, + "output_cost_per_token": 5.6e-07, "max_input_tokens": 30000, "max_output_tokens": 8000, "max_tokens": 8000, @@ -33586,8 +33586,8 @@ "novita/baidu/ernie-4.5-21B-a3b": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-8, - "output_cost_per_token": 2.8e-7, + "input_cost_per_token": 7e-08, + "output_cost_per_token": 2.8e-07, "max_input_tokens": 120000, "max_output_tokens": 8000, "max_tokens": 8000, @@ -33599,8 +33599,8 @@ "novita/qwen/qwen3-8b-fp8": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3.5e-8, - "output_cost_per_token": 1.38e-7, + "input_cost_per_token": 3.5e-08, + "output_cost_per_token": 1.38e-07, "max_input_tokens": 128000, "max_output_tokens": 20000, "max_tokens": 20000, @@ -33610,8 +33610,8 @@ "novita/qwen/qwen3-4b-fp8": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-8, - "output_cost_per_token": 3e-8, + "input_cost_per_token": 3e-08, + "output_cost_per_token": 3e-08, "max_input_tokens": 128000, "max_output_tokens": 20000, "max_tokens": 20000, @@ -33621,8 +33621,8 @@ "novita/qwen/qwen2.5-7b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 7e-8, - "output_cost_per_token": 7e-8, + "input_cost_per_token": 7e-08, + "output_cost_per_token": 7e-08, "max_input_tokens": 32000, "max_output_tokens": 32000, "max_tokens": 32000, @@ -33635,8 +33635,8 @@ "novita/meta-llama/llama-3.2-3b-instruct": { "litellm_provider": "novita", "mode": "chat", - "input_cost_per_token": 3e-8, - "output_cost_per_token": 5e-8, + "input_cost_per_token": 3e-08, + "output_cost_per_token": 5e-08, "max_input_tokens": 32768, "max_output_tokens": 32000, "max_tokens": 32000, @@ -33661,7 +33661,7 @@ "novita/qwen/qwen3-embedding-0.6b": { "litellm_provider": "novita", "mode": "embedding", - "input_cost_per_token": 7e-8, + "input_cost_per_token": 7e-08, "output_cost_per_token": 0, "max_input_tokens": 32768, "max_output_tokens": 32768, @@ -33670,7 +33670,7 @@ "novita/qwen/qwen3-embedding-8b": { "litellm_provider": "novita", "mode": "embedding", - "input_cost_per_token": 7e-8, + "input_cost_per_token": 7e-08, "output_cost_per_token": 0, "max_input_tokens": 32768, "max_output_tokens": 4096, @@ -33679,8 +33679,8 @@ "novita/baai/bge-m3": { "litellm_provider": "novita", "mode": "embedding", - "input_cost_per_token": 1e-8, - "output_cost_per_token": 1e-8, + "input_cost_per_token": 1e-08, + "output_cost_per_token": 1e-08, "max_input_tokens": 8192, "max_output_tokens": 96000, "max_tokens": 96000 @@ -33688,8 +33688,8 @@ "novita/qwen/qwen3-reranker-8b": { "litellm_provider": "novita", "mode": "rerank", - "input_cost_per_token": 5e-8, - "output_cost_per_token": 5e-8, + "input_cost_per_token": 5e-08, + "output_cost_per_token": 5e-08, "max_input_tokens": 32768, "max_output_tokens": 4096, "max_tokens": 4096 @@ -33697,8 +33697,8 @@ "novita/baai/bge-reranker-v2-m3": { "litellm_provider": "novita", "mode": "rerank", - "input_cost_per_token": 1e-8, - "output_cost_per_token": 1e-8, + "input_cost_per_token": 1e-08, + "output_cost_per_token": 1e-08, "max_input_tokens": 8000, "max_output_tokens": 8000, "max_tokens": 8000 @@ -33878,4 +33878,4 @@ "litellm_provider": "llamagate", "mode": "embedding" } -} +} \ No newline at end of file diff --git a/model_prices_and_context_window.sha256 b/model_prices_and_context_window.sha256 index 10b0f27f..ea8d8cb2 100644 --- a/model_prices_and_context_window.sha256 +++ b/model_prices_and_context_window.sha256 @@ -1 +1 @@ -4b7cbba1dd9dcec6881fba5d75ef8320a4c189d3da1944913570404a6e0adde5 +3f47198bec7bed0d3abaaed772d3c2a3b9f52d8157470238d4ea9ef7862bd585