mirror of
https://github.com/Wei-Shaw/claude-relay-service.git
synced 2026-01-23 19:09:15 +00:00
chore: 同步模型价格数据 (f64b517b218c0a125b7d75fb268de7705f8a66d7ec861ba8b45b5bf9b72b0415)
This commit is contained in:
@@ -934,7 +934,7 @@
|
|||||||
"litellm_provider": "bedrock_converse",
|
"litellm_provider": "bedrock_converse",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.65e-05,
|
"output_cost_per_token": 1.65e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -4981,7 +4981,7 @@
|
|||||||
"litellm_provider": "anthropic",
|
"litellm_provider": "anthropic",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.5e-05,
|
"output_cost_per_token": 1.5e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -5011,7 +5011,7 @@
|
|||||||
"litellm_provider": "anthropic",
|
"litellm_provider": "anthropic",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.5e-05,
|
"output_cost_per_token": 1.5e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -8051,7 +8051,7 @@
|
|||||||
"litellm_provider": "bedrock_converse",
|
"litellm_provider": "bedrock_converse",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.65e-05,
|
"output_cost_per_token": 1.65e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -12130,7 +12130,7 @@
|
|||||||
"litellm_provider": "bedrock_converse",
|
"litellm_provider": "bedrock_converse",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.5e-05,
|
"output_cost_per_token": 1.5e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -14751,7 +14751,7 @@
|
|||||||
"litellm_provider": "bedrock_converse",
|
"litellm_provider": "bedrock_converse",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.65e-05,
|
"output_cost_per_token": 1.65e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -20630,7 +20630,7 @@
|
|||||||
"litellm_provider": "bedrock_converse",
|
"litellm_provider": "bedrock_converse",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.65e-05,
|
"output_cost_per_token": 1.65e-05,
|
||||||
"search_context_cost_per_query": {
|
"search_context_cost_per_query": {
|
||||||
@@ -22049,7 +22049,7 @@
|
|||||||
"litellm_provider": "vertex_ai-anthropic_models",
|
"litellm_provider": "vertex_ai-anthropic_models",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.5e-05,
|
"output_cost_per_token": 1.5e-05,
|
||||||
"output_cost_per_token_batches": 7.5e-06,
|
"output_cost_per_token_batches": 7.5e-06,
|
||||||
@@ -22075,7 +22075,7 @@
|
|||||||
"litellm_provider": "vertex_ai-anthropic_models",
|
"litellm_provider": "vertex_ai-anthropic_models",
|
||||||
"max_input_tokens": 200000,
|
"max_input_tokens": 200000,
|
||||||
"max_output_tokens": 64000,
|
"max_output_tokens": 64000,
|
||||||
"max_tokens": 200000,
|
"max_tokens": 64000,
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"output_cost_per_token": 1.5e-05,
|
"output_cost_per_token": 1.5e-05,
|
||||||
"output_cost_per_token_batches": 7.5e-06,
|
"output_cost_per_token_batches": 7.5e-06,
|
||||||
|
|||||||
@@ -1 +1 @@
|
|||||||
4a7b0978254bb5bcf882de6cf72c00bfb21e2be13d9c217700b2ac3ac129935e
|
f64b517b218c0a125b7d75fb268de7705f8a66d7ec861ba8b45b5bf9b72b0415
|
||||||
|
|||||||
Reference in New Issue
Block a user