Merge pull request #4087 from BerriAI/litellm_update_together_ai_pricing

Update together ai pricing
This commit is contained in:
Krish Dholakia 2024-06-08 22:07:12 -07:00 committed by GitHub
commit 90ae7f040a
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
4 changed files with 41 additions and 30 deletions

View file

@ -42,16 +42,18 @@ def get_model_params_and_category(model_name) -> str:
else: else:
return model_name return model_name
# Determine the category based on the number of parameters # Determine the category based on the number of parameters
if params_billion <= 3.0: if params_billion <= 4.0:
category = "together-ai-up-to-3b" category = "together-ai-up-to-4b"
elif params_billion <= 7.0: elif params_billion <= 8.0:
category = "together-ai-3.1b-7b" category = "together-ai-4.1b-8b"
elif params_billion <= 20.0: elif params_billion <= 21.0:
category = "together-ai-7.1b-20b" category = "together-ai-8.1b-21b"
elif params_billion <= 40.0: elif params_billion <= 41.0:
category = "together-ai-20.1b-40b" category = "together-ai-21.1b-41b"
elif params_billion <= 70.0: elif params_billion <= 80.0:
category = "together-ai-40.1b-70b" category = "together-ai-41.1b-80b"
elif params_billion <= 110.0:
category = "together-ai-81.1b-110b"
if category is not None: if category is not None:
return category return category

View file

@ -3009,32 +3009,37 @@
"litellm_provider": "sagemaker", "litellm_provider": "sagemaker",
"mode": "chat" "mode": "chat"
}, },
"together-ai-up-to-3b": { "together-ai-up-to-4b": {
"input_cost_per_token": 0.0000001, "input_cost_per_token": 0.0000001,
"output_cost_per_token": 0.0000001, "output_cost_per_token": 0.0000001,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-3.1b-7b": { "together-ai-4.1b-8b": {
"input_cost_per_token": 0.0000002, "input_cost_per_token": 0.0000002,
"output_cost_per_token": 0.0000002, "output_cost_per_token": 0.0000002,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-7.1b-20b": { "together-ai-8.1b-21b": {
"max_tokens": 1000, "max_tokens": 1000,
"input_cost_per_token": 0.0000004, "input_cost_per_token": 0.0000003,
"output_cost_per_token": 0.0000004, "output_cost_per_token": 0.0000003,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-20.1b-40b": { "together-ai-21.1b-41b": {
"input_cost_per_token": 0.0000008, "input_cost_per_token": 0.0000008,
"output_cost_per_token": 0.0000008, "output_cost_per_token": 0.0000008,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-40.1b-70b": { "together-ai-41.1b-80b": {
"input_cost_per_token": 0.0000009, "input_cost_per_token": 0.0000009,
"output_cost_per_token": 0.0000009, "output_cost_per_token": 0.0000009,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-81.1b-110b": {
"input_cost_per_token": 0.0000018,
"output_cost_per_token": 0.0000018,
"litellm_provider": "together_ai"
},
"together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1": { "together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1": {
"input_cost_per_token": 0.0000006, "input_cost_per_token": 0.0000006,
"output_cost_per_token": 0.0000006, "output_cost_per_token": 0.0000006,

View file

@ -560,9 +560,8 @@ def test_together_ai_qwen_completion_cost():
"custom_cost_per_second": None, "custom_cost_per_second": None,
} }
try: response = litellm.cost_calculator.get_model_params_and_category(
litellm.completion_cost(**input_kwargs) model_name="qwen/Qwen2-72B-Instruct"
except litellm.NotFoundError: )
pass
except Exception: assert response == "together-ai-41.1b-80b"
pytest.fail("This should have returned a 'not found error'")

View file

@ -3009,32 +3009,37 @@
"litellm_provider": "sagemaker", "litellm_provider": "sagemaker",
"mode": "chat" "mode": "chat"
}, },
"together-ai-up-to-3b": { "together-ai-up-to-4b": {
"input_cost_per_token": 0.0000001, "input_cost_per_token": 0.0000001,
"output_cost_per_token": 0.0000001, "output_cost_per_token": 0.0000001,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-3.1b-7b": { "together-ai-4.1b-8b": {
"input_cost_per_token": 0.0000002, "input_cost_per_token": 0.0000002,
"output_cost_per_token": 0.0000002, "output_cost_per_token": 0.0000002,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-7.1b-20b": { "together-ai-8.1b-21b": {
"max_tokens": 1000, "max_tokens": 1000,
"input_cost_per_token": 0.0000004, "input_cost_per_token": 0.0000003,
"output_cost_per_token": 0.0000004, "output_cost_per_token": 0.0000003,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-20.1b-40b": { "together-ai-21.1b-41b": {
"input_cost_per_token": 0.0000008, "input_cost_per_token": 0.0000008,
"output_cost_per_token": 0.0000008, "output_cost_per_token": 0.0000008,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-40.1b-70b": { "together-ai-41.1b-80b": {
"input_cost_per_token": 0.0000009, "input_cost_per_token": 0.0000009,
"output_cost_per_token": 0.0000009, "output_cost_per_token": 0.0000009,
"litellm_provider": "together_ai" "litellm_provider": "together_ai"
}, },
"together-ai-81.1b-110b": {
"input_cost_per_token": 0.0000018,
"output_cost_per_token": 0.0000018,
"litellm_provider": "together_ai"
},
"together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1": { "together_ai/mistralai/Mixtral-8x7B-Instruct-v0.1": {
"input_cost_per_token": 0.0000006, "input_cost_per_token": 0.0000006,
"output_cost_per_token": 0.0000006, "output_cost_per_token": 0.0000006,