From 46fc8be4c82e2811b0c09cc61efb2cf64f6d0a52 Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Thu, 23 Nov 2023 13:58:59 -0800 Subject: [PATCH] (feat) cost tracking ft:gpt-3.5-turbo --- litellm/utils.py | 9 +++++++++ model_prices_and_context_window.json | 7 +++++++ 2 files changed, 16 insertions(+) diff --git a/litellm/utils.py b/litellm/utils.py index 8ca70fa64a..dce870ce6f 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -1611,6 +1611,15 @@ def cost_per_token(model="gpt-3.5-turbo", prompt_tokens=0, completion_tokens=0): model_cost_ref[model]["output_cost_per_token"] * completion_tokens ) return prompt_tokens_cost_usd_dollar, completion_tokens_cost_usd_dollar + elif "ft:gpt-3.5-turbo" in model: + # fuzzy match ft:gpt-3.5-turbo:abcd-id-cool-litellm + prompt_tokens_cost_usd_dollar = ( + model_cost_ref["ft:gpt-3.5-turbo"]["input_cost_per_token"] * prompt_tokens + ) + completion_tokens_cost_usd_dollar = ( + model_cost_ref["ft:gpt-3.5-turbo"]["output_cost_per_token"] * completion_tokens + ) + return prompt_tokens_cost_usd_dollar, completion_tokens_cost_usd_dollar else: # calculate average input cost, azure/gpt-deployments can potentially go here if users don't specify, gpt-4, gpt-3.5-turbo. LLMs litellm knows input_cost_sum = 0 diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 4e98b3ed06..ff67038836 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -97,6 +97,13 @@ "litellm_provider": "openai", "mode": "chat" }, + "ft:gpt-3.5-turbo": { + "max_tokens": 4097, + "input_cost_per_token": 0.000012, + "output_cost_per_token": 0.000016, + "litellm_provider": "openai", + "mode": "chat" + }, "azure/gpt-4-32k": { "max_tokens": 8192, "input_cost_per_token": 0.00006,