diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index daf2c502a..42863e8bc 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -286,8 +286,35 @@ "mode": "chat" }, "ft:gpt-3.5-turbo": { - "max_tokens": 4097, - "max_input_tokens": 4097, + "max_tokens": 4096, + "max_input_tokens": 16385, + "max_output_tokens": 4096, + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.000006, + "litellm_provider": "openai", + "mode": "chat" + }, + "ft:gpt-3.5-turbo-0125": { + "max_tokens": 4096, + "max_input_tokens": 16385, + "max_output_tokens": 4096, + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.000006, + "litellm_provider": "openai", + "mode": "chat" + }, + "ft:gpt-3.5-turbo-1106": { + "max_tokens": 4096, + "max_input_tokens": 16385, + "max_output_tokens": 4096, + "input_cost_per_token": 0.000003, + "output_cost_per_token": 0.000006, + "litellm_provider": "openai", + "mode": "chat" + }, + "ft:gpt-3.5-turbo-0613": { + "max_tokens": 4096, + "max_input_tokens": 4096, "max_output_tokens": 4096, "input_cost_per_token": 0.000003, "output_cost_per_token": 0.000006, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index e94ff25b8..42863e8bc 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -286,7 +286,7 @@ "mode": "chat" }, "ft:gpt-3.5-turbo": { - "max_tokens": 16385, + "max_tokens": 4096, "max_input_tokens": 16385, "max_output_tokens": 4096, "input_cost_per_token": 0.000003, @@ -295,7 +295,7 @@ "mode": "chat" }, "ft:gpt-3.5-turbo-0125": { - "max_tokens": 16385, + "max_tokens": 4096, "max_input_tokens": 16385, "max_output_tokens": 4096, "input_cost_per_token": 0.000003, @@ -304,7 +304,7 @@ "mode": "chat" }, "ft:gpt-3.5-turbo-1106": { - "max_tokens": 16385, + "max_tokens": 4096, "max_input_tokens": 16385, "max_output_tokens": 4096, "input_cost_per_token": 0.000003, @@ -313,8 +313,8 @@ "mode": "chat" }, "ft:gpt-3.5-turbo-0613": { - "max_tokens": 4097, - "max_input_tokens": 4097, + "max_tokens": 4096, + "max_input_tokens": 4096, "max_output_tokens": 4096, "input_cost_per_token": 0.000003, "output_cost_per_token": 0.000006,