diff --git a/docs/my-website/docs/providers/azure.md b/docs/my-website/docs/providers/azure.md index 4b78f12be..dda7384c0 100644 --- a/docs/my-website/docs/providers/azure.md +++ b/docs/my-website/docs/providers/azure.md @@ -168,6 +168,13 @@ response = completion( ) ``` +## Azure Instruct Models + +| Model Name | Function Call | +|---------------------|----------------------------------------------------| +| gpt-3.5-turbo-instruct | `response = completion(model="azure/", messages=messages)` | +| gpt-3.5-turbo-instruct-0914 | `response = completion(model="azure/", messages=messages)` | + ## Advanced ### Azure API Load-Balancing diff --git a/docs/my-website/docs/providers/openai.md b/docs/my-website/docs/providers/openai.md index 1ca13b600..aa42847b1 100644 --- a/docs/my-website/docs/providers/openai.md +++ b/docs/my-website/docs/providers/openai.md @@ -93,6 +93,7 @@ response = completion( | Model Name | Function Call | |---------------------|----------------------------------------------------| | gpt-3.5-turbo-instruct | `response = completion(model="gpt-3.5-turbo-instruct", messages=messages)` | +| gpt-3.5-turbo-instruct-0914 | `response = completion(model="gpt-3.5-turbo-instruct-091", messages=messages)` | | text-davinci-003 | `response = completion(model="text-davinci-003", messages=messages)` | | ada-001 | `response = completion(model="ada-001", messages=messages)` | | curie-001 | `response = completion(model="curie-001", messages=messages)` | diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 66061acc4..d97c042c5 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -424,6 +424,23 @@ "mode": "chat", "supports_function_calling": true }, + "azure/gpt-3.5-turbo-instruct-0914": { + "max_tokens": 4097, + + "input_cost_per_token": 0.0000015, + "output_cost_per_token": 0.000002, + "litellm_provider": "text-completion-openai", + "mode": "completion" + + }, + "azure/gpt-35-turbo-instruct": { + "max_tokens": 4097, + "input_cost_per_token": 0.0000015, + "output_cost_per_token": 0.000002, + "litellm_provider": "text-completion-openai", + "mode": "completion" + + }, "azure/mistral-large-latest": { "max_tokens": 32000, "input_cost_per_token": 0.000008, @@ -537,6 +554,14 @@ "litellm_provider": "text-completion-openai", "mode": "completion" }, + "gpt-3.5-turbo-instruct-0914": { + "max_tokens": 4097, + "input_cost_per_token": 0.0000015, + "output_cost_per_token": 0.000002, + "litellm_provider": "text-completion-openai", + "mode": "completion" + + }, "claude-instant-1": { "max_tokens": 100000, "max_output_tokens": 8191, diff --git a/model_prices_and_context_window.json b/model_prices_and_context_window.json index 66061acc4..d97c042c5 100644 --- a/model_prices_and_context_window.json +++ b/model_prices_and_context_window.json @@ -424,6 +424,23 @@ "mode": "chat", "supports_function_calling": true }, + "azure/gpt-3.5-turbo-instruct-0914": { + "max_tokens": 4097, + + "input_cost_per_token": 0.0000015, + "output_cost_per_token": 0.000002, + "litellm_provider": "text-completion-openai", + "mode": "completion" + + }, + "azure/gpt-35-turbo-instruct": { + "max_tokens": 4097, + "input_cost_per_token": 0.0000015, + "output_cost_per_token": 0.000002, + "litellm_provider": "text-completion-openai", + "mode": "completion" + + }, "azure/mistral-large-latest": { "max_tokens": 32000, "input_cost_per_token": 0.000008, @@ -537,6 +554,14 @@ "litellm_provider": "text-completion-openai", "mode": "completion" }, + "gpt-3.5-turbo-instruct-0914": { + "max_tokens": 4097, + "input_cost_per_token": 0.0000015, + "output_cost_per_token": 0.000002, + "litellm_provider": "text-completion-openai", + "mode": "completion" + + }, "claude-instant-1": { "max_tokens": 100000, "max_output_tokens": 8191,