forked from phoenix/litellm-mirror
docs - show correct rpm - > tpm conversion for Azure
This commit is contained in:
parent
70df474e64
commit
bbf4db79c1
2 changed files with 2 additions and 2 deletions
|
@ -167,7 +167,7 @@ Routes to **deployment with lowest TPM usage** for that minute.
|
||||||
|
|
||||||
In production, we use Redis to track usage (TPM/RPM) across multiple deployments. This implementation uses **async redis calls** (redis.incr and redis.mget).
|
In production, we use Redis to track usage (TPM/RPM) across multiple deployments. This implementation uses **async redis calls** (redis.incr and redis.mget).
|
||||||
|
|
||||||
For Azure, your RPM = TPM/6.
|
For Azure, [you get 6 RPM per 1000 TPM](https://stackoverflow.com/questions/77368844/what-is-the-request-per-minute-rate-limit-for-azure-openai-models-for-gpt-3-5-tu)
|
||||||
|
|
||||||
<Tabs>
|
<Tabs>
|
||||||
<TabItem value="sdk" label="sdk">
|
<TabItem value="sdk" label="sdk">
|
||||||
|
|
|
@ -228,12 +228,12 @@ const sidebars = {
|
||||||
},
|
},
|
||||||
],
|
],
|
||||||
},
|
},
|
||||||
|
"routing",
|
||||||
"scheduler",
|
"scheduler",
|
||||||
{
|
{
|
||||||
type: "category",
|
type: "category",
|
||||||
label: "🚅 LiteLLM Python SDK",
|
label: "🚅 LiteLLM Python SDK",
|
||||||
items: [
|
items: [
|
||||||
"routing",
|
|
||||||
"set_keys",
|
"set_keys",
|
||||||
"completion/token_usage",
|
"completion/token_usage",
|
||||||
"sdk_custom_pricing",
|
"sdk_custom_pricing",
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue