forked from phoenix/litellm-mirror
(docs) add example load balancer
This commit is contained in:
parent
7376e57e9c
commit
42432bedaa
1 changed files with 28 additions and 0 deletions
28
litellm/proxy/example_config_yaml/load_balancer.yaml
Normal file
28
litellm/proxy/example_config_yaml/load_balancer.yaml
Normal file
|
@ -0,0 +1,28 @@
|
||||||
|
litellm_settings:
|
||||||
|
drop_params: True
|
||||||
|
|
||||||
|
# Model-specific settings
|
||||||
|
model_list: # use the same model_name for using the litellm router. LiteLLM will use the router between gpt-3.5-turbo
|
||||||
|
- model_name: gpt-3.5-turbo # litellm will
|
||||||
|
litellm_params:
|
||||||
|
model: gpt-3.5-turbo
|
||||||
|
api_key: sk-uj6F
|
||||||
|
tpm: 20000 # [OPTIONAL] REPLACE with your openai tpm
|
||||||
|
rpm: 3 # [OPTIONAL] REPLACE with your openai rpm
|
||||||
|
- model_name: gpt-3.5-turbo
|
||||||
|
litellm_params:
|
||||||
|
model: gpt-3.5-turbo
|
||||||
|
api_key: sk-Imn
|
||||||
|
tpm: 20000 # [OPTIONAL] REPLACE with your openai tpm
|
||||||
|
rpm: 3 # [OPTIONAL] REPLACE with your openai rpm
|
||||||
|
- model_name: gpt-3.5-turbo
|
||||||
|
litellm_params:
|
||||||
|
model: openrouter/gpt-3.5-turbo
|
||||||
|
- model_name: mistral-7b-instruct
|
||||||
|
litellm_params:
|
||||||
|
model: mistralai/mistral-7b-instruct
|
||||||
|
|
||||||
|
environment_variables:
|
||||||
|
REDIS_HOST: localhost
|
||||||
|
REDIS_PASSWORD: "eYVX71"
|
||||||
|
REDIS_PORT: 6329
|
Loading…
Add table
Add a link
Reference in a new issue