mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
docs(config.md): update wildcard docs
This commit is contained in:
parent
9d87767639
commit
e39ff46222
3 changed files with 5 additions and 5 deletions
|
@ -295,7 +295,7 @@ Dynamically call any model from any given provider without the need to predefine
|
|||
model_list:
|
||||
- model_name: "*" # all requests where model not in your config go to this deployment
|
||||
litellm_params:
|
||||
model: "openai/*" # passes our validation check that a real provider is given
|
||||
model: "*" # passes our validation check that a real provider is given
|
||||
```
|
||||
|
||||
2. Start LiteLLM proxy
|
||||
|
|
|
@ -1,4 +1,4 @@
|
|||
model_list:
|
||||
- model_name: "gpt-3.5-turbo"
|
||||
- model_name: "*"
|
||||
litellm_params:
|
||||
model: "openai/gpt-3.5-turbo"
|
||||
model: "*"
|
||||
|
|
|
@ -2937,8 +2937,8 @@ class Router:
|
|||
model_group = kwargs["litellm_params"]["metadata"].get(
|
||||
"model_group", None
|
||||
)
|
||||
|
||||
id = kwargs["litellm_params"].get("model_info", {}).get("id", None)
|
||||
model_info = kwargs["litellm_params"].get("model_info", {}) or {}
|
||||
id = model_info.get("id", None)
|
||||
if model_group is None or id is None:
|
||||
return
|
||||
elif isinstance(id, int):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue