mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 19:24:27 +00:00
docs(config.md): update wildcard docs
This commit is contained in:
parent
9d87767639
commit
e39ff46222
3 changed files with 5 additions and 5 deletions
|
@ -295,7 +295,7 @@ Dynamically call any model from any given provider without the need to predefine
|
||||||
model_list:
|
model_list:
|
||||||
- model_name: "*" # all requests where model not in your config go to this deployment
|
- model_name: "*" # all requests where model not in your config go to this deployment
|
||||||
litellm_params:
|
litellm_params:
|
||||||
model: "openai/*" # passes our validation check that a real provider is given
|
model: "*" # passes our validation check that a real provider is given
|
||||||
```
|
```
|
||||||
|
|
||||||
2. Start LiteLLM proxy
|
2. Start LiteLLM proxy
|
||||||
|
|
|
@ -1,4 +1,4 @@
|
||||||
model_list:
|
model_list:
|
||||||
- model_name: "gpt-3.5-turbo"
|
- model_name: "*"
|
||||||
litellm_params:
|
litellm_params:
|
||||||
model: "openai/gpt-3.5-turbo"
|
model: "*"
|
||||||
|
|
|
@ -2937,8 +2937,8 @@ class Router:
|
||||||
model_group = kwargs["litellm_params"]["metadata"].get(
|
model_group = kwargs["litellm_params"]["metadata"].get(
|
||||||
"model_group", None
|
"model_group", None
|
||||||
)
|
)
|
||||||
|
model_info = kwargs["litellm_params"].get("model_info", {}) or {}
|
||||||
id = kwargs["litellm_params"].get("model_info", {}).get("id", None)
|
id = model_info.get("id", None)
|
||||||
if model_group is None or id is None:
|
if model_group is None or id is None:
|
||||||
return
|
return
|
||||||
elif isinstance(id, int):
|
elif isinstance(id, int):
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue