From c9d6925a42f9976ac1310f9a64ae461b9a6eaeed Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Sat, 28 Sep 2024 19:08:20 -0700 Subject: [PATCH] docs(reliability.md): add tutorial on setting wildcard models as fallbacks --- docs/my-website/docs/proxy/reliability.md | 50 +++++++++++++++++++++++ 1 file changed, 50 insertions(+) diff --git a/docs/my-website/docs/proxy/reliability.md b/docs/my-website/docs/proxy/reliability.md index 4aa03228d..51fa6d3ca 100644 --- a/docs/my-website/docs/proxy/reliability.md +++ b/docs/my-website/docs/proxy/reliability.md @@ -699,3 +699,53 @@ print(response) ``` + +### Setting Fallbacks for Wildcard Models + +You can set fallbacks for wildcard models (e.g. `azure/*`) in your config file. + +1. Setup config +```yaml +model_list: + - model_name: "gpt-4o" + litellm_params: + model: "openai/gpt-4o" + api_key: os.environ/OPENAI_API_KEY + - model_name: "azure/*" + litellm_params: + model: "azure/*" + api_key: os.environ/AZURE_API_KEY + api_base: os.environ/AZURE_API_BASE + +litellm_settings: + fallbacks: [{"gpt-4o": ["azure/gpt-4o"]}] +``` + +2. Start Proxy +```bash +litellm --config /path/to/config.yaml +``` + +3. Test it! + +```bash +curl -L -X POST 'http://0.0.0.0:4000/v1/chat/completions' \ +-H 'Content-Type: application/json' \ +-H 'Authorization: Bearer sk-1234' \ +-d '{ + "model": "gpt-4o", + "messages": [ + { + "role": "user", + "content": [ + { + "type": "text", + "text": "what color is red" + } + ] + } + ], + "max_tokens": 300, + "mock_testing_fallbacks": true +}' +``` \ No newline at end of file