diff --git a/docs/my-website/docs/proxy/configs.md b/docs/my-website/docs/proxy/configs.md index e9fd4cda4..0a155828b 100644 --- a/docs/my-website/docs/proxy/configs.md +++ b/docs/my-website/docs/proxy/configs.md @@ -227,6 +227,7 @@ model_list: - model_name: openai-gpt-3.5 litellm_params: model: openai/gpt-3.5-turbo + extra_headers: {"AI-Resource Group": "ishaan-resource"} api_key: sk-123 organization: org-ikDc4ex8NB temperature: 0.2 @@ -234,10 +235,6 @@ model_list: litellm_params: model: ollama/mistral api_base: your_ollama_api_base - headers: { - "HTTP-Referer": "litellm.ai", - "X-Title": "LiteLLM Server" - } ``` **Step 2**: Start server with config diff --git a/docs/my-website/docs/simple_proxy_old_doc.md b/docs/my-website/docs/simple_proxy_old_doc.md index 01c8a5754..b48e345e1 100644 --- a/docs/my-website/docs/simple_proxy_old_doc.md +++ b/docs/my-website/docs/simple_proxy_old_doc.md @@ -787,10 +787,6 @@ model_list: litellm_params: model: ollama/mistral api_base: your_ollama_api_base - headers: { - "HTTP-Referer": "litellm.ai", - "X-Title": "LiteLLM Server" - } ``` **Step 2**: Start server with config diff --git a/litellm/main.py b/litellm/main.py index 120f85cb1..acd2ba513 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -436,6 +436,7 @@ def completion( api_version (str, optional): API version (default is None). api_key (str, optional): API key (default is None). model_list (list, optional): List of api base, version, keys + extra_headers (dict, optional): Additional headers to include in the request. LITELLM Specific Params mock_response (str, optional): If provided, return a mock completion response for testing or debugging purposes (default is None).