diff --git a/litellm/llms/openai.py b/litellm/llms/openai.py index 2f6c7e673..7121d7bc7 100644 --- a/litellm/llms/openai.py +++ b/litellm/llms/openai.py @@ -759,8 +759,11 @@ class OpenAIChatCompletion(BaseLLM): messages: Optional[list] = None, input: Optional[list] = None, prompt: Optional[str] = None, + organization: Optional[str] = None, ): - client = AsyncOpenAI(api_key=api_key, timeout=timeout) + client = AsyncOpenAI( + api_key=api_key, timeout=timeout, organization=organization + ) if model is None and mode != "image_generation": raise Exception("model is not set") diff --git a/litellm/main.py b/litellm/main.py index 0d6b0caec..851a89cb3 100644 --- a/litellm/main.py +++ b/litellm/main.py @@ -3227,6 +3227,7 @@ async def ahealth_check( or custom_llm_provider == "text-completion-openai" ): api_key = model_params.get("api_key") or get_secret("OPENAI_API_KEY") + organization = model_params.get("organization") timeout = ( model_params.get("timeout") @@ -3244,6 +3245,7 @@ async def ahealth_check( mode=mode, prompt=prompt, input=input, + organization=organization, ) else: if mode == "embedding": diff --git a/litellm/router.py b/litellm/router.py index 87755be9a..0d6d108e4 100644 --- a/litellm/router.py +++ b/litellm/router.py @@ -1411,7 +1411,7 @@ class Router: max_retries = litellm.get_secret(max_retries_env_name) litellm_params["max_retries"] = max_retries - organization = litellm_params.pop("organization", None) + organization = litellm_params.get("organization", None) if isinstance(organization, str) and organization.startswith("os.environ/"): organization_env_name = organization.replace("os.environ/", "") organization = litellm.get_secret(organization_env_name)