(feat) add mode for config.yaml health checks

This commit is contained in:
ishaan-jaff 2023-12-06 11:16:28 -08:00
parent 4f02b3c161
commit 7c77cc3cfa
3 changed files with 6 additions and 8 deletions

View file

@ -50,7 +50,6 @@ async def _perform_health_check(model_list: list):
try: try:
await litellm.aembedding(**model_params) await litellm.aembedding(**model_params)
except Exception as e: except Exception as e:
print_verbose(f"\n\n Got Exception, {e}")
print_verbose(f"Health check failed for model {model_params['model']}. Error: {e}") print_verbose(f"Health check failed for model {model_params['model']}. Error: {e}")
return False return False
return True return True
@ -60,11 +59,6 @@ async def _perform_health_check(model_list: list):
try: try:
await litellm.acompletion(**model_params) await litellm.acompletion(**model_params)
except Exception as e: except Exception as e:
print_verbose(f"\n\n Got Exception, {e}")
error_str = (str(e))
if "This is not a chat model" in error_str or "The chatCompletion operation does not work with the specified model" in error_str:
return await _check_embedding_model(model_params)
print_verbose(f"Health check failed for model {model_params['model']}. Error: {e}") print_verbose(f"Health check failed for model {model_params['model']}. Error: {e}")
return False return False
@ -78,7 +72,8 @@ async def _perform_health_check(model_list: list):
litellm_params["messages"] = _get_random_llm_message() litellm_params["messages"] = _get_random_llm_message()
prepped_params.append(litellm_params) prepped_params.append(litellm_params)
if "embedding" in litellm_params["model"]: if model.get("mode", None) == "embedding":
# this is an embedding model
tasks.append(_check_embedding_model(litellm_params)) tasks.append(_check_embedding_model(litellm_params))
else: else:
tasks.append(_check_model(litellm_params)) tasks.append(_check_model(litellm_params))

View file

@ -11,6 +11,7 @@ model_list:
api_base: os.environ/AZURE_API_BASE api_base: os.environ/AZURE_API_BASE
api_key: os.environ/AZURE_API_KEY api_key: os.environ/AZURE_API_KEY
api_version: "2023-07-01-preview" api_version: "2023-07-01-preview"
mode: embedding
- model_name: openai-gpt-3.5 - model_name: openai-gpt-3.5
litellm_params: litellm_params:
model: gpt-3.5-turbo model: gpt-3.5-turbo
@ -19,6 +20,7 @@ model_list:
litellm_params: litellm_params:
model: text-embedding-ada-002 model: text-embedding-ada-002
api_key: os.environ/OPENAI_API_KEY api_key: os.environ/OPENAI_API_KEY
mode: embedding
litellm_settings: litellm_settings:

View file

@ -40,6 +40,7 @@ class ProxyChatCompletionRequest(BaseModel):
class ModelInfo(BaseModel): class ModelInfo(BaseModel):
id: Optional[str] id: Optional[str]
mode: Optional[str]
class Config: class Config:
extra = Extra.allow # Allow extra fields extra = Extra.allow # Allow extra fields