forked from phoenix/litellm-mirror
fix(types/router.py): add 'drop_params' as a litellm param to router types
This commit is contained in:
parent
b360ab4c89
commit
bdffb1f23d
2 changed files with 4 additions and 7 deletions
|
@ -28,7 +28,8 @@ model_list:
|
|||
api_key: os.environ/AZURE_API_KEY
|
||||
api_version: '2023-05-15'
|
||||
model: azure/chatgpt-v-2
|
||||
model_name: gpt-3.5-turbo
|
||||
drop_params: True
|
||||
model_name: gpt-3.5-turbo-drop-params
|
||||
- model_name: tts
|
||||
litellm_params:
|
||||
model: openai/tts-1
|
||||
|
|
|
@ -283,12 +283,6 @@ class updateDeployment(BaseModel):
|
|||
|
||||
|
||||
class LiteLLMParamsTypedDict(TypedDict, total=False):
|
||||
"""
|
||||
[TODO]
|
||||
- allow additional params (not in list)
|
||||
- set value to none if not set -> don't raise error if value not set
|
||||
"""
|
||||
|
||||
model: str
|
||||
custom_llm_provider: Optional[str]
|
||||
tpm: Optional[int]
|
||||
|
@ -300,6 +294,8 @@ class LiteLLMParamsTypedDict(TypedDict, total=False):
|
|||
stream_timeout: Optional[Union[float, str]]
|
||||
max_retries: Optional[int]
|
||||
organization: Optional[str] # for openai orgs
|
||||
## DROP PARAMS ##
|
||||
drop_params: Optional[bool]
|
||||
## UNIFIED PROJECT/REGION ##
|
||||
region_name: Optional[str]
|
||||
## VERTEX AI ##
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue