mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
Revert "(feat) add typehints for litellm.acompletion"
This reverts commit a9cf6cec80
.
This commit is contained in:
parent
0124de558d
commit
b4cc227d1c
1 changed files with 5 additions and 67 deletions
|
@ -353,77 +353,15 @@ class Router:
|
||||||
)
|
)
|
||||||
raise e
|
raise e
|
||||||
|
|
||||||
async def acompletion(
|
async def acompletion(self, model: str, messages: List[Dict[str, str]], **kwargs):
|
||||||
self,
|
|
||||||
model: str,
|
|
||||||
# Optional OpenAI params: see https://platform.openai.com/docs/api-reference/chat/create
|
|
||||||
messages: List = [],
|
|
||||||
functions: Optional[List] = None,
|
|
||||||
function_call: Optional[str] = None,
|
|
||||||
timeout: Optional[Union[float, int]] = None,
|
|
||||||
temperature: Optional[float] = None,
|
|
||||||
top_p: Optional[float] = None,
|
|
||||||
n: Optional[int] = None,
|
|
||||||
stream: Optional[bool] = False,
|
|
||||||
stop=None,
|
|
||||||
max_tokens: Optional[float] = None,
|
|
||||||
presence_penalty: Optional[float] = None,
|
|
||||||
frequency_penalty: Optional[float] = None,
|
|
||||||
logit_bias: Optional[dict] = None,
|
|
||||||
user: Optional[str] = None,
|
|
||||||
# openai v1.0+ new params
|
|
||||||
response_format: Optional[dict] = None,
|
|
||||||
seed: Optional[int] = None,
|
|
||||||
tools: Optional[List] = None,
|
|
||||||
tool_choice: Optional[str] = None,
|
|
||||||
logprobs: Optional[bool] = None,
|
|
||||||
top_logprobs: Optional[int] = None,
|
|
||||||
deployment_id=None,
|
|
||||||
# set api_base, api_version, api_key
|
|
||||||
base_url: Optional[str] = None,
|
|
||||||
api_version: Optional[str] = None,
|
|
||||||
api_key: Optional[str] = None,
|
|
||||||
model_list: Optional[list] = None, # pass in a list of api_base,keys, etc.
|
|
||||||
# Optional liteLLM function params
|
|
||||||
**kwargs,
|
|
||||||
):
|
|
||||||
try:
|
try:
|
||||||
completion_kwargs = {
|
kwargs["model"] = model
|
||||||
"model": model,
|
kwargs["messages"] = messages
|
||||||
"messages": messages,
|
kwargs["original_function"] = self._acompletion
|
||||||
"functions": functions,
|
|
||||||
"function_call": function_call,
|
|
||||||
"timeout": timeout or self.timeout,
|
|
||||||
"temperature": temperature,
|
|
||||||
"top_p": top_p,
|
|
||||||
"n": n,
|
|
||||||
"stream": stream,
|
|
||||||
"stop": stop,
|
|
||||||
"max_tokens": max_tokens,
|
|
||||||
"presence_penalty": presence_penalty,
|
|
||||||
"frequency_penalty": frequency_penalty,
|
|
||||||
"logit_bias": logit_bias,
|
|
||||||
"user": user,
|
|
||||||
"response_format": response_format,
|
|
||||||
"seed": seed,
|
|
||||||
"tools": tools,
|
|
||||||
"tool_choice": tool_choice,
|
|
||||||
"logprobs": logprobs,
|
|
||||||
"top_logprobs": top_logprobs,
|
|
||||||
"deployment_id": deployment_id,
|
|
||||||
"base_url": base_url,
|
|
||||||
"api_version": api_version,
|
|
||||||
"api_key": api_key,
|
|
||||||
"model_list": model_list,
|
|
||||||
"original_function": self._acompletion,
|
|
||||||
}
|
|
||||||
kwargs["num_retries"] = kwargs.get("num_retries", self.num_retries)
|
kwargs["num_retries"] = kwargs.get("num_retries", self.num_retries)
|
||||||
kwargs.setdefault("metadata", {}).update({"model_group": model})
|
kwargs.setdefault("metadata", {}).update({"model_group": model})
|
||||||
|
|
||||||
response = await self.async_function_with_fallbacks(
|
response = await self.async_function_with_fallbacks(**kwargs)
|
||||||
**completion_kwargs,
|
|
||||||
**kwargs,
|
|
||||||
)
|
|
||||||
|
|
||||||
return response
|
return response
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue