forked from phoenix/litellm-mirror
Merge pull request #4420 from guitard0g/add-init-return-annos
Add return type annotations to util types
This commit is contained in:
commit
9ca2193bdf
1 changed files with 13 additions and 13 deletions
|
@ -171,7 +171,7 @@ class Function(OpenAIObject):
|
||||||
arguments: Union[Dict, str],
|
arguments: Union[Dict, str],
|
||||||
name: Optional[str] = None,
|
name: Optional[str] = None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
if isinstance(arguments, Dict):
|
if isinstance(arguments, Dict):
|
||||||
arguments = json.dumps(arguments)
|
arguments = json.dumps(arguments)
|
||||||
else:
|
else:
|
||||||
|
@ -242,7 +242,7 @@ class ChatCompletionMessageToolCall(OpenAIObject):
|
||||||
id: Optional[str] = None,
|
id: Optional[str] = None,
|
||||||
type: Optional[str] = None,
|
type: Optional[str] = None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
super(ChatCompletionMessageToolCall, self).__init__(**params)
|
super(ChatCompletionMessageToolCall, self).__init__(**params)
|
||||||
if isinstance(function, Dict):
|
if isinstance(function, Dict):
|
||||||
self.function = Function(**function)
|
self.function = Function(**function)
|
||||||
|
@ -285,7 +285,7 @@ class Message(OpenAIObject):
|
||||||
function_call=None,
|
function_call=None,
|
||||||
tool_calls=None,
|
tool_calls=None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
super(Message, self).__init__(**params)
|
super(Message, self).__init__(**params)
|
||||||
self.content = content
|
self.content = content
|
||||||
self.role = role
|
self.role = role
|
||||||
|
@ -328,7 +328,7 @@ class Delta(OpenAIObject):
|
||||||
function_call=None,
|
function_call=None,
|
||||||
tool_calls=None,
|
tool_calls=None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
super(Delta, self).__init__(**params)
|
super(Delta, self).__init__(**params)
|
||||||
self.content = content
|
self.content = content
|
||||||
self.role = role
|
self.role = role
|
||||||
|
@ -375,7 +375,7 @@ class Choices(OpenAIObject):
|
||||||
logprobs=None,
|
logprobs=None,
|
||||||
enhancements=None,
|
enhancements=None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
super(Choices, self).__init__(**params)
|
super(Choices, self).__init__(**params)
|
||||||
if finish_reason is not None:
|
if finish_reason is not None:
|
||||||
self.finish_reason = map_finish_reason(
|
self.finish_reason = map_finish_reason(
|
||||||
|
@ -416,7 +416,7 @@ class Choices(OpenAIObject):
|
||||||
class Usage(OpenAIObject):
|
class Usage(OpenAIObject):
|
||||||
def __init__(
|
def __init__(
|
||||||
self, prompt_tokens=None, completion_tokens=None, total_tokens=None, **params
|
self, prompt_tokens=None, completion_tokens=None, total_tokens=None, **params
|
||||||
):
|
) -> None:
|
||||||
super(Usage, self).__init__(**params)
|
super(Usage, self).__init__(**params)
|
||||||
if prompt_tokens:
|
if prompt_tokens:
|
||||||
self.prompt_tokens = prompt_tokens
|
self.prompt_tokens = prompt_tokens
|
||||||
|
@ -451,7 +451,7 @@ class StreamingChoices(OpenAIObject):
|
||||||
logprobs=None,
|
logprobs=None,
|
||||||
enhancements=None,
|
enhancements=None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
super(StreamingChoices, self).__init__(**params)
|
super(StreamingChoices, self).__init__(**params)
|
||||||
if finish_reason:
|
if finish_reason:
|
||||||
self.finish_reason = finish_reason
|
self.finish_reason = finish_reason
|
||||||
|
@ -657,7 +657,7 @@ class EmbeddingResponse(OpenAIObject):
|
||||||
response_ms=None,
|
response_ms=None,
|
||||||
data=None,
|
data=None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
object = "list"
|
object = "list"
|
||||||
if response_ms:
|
if response_ms:
|
||||||
_response_ms = response_ms
|
_response_ms = response_ms
|
||||||
|
@ -708,7 +708,7 @@ class Logprobs(OpenAIObject):
|
||||||
|
|
||||||
|
|
||||||
class TextChoices(OpenAIObject):
|
class TextChoices(OpenAIObject):
|
||||||
def __init__(self, finish_reason=None, index=0, text=None, logprobs=None, **params):
|
def __init__(self, finish_reason=None, index=0, text=None, logprobs=None, **params) -> None:
|
||||||
super(TextChoices, self).__init__(**params)
|
super(TextChoices, self).__init__(**params)
|
||||||
if finish_reason:
|
if finish_reason:
|
||||||
self.finish_reason = map_finish_reason(finish_reason)
|
self.finish_reason = map_finish_reason(finish_reason)
|
||||||
|
@ -790,7 +790,7 @@ class TextCompletionResponse(OpenAIObject):
|
||||||
response_ms=None,
|
response_ms=None,
|
||||||
object=None,
|
object=None,
|
||||||
**params,
|
**params,
|
||||||
):
|
) -> None:
|
||||||
if stream:
|
if stream:
|
||||||
object = "text_completion.chunk"
|
object = "text_completion.chunk"
|
||||||
choices = [TextChoices()]
|
choices = [TextChoices()]
|
||||||
|
@ -873,7 +873,7 @@ class ImageObject(OpenAIObject):
|
||||||
url: Optional[str] = None
|
url: Optional[str] = None
|
||||||
revised_prompt: Optional[str] = None
|
revised_prompt: Optional[str] = None
|
||||||
|
|
||||||
def __init__(self, b64_json=None, url=None, revised_prompt=None):
|
def __init__(self, b64_json=None, url=None, revised_prompt=None) -> None:
|
||||||
super().__init__(b64_json=b64_json, url=url, revised_prompt=revised_prompt)
|
super().__init__(b64_json=b64_json, url=url, revised_prompt=revised_prompt)
|
||||||
|
|
||||||
def __contains__(self, key):
|
def __contains__(self, key):
|
||||||
|
@ -909,7 +909,7 @@ class ImageResponse(OpenAIObject):
|
||||||
|
|
||||||
_hidden_params: dict = {}
|
_hidden_params: dict = {}
|
||||||
|
|
||||||
def __init__(self, created=None, data=None, response_ms=None):
|
def __init__(self, created=None, data=None, response_ms=None) -> None:
|
||||||
if response_ms:
|
if response_ms:
|
||||||
_response_ms = response_ms
|
_response_ms = response_ms
|
||||||
else:
|
else:
|
||||||
|
@ -956,7 +956,7 @@ class TranscriptionResponse(OpenAIObject):
|
||||||
|
|
||||||
_hidden_params: dict = {}
|
_hidden_params: dict = {}
|
||||||
|
|
||||||
def __init__(self, text=None):
|
def __init__(self, text=None) -> None:
|
||||||
super().__init__(text=text)
|
super().__init__(text=text)
|
||||||
|
|
||||||
def __contains__(self, key):
|
def __contains__(self, key):
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue