forked from phoenix/litellm-mirror
fix import error
This commit is contained in:
parent
ed627bc5d2
commit
25887c1846
9 changed files with 40 additions and 32 deletions
|
@ -923,7 +923,7 @@ from .llms.bedrock.embed.amazon_titan_v2_transformation import (
|
||||||
AmazonTitanV2Config,
|
AmazonTitanV2Config,
|
||||||
)
|
)
|
||||||
from .llms.bedrock.embed.cohere_transformation import BedrockCohereEmbeddingConfig
|
from .llms.bedrock.embed.cohere_transformation import BedrockCohereEmbeddingConfig
|
||||||
from .llms.openai import (
|
from .llms.OpenAI.openai import (
|
||||||
OpenAIConfig,
|
OpenAIConfig,
|
||||||
OpenAITextCompletionConfig,
|
OpenAITextCompletionConfig,
|
||||||
MistralConfig,
|
MistralConfig,
|
||||||
|
@ -938,7 +938,7 @@ from .llms.AI21.chat import AI21ChatConfig
|
||||||
from .llms.fireworks_ai import FireworksAIConfig
|
from .llms.fireworks_ai import FireworksAIConfig
|
||||||
from .llms.volcengine import VolcEngineConfig
|
from .llms.volcengine import VolcEngineConfig
|
||||||
from .llms.text_completion_codestral import MistralTextCompletionConfig
|
from .llms.text_completion_codestral import MistralTextCompletionConfig
|
||||||
from .llms.azure import (
|
from .llms.AzureOpenAI.azure import (
|
||||||
AzureOpenAIConfig,
|
AzureOpenAIConfig,
|
||||||
AzureOpenAIError,
|
AzureOpenAIError,
|
||||||
AzureOpenAIAssistantsAPIConfig,
|
AzureOpenAIAssistantsAPIConfig,
|
||||||
|
|
|
@ -16,7 +16,7 @@ import httpx
|
||||||
import litellm
|
import litellm
|
||||||
from litellm import client, get_secret
|
from litellm import client, get_secret
|
||||||
from litellm.llms.files_apis.azure import AzureOpenAIFilesAPI
|
from litellm.llms.files_apis.azure import AzureOpenAIFilesAPI
|
||||||
from litellm.llms.openai import FileDeleted, FileObject, OpenAIFilesAPI
|
from litellm.llms.OpenAI.openai import FileDeleted, FileObject, OpenAIFilesAPI
|
||||||
from litellm.types.llms.openai import (
|
from litellm.types.llms.openai import (
|
||||||
Batch,
|
Batch,
|
||||||
CreateFileRequest,
|
CreateFileRequest,
|
||||||
|
|
23
litellm/litellm_core_utils/audio_utils/utils.py
Normal file
23
litellm/litellm_core_utils/audio_utils/utils.py
Normal file
|
@ -0,0 +1,23 @@
|
||||||
|
"""
|
||||||
|
Utils used for litellm.transcription() and litellm.atranscription()
|
||||||
|
"""
|
||||||
|
|
||||||
|
from litellm.types.utils import FileTypes
|
||||||
|
|
||||||
|
|
||||||
|
def get_audio_file_name(file_obj: FileTypes) -> str:
|
||||||
|
"""
|
||||||
|
Safely get the name of a file-like object or return its string representation.
|
||||||
|
|
||||||
|
Args:
|
||||||
|
file_obj (Any): A file-like object or any other object.
|
||||||
|
|
||||||
|
Returns:
|
||||||
|
str: The name of the file if available, otherwise a string representation of the object.
|
||||||
|
"""
|
||||||
|
if hasattr(file_obj, "name"):
|
||||||
|
return getattr(file_obj, "name")
|
||||||
|
elif hasattr(file_obj, "__str__"):
|
||||||
|
return str(file_obj)
|
||||||
|
else:
|
||||||
|
return repr(file_obj)
|
|
@ -4,7 +4,6 @@ import os
|
||||||
from typing import List, Literal, Optional, Tuple
|
from typing import List, Literal, Optional, Tuple
|
||||||
|
|
||||||
from litellm._logging import verbose_logger
|
from litellm._logging import verbose_logger
|
||||||
from litellm.types.utils import FileTypes
|
|
||||||
|
|
||||||
|
|
||||||
def map_finish_reason(
|
def map_finish_reason(
|
||||||
|
@ -87,21 +86,3 @@ def _get_parent_otel_span_from_kwargs(kwargs: Optional[dict] = None):
|
||||||
return kwargs["litellm_parent_otel_span"]
|
return kwargs["litellm_parent_otel_span"]
|
||||||
except:
|
except:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def get_audio_file_name(file_obj: FileTypes) -> str:
|
|
||||||
"""
|
|
||||||
Safely get the name of a file-like object or return its string representation.
|
|
||||||
|
|
||||||
Args:
|
|
||||||
file_obj (Any): A file-like object or any other object.
|
|
||||||
|
|
||||||
Returns:
|
|
||||||
str: The name of the file if available, otherwise a string representation of the object.
|
|
||||||
"""
|
|
||||||
if hasattr(file_obj, "name"):
|
|
||||||
return getattr(file_obj, "name")
|
|
||||||
elif hasattr(file_obj, "__str__"):
|
|
||||||
return str(file_obj)
|
|
||||||
else:
|
|
||||||
return repr(file_obj)
|
|
||||||
|
|
|
@ -6,7 +6,7 @@ from openai import AsyncAzureOpenAI, AzureOpenAI
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
|
|
||||||
import litellm
|
import litellm
|
||||||
from litellm.litellm_core_utils.core_helpers import get_audio_file_name
|
from litellm.litellm_core_utils.audio_utils.utils import get_audio_file_name
|
||||||
from litellm.litellm_core_utils.litellm_logging import Logging as LiteLLMLoggingObj
|
from litellm.litellm_core_utils.litellm_logging import Logging as LiteLLMLoggingObj
|
||||||
from litellm.types.utils import FileTypes
|
from litellm.types.utils import FileTypes
|
||||||
from litellm.utils import TranscriptionResponse, convert_to_model_response_object
|
from litellm.utils import TranscriptionResponse, convert_to_model_response_object
|
||||||
|
|
|
@ -5,7 +5,7 @@ from openai import AsyncOpenAI, OpenAI
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
|
|
||||||
import litellm
|
import litellm
|
||||||
from litellm.litellm_core_utils.core_helpers import get_audio_file_name
|
from litellm.litellm_core_utils.audio_utils.utils import get_audio_file_name
|
||||||
from litellm.litellm_core_utils.litellm_logging import Logging as LiteLLMLoggingObj
|
from litellm.litellm_core_utils.litellm_logging import Logging as LiteLLMLoggingObj
|
||||||
from litellm.types.utils import FileTypes
|
from litellm.types.utils import FileTypes
|
||||||
from litellm.utils import TranscriptionResponse, convert_to_model_response_object
|
from litellm.utils import TranscriptionResponse, convert_to_model_response_object
|
||||||
|
|
|
@ -12,7 +12,7 @@ from litellm.llms.custom_httpx.http_handler import (
|
||||||
_get_async_httpx_client,
|
_get_async_httpx_client,
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.llms.openai import HttpxBinaryResponseContent
|
from litellm.llms.OpenAI.openai import HttpxBinaryResponseContent
|
||||||
from litellm.llms.vertex_ai_and_google_ai_studio.gemini.vertex_and_google_ai_studio_gemini import (
|
from litellm.llms.vertex_ai_and_google_ai_studio.gemini.vertex_and_google_ai_studio_gemini import (
|
||||||
VertexLLM,
|
VertexLLM,
|
||||||
)
|
)
|
||||||
|
|
|
@ -8,7 +8,7 @@ import openai
|
||||||
|
|
||||||
import litellm
|
import litellm
|
||||||
from litellm._logging import verbose_router_logger
|
from litellm._logging import verbose_router_logger
|
||||||
from litellm.llms.azure import get_azure_ad_token_from_oidc
|
from litellm.llms.AzureOpenAI.azure import get_azure_ad_token_from_oidc
|
||||||
from litellm.secret_managers.get_azure_ad_token_provider import (
|
from litellm.secret_managers.get_azure_ad_token_provider import (
|
||||||
get_azure_ad_token_provider,
|
get_azure_ad_token_provider,
|
||||||
)
|
)
|
||||||
|
@ -337,7 +337,9 @@ def set_client(litellm_router_instance: LitellmRouter, model: dict):
|
||||||
azure_client_params["azure_ad_token_provider"] = (
|
azure_client_params["azure_ad_token_provider"] = (
|
||||||
azure_ad_token_provider
|
azure_ad_token_provider
|
||||||
)
|
)
|
||||||
from litellm.llms.azure import select_azure_base_url_or_endpoint
|
from litellm.llms.AzureOpenAI.azure import (
|
||||||
|
select_azure_base_url_or_endpoint,
|
||||||
|
)
|
||||||
|
|
||||||
# this decides if we should set azure_endpoint or base_url on Azure OpenAI Client
|
# this decides if we should set azure_endpoint or base_url on Azure OpenAI Client
|
||||||
# required to support GPT-4 vision enhancements, since base_url needs to be set on Azure OpenAI Client
|
# required to support GPT-4 vision enhancements, since base_url needs to be set on Azure OpenAI Client
|
||||||
|
|
|
@ -55,12 +55,10 @@ from tokenizers import Tokenizer
|
||||||
import litellm
|
import litellm
|
||||||
import litellm._service_logger # for storing API inputs, outputs, and metadata
|
import litellm._service_logger # for storing API inputs, outputs, and metadata
|
||||||
import litellm.litellm_core_utils
|
import litellm.litellm_core_utils
|
||||||
|
import litellm.litellm_core_utils.audio_utils.utils
|
||||||
import litellm.litellm_core_utils.json_validation_rule
|
import litellm.litellm_core_utils.json_validation_rule
|
||||||
from litellm.caching import DualCache
|
from litellm.caching import DualCache
|
||||||
from litellm.litellm_core_utils.core_helpers import (
|
from litellm.litellm_core_utils.core_helpers import map_finish_reason
|
||||||
get_audio_file_name,
|
|
||||||
map_finish_reason,
|
|
||||||
)
|
|
||||||
from litellm.litellm_core_utils.exception_mapping_utils import get_error_message
|
from litellm.litellm_core_utils.exception_mapping_utils import get_error_message
|
||||||
from litellm.litellm_core_utils.get_llm_provider_logic import (
|
from litellm.litellm_core_utils.get_llm_provider_logic import (
|
||||||
_is_non_openai_azure_model,
|
_is_non_openai_azure_model,
|
||||||
|
@ -567,7 +565,11 @@ def function_setup(
|
||||||
or call_type == CallTypes.transcription.value
|
or call_type == CallTypes.transcription.value
|
||||||
):
|
):
|
||||||
_file_obj: FileTypes = args[1] if len(args) > 1 else kwargs["file"]
|
_file_obj: FileTypes = args[1] if len(args) > 1 else kwargs["file"]
|
||||||
file_checksum = get_audio_file_name(file_obj=_file_obj)
|
file_checksum = (
|
||||||
|
litellm.litellm_core_utils.audio_utils.utils.get_audio_file_name(
|
||||||
|
file_obj=_file_obj
|
||||||
|
)
|
||||||
|
)
|
||||||
if "metadata" in kwargs:
|
if "metadata" in kwargs:
|
||||||
kwargs["metadata"]["file_checksum"] = file_checksum
|
kwargs["metadata"]["file_checksum"] = file_checksum
|
||||||
else:
|
else:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue