diff --git a/litellm/__init__.py b/litellm/__init__.py index 4c9baac19..353d7ac5b 100644 --- a/litellm/__init__.py +++ b/litellm/__init__.py @@ -715,6 +715,7 @@ openai_image_generation_models = ["dall-e-2", "dall-e-3"] from .timeout import timeout from .cost_calculator import completion_cost +from litellm.litellm_core_utils.litellm_logging import Logging from .utils import ( client, exception_type, @@ -827,4 +828,3 @@ from .assistants.main import * from .batches.main import * from .scheduler import * from .cost_calculator import response_cost_calculator, cost_per_token -from litellm.litellm_core_utils.litellm_logging import Logging diff --git a/litellm/llms/bedrock.py b/litellm/llms/bedrock.py index 8d88cdd3d..73fa18023 100644 --- a/litellm/llms/bedrock.py +++ b/litellm/llms/bedrock.py @@ -7,7 +7,7 @@ import litellm from litellm.utils import ( get_secret, ) -from litellm.litellm_core_utils.model_response_helpers import map_finish_reason +from litellm.litellm_core_utils.core_helpers import map_finish_reason from litellm.types.utils import ImageResponse, ModelResponse, Usage from .prompt_templates.factory import ( prompt_factory, diff --git a/litellm/llms/vertex_ai.py b/litellm/llms/vertex_ai.py index 28cdde518..60d3d5897 100644 --- a/litellm/llms/vertex_ai.py +++ b/litellm/llms/vertex_ai.py @@ -6,7 +6,7 @@ import time from typing import Callable, Optional, Union, List, Literal, Any from pydantic import BaseModel from litellm.utils import ModelResponse, Usage, CustomStreamWrapper -from litellm.litellm_core_utils.model_response_helpers import map_finish_reason +from litellm.litellm_core_utils.core_helpers import map_finish_reason import litellm, uuid import httpx, inspect # type: ignore from litellm.types.llms.vertex_ai import * diff --git a/litellm/llms/vertex_ai_anthropic.py b/litellm/llms/vertex_ai_anthropic.py index 1907ad5f0..fd43d4378 100644 --- a/litellm/llms/vertex_ai_anthropic.py +++ b/litellm/llms/vertex_ai_anthropic.py @@ -7,7 +7,7 @@ import requests, copy # type: ignore import time, uuid from typing import Callable, Optional, List from litellm.utils import ModelResponse, Usage, CustomStreamWrapper -from litellm.litellm_core_utils.model_response_helpers import map_finish_reason +from litellm.litellm_core_utils.core_helpers import map_finish_reason import litellm from litellm.llms.custom_httpx.http_handler import AsyncHTTPHandler, HTTPHandler from .prompt_templates.factory import ( diff --git a/litellm/types/utils.py b/litellm/types/utils.py index 6e7a536e2..29d21143e 100644 --- a/litellm/types/utils.py +++ b/litellm/types/utils.py @@ -8,6 +8,7 @@ from openai._models import BaseModel as OpenAIObject from pydantic import ConfigDict import uuid import json +import time def _generate_id(): # private helper function diff --git a/litellm/utils.py b/litellm/utils.py index a126a10cd..498e2639e 100644 --- a/litellm/utils.py +++ b/litellm/utils.py @@ -32,8 +32,7 @@ from dataclasses import ( ) import os import litellm._service_logger # for storing API inputs, outputs, and metadata -import litellm.litellm_core_utils -import litellm.litellm_core_utils.litellm_logging +from litellm.litellm_core_utils.core_helpers import map_finish_reason from litellm.llms.custom_httpx.http_handler import HTTPHandler, AsyncHTTPHandler from litellm.caching import DualCache from litellm.types.utils import (