rename get_async_httpx_client

This commit is contained in:
Ishaan Jaff 2024-09-10 10:38:01 -07:00
parent 1e8cf9f2a6
commit d4b9a1307d
13 changed files with 27 additions and 27 deletions

View file

@ -1242,8 +1242,8 @@ class QdrantSemanticCache(BaseCache):
import os import os
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
if collection_name is None: if collection_name is None:
@ -1290,7 +1290,7 @@ class QdrantSemanticCache(BaseCache):
self.headers = headers self.headers = headers
self.sync_client = _get_httpx_client() self.sync_client = _get_httpx_client()
self.async_client = _get_async_httpx_client() self.async_client = get_async_httpx_client()
if quantization_config is None: if quantization_config is None:
print_verbose( print_verbose(

View file

@ -25,8 +25,8 @@ from litellm.litellm_core_utils.core_helpers import map_finish_reason
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
HTTPHandler, HTTPHandler,
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.types.llms.anthropic import ( from litellm.types.llms.anthropic import (
AnthopicMessagesAssistantMessageParam, AnthopicMessagesAssistantMessageParam,
@ -918,7 +918,7 @@ class AnthropicChatCompletion(BaseLLM):
headers={}, headers={},
client=None, client=None,
) -> Union[ModelResponse, CustomStreamWrapper]: ) -> Union[ModelResponse, CustomStreamWrapper]:
async_handler = _get_async_httpx_client() async_handler = get_async_httpx_client()
try: try:
response = await async_handler.post( response = await async_handler.post(

View file

@ -35,8 +35,8 @@ from litellm.litellm_core_utils.litellm_logging import Logging
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
HTTPHandler, HTTPHandler,
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.types.llms.bedrock import * from litellm.types.llms.bedrock import *
from litellm.types.llms.openai import ( from litellm.types.llms.openai import (
@ -209,7 +209,7 @@ async def make_call(
): ):
try: try:
if client is None: if client is None:
client = _get_async_httpx_client() # Create a new client if none provided client = get_async_httpx_client() # Create a new client if none provided
response = await client.post( response = await client.post(
api_base, api_base,
@ -1041,7 +1041,7 @@ class BedrockLLM(BaseAWSLLM):
if isinstance(timeout, float) or isinstance(timeout, int): if isinstance(timeout, float) or isinstance(timeout, int):
timeout = httpx.Timeout(timeout) timeout = httpx.Timeout(timeout)
_params["timeout"] = timeout _params["timeout"] = timeout
client = _get_async_httpx_client(_params) # type: ignore client = get_async_httpx_client(_params) # type: ignore
else: else:
client = client # type: ignore client = client # type: ignore
@ -1498,7 +1498,7 @@ class BedrockConverseLLM(BaseAWSLLM):
if isinstance(timeout, float) or isinstance(timeout, int): if isinstance(timeout, float) or isinstance(timeout, int):
timeout = httpx.Timeout(timeout) timeout = httpx.Timeout(timeout)
_params["timeout"] = timeout _params["timeout"] = timeout
client = _get_async_httpx_client(_params) # type: ignore client = get_async_httpx_client(_params) # type: ignore
else: else:
client = client # type: ignore client = client # type: ignore

View file

@ -15,8 +15,8 @@ from litellm.llms.cohere.embed import embedding as cohere_embedding
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
HTTPHandler, HTTPHandler,
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.secret_managers.main import get_secret from litellm.secret_managers.main import get_secret
from litellm.types.llms.bedrock import AmazonEmbeddingRequest, CohereEmbeddingRequest from litellm.types.llms.bedrock import AmazonEmbeddingRequest, CohereEmbeddingRequest
@ -130,7 +130,7 @@ class BedrockEmbedding(BaseAWSLLM):
if isinstance(timeout, float) or isinstance(timeout, int): if isinstance(timeout, float) or isinstance(timeout, int):
timeout = httpx.Timeout(timeout) timeout = httpx.Timeout(timeout)
_params["timeout"] = timeout _params["timeout"] = timeout
client = _get_async_httpx_client(_params) # type: ignore client = get_async_httpx_client(_params) # type: ignore
else: else:
client = client client = client

View file

@ -11,8 +11,8 @@ from pydantic import BaseModel
from litellm.llms.base import BaseLLM from litellm.llms.base import BaseLLM
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.rerank_api.types import RerankRequest, RerankResponse from litellm.rerank_api.types import RerankRequest, RerankResponse
@ -65,7 +65,7 @@ class CohereRerank(BaseLLM):
api_key: str, api_key: str,
api_base: str, api_base: str,
) -> RerankResponse: ) -> RerankResponse:
client = _get_async_httpx_client() client = get_async_httpx_client()
response = await client.post( response = await client.post(
api_base, api_base,

View file

@ -378,7 +378,7 @@ class HTTPHandler:
pass pass
def _get_async_httpx_client(params: Optional[dict] = None) -> AsyncHTTPHandler: def get_async_httpx_client(params: Optional[dict] = None) -> AsyncHTTPHandler:
""" """
Retrieves the async HTTP client from the cache Retrieves the async HTTP client from the cache
If not present, creates a new client If not present, creates a new client

View file

@ -19,8 +19,8 @@ from litellm.litellm_core_utils.asyncify import asyncify
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
HTTPHandler, HTTPHandler,
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.types.llms.openai import ( from litellm.types.llms.openai import (
ChatCompletionToolCallChunk, ChatCompletionToolCallChunk,
@ -566,7 +566,7 @@ class SagemakerLLM(BaseAWSLLM):
try: try:
if client is None: if client is None:
client = ( client = (
_get_async_httpx_client() get_async_httpx_client()
) # Create a new client if none provided ) # Create a new client if none provided
response = await client.post( response = await client.post(
api_base, api_base,
@ -673,7 +673,7 @@ class SagemakerLLM(BaseAWSLLM):
model_id: Optional[str], model_id: Optional[str],
): ):
timeout = 300.0 timeout = 300.0
async_handler = _get_async_httpx_client() async_handler = get_async_httpx_client()
async_transform_prompt = asyncify(self._transform_prompt) async_transform_prompt = asyncify(self._transform_prompt)

View file

@ -11,8 +11,8 @@ from pydantic import BaseModel
from litellm.llms.base import BaseLLM from litellm.llms.base import BaseLLM
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.rerank_api.types import RerankRequest, RerankResponse from litellm.rerank_api.types import RerankRequest, RerankResponse
@ -77,7 +77,7 @@ class TogetherAIRerank(BaseLLM):
request_data_dict: Dict[str, Any], request_data_dict: Dict[str, Any],
api_key: str, api_key: str,
) -> RerankResponse: ) -> RerankResponse:
client = _get_async_httpx_client() # Use async client client = get_async_httpx_client() # Use async client
response = await client.post( response = await client.post(
"https://api.together.xyz/v1/rerank", "https://api.together.xyz/v1/rerank",

View file

@ -22,7 +22,7 @@ from litellm.litellm_core_utils.core_helpers import map_finish_reason
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
HTTPHandler, HTTPHandler,
_get_async_httpx_client, get_async_httpx_client,
) )
from litellm.llms.prompt_templates.factory import ( from litellm.llms.prompt_templates.factory import (
convert_url_to_base64, convert_url_to_base64,
@ -1293,7 +1293,7 @@ class VertexLLM(BaseLLM):
_async_client_params = {} _async_client_params = {}
if timeout: if timeout:
_async_client_params["timeout"] = timeout _async_client_params["timeout"] = timeout
client = _get_async_httpx_client(params=_async_client_params) client = get_async_httpx_client(params=_async_client_params)
## LOGGING ## LOGGING
logging_obj.pre_call( logging_obj.pre_call(
input=messages, input=messages,

View file

@ -9,8 +9,8 @@ from litellm.llms.base import BaseLLM
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
HTTPHandler, HTTPHandler,
_get_async_httpx_client,
_get_httpx_client, _get_httpx_client,
get_async_httpx_client,
) )
from litellm.llms.OpenAI.openai import HttpxBinaryResponseContent from litellm.llms.OpenAI.openai import HttpxBinaryResponseContent
from litellm.llms.vertex_ai_and_google_ai_studio.gemini.vertex_and_google_ai_studio_gemini import ( from litellm.llms.vertex_ai_and_google_ai_studio.gemini.vertex_and_google_ai_studio_gemini import (
@ -178,7 +178,7 @@ class VertexTextToSpeechAPI(VertexLLM):
) -> HttpxBinaryResponseContent: ) -> HttpxBinaryResponseContent:
import base64 import base64
async_handler = _get_async_httpx_client() async_handler = get_async_httpx_client()
response = await async_handler.post( response = await async_handler.post(
url=url, url=url,

View file

@ -20,7 +20,7 @@ async def check_oauth2_token(token: str) -> UserAPIKeyAuth:
import httpx import httpx
from litellm._logging import verbose_proxy_logger from litellm._logging import verbose_proxy_logger
from litellm.llms.custom_httpx.http_handler import _get_async_httpx_client from litellm.llms.custom_httpx.http_handler import get_async_httpx_client
from litellm.proxy._types import CommonProxyErrors from litellm.proxy._types import CommonProxyErrors
from litellm.proxy.proxy_server import premium_user from litellm.proxy.proxy_server import premium_user
@ -40,7 +40,7 @@ async def check_oauth2_token(token: str) -> UserAPIKeyAuth:
if not token_info_endpoint: if not token_info_endpoint:
raise ValueError("OAUTH_TOKEN_INFO_ENDPOINT environment variable is not set") raise ValueError("OAUTH_TOKEN_INFO_ENDPOINT environment variable is not set")
client = _get_async_httpx_client() client = get_async_httpx_client()
headers = {"Authorization": f"Bearer {token}", "Content-Type": "application/json"} headers = {"Authorization": f"Bearer {token}", "Content-Type": "application/json"}
try: try:

View file

@ -33,7 +33,7 @@ from litellm.litellm_core_utils.logging_utils import (
from litellm.llms.base_aws_llm import BaseAWSLLM from litellm.llms.base_aws_llm import BaseAWSLLM
from litellm.llms.custom_httpx.http_handler import ( from litellm.llms.custom_httpx.http_handler import (
AsyncHTTPHandler, AsyncHTTPHandler,
_get_async_httpx_client, get_async_httpx_client,
) )
from litellm.proxy._types import UserAPIKeyAuth from litellm.proxy._types import UserAPIKeyAuth
from litellm.proxy.guardrails.guardrail_helpers import should_proceed_based_on_metadata from litellm.proxy.guardrails.guardrail_helpers import should_proceed_based_on_metadata
@ -55,7 +55,7 @@ class BedrockGuardrail(CustomGuardrail, BaseAWSLLM):
guardrailVersion: Optional[str] = None, guardrailVersion: Optional[str] = None,
**kwargs, **kwargs,
): ):
self.async_handler = _get_async_httpx_client() self.async_handler = get_async_httpx_client()
self.guardrailIdentifier = guardrailIdentifier self.guardrailIdentifier = guardrailIdentifier
self.guardrailVersion = guardrailVersion self.guardrailVersion = guardrailVersion

View file

@ -22,7 +22,7 @@ import litellm # noqa: E401
from litellm._logging import verbose_proxy_logger from litellm._logging import verbose_proxy_logger
from litellm.caching import DualCache from litellm.caching import DualCache
from litellm.integrations.custom_guardrail import CustomGuardrail from litellm.integrations.custom_guardrail import CustomGuardrail
from litellm.llms.custom_httpx.http_handler import _get_async_httpx_client from litellm.llms.custom_httpx.http_handler import get_async_httpx_client
from litellm.proxy._types import UserAPIKeyAuth from litellm.proxy._types import UserAPIKeyAuth
from litellm.utils import ( from litellm.utils import (
EmbeddingResponse, EmbeddingResponse,