forked from phoenix/litellm-mirror
rename get_async_httpx_client
This commit is contained in:
parent
1e8cf9f2a6
commit
d4b9a1307d
13 changed files with 27 additions and 27 deletions
|
@ -1242,8 +1242,8 @@ class QdrantSemanticCache(BaseCache):
|
||||||
import os
|
import os
|
||||||
|
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
|
|
||||||
if collection_name is None:
|
if collection_name is None:
|
||||||
|
@ -1290,7 +1290,7 @@ class QdrantSemanticCache(BaseCache):
|
||||||
self.headers = headers
|
self.headers = headers
|
||||||
|
|
||||||
self.sync_client = _get_httpx_client()
|
self.sync_client = _get_httpx_client()
|
||||||
self.async_client = _get_async_httpx_client()
|
self.async_client = get_async_httpx_client()
|
||||||
|
|
||||||
if quantization_config is None:
|
if quantization_config is None:
|
||||||
print_verbose(
|
print_verbose(
|
||||||
|
|
|
@ -25,8 +25,8 @@ from litellm.litellm_core_utils.core_helpers import map_finish_reason
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
HTTPHandler,
|
HTTPHandler,
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.types.llms.anthropic import (
|
from litellm.types.llms.anthropic import (
|
||||||
AnthopicMessagesAssistantMessageParam,
|
AnthopicMessagesAssistantMessageParam,
|
||||||
|
@ -918,7 +918,7 @@ class AnthropicChatCompletion(BaseLLM):
|
||||||
headers={},
|
headers={},
|
||||||
client=None,
|
client=None,
|
||||||
) -> Union[ModelResponse, CustomStreamWrapper]:
|
) -> Union[ModelResponse, CustomStreamWrapper]:
|
||||||
async_handler = _get_async_httpx_client()
|
async_handler = get_async_httpx_client()
|
||||||
|
|
||||||
try:
|
try:
|
||||||
response = await async_handler.post(
|
response = await async_handler.post(
|
||||||
|
|
|
@ -35,8 +35,8 @@ from litellm.litellm_core_utils.litellm_logging import Logging
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
HTTPHandler,
|
HTTPHandler,
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.types.llms.bedrock import *
|
from litellm.types.llms.bedrock import *
|
||||||
from litellm.types.llms.openai import (
|
from litellm.types.llms.openai import (
|
||||||
|
@ -209,7 +209,7 @@ async def make_call(
|
||||||
):
|
):
|
||||||
try:
|
try:
|
||||||
if client is None:
|
if client is None:
|
||||||
client = _get_async_httpx_client() # Create a new client if none provided
|
client = get_async_httpx_client() # Create a new client if none provided
|
||||||
|
|
||||||
response = await client.post(
|
response = await client.post(
|
||||||
api_base,
|
api_base,
|
||||||
|
@ -1041,7 +1041,7 @@ class BedrockLLM(BaseAWSLLM):
|
||||||
if isinstance(timeout, float) or isinstance(timeout, int):
|
if isinstance(timeout, float) or isinstance(timeout, int):
|
||||||
timeout = httpx.Timeout(timeout)
|
timeout = httpx.Timeout(timeout)
|
||||||
_params["timeout"] = timeout
|
_params["timeout"] = timeout
|
||||||
client = _get_async_httpx_client(_params) # type: ignore
|
client = get_async_httpx_client(_params) # type: ignore
|
||||||
else:
|
else:
|
||||||
client = client # type: ignore
|
client = client # type: ignore
|
||||||
|
|
||||||
|
@ -1498,7 +1498,7 @@ class BedrockConverseLLM(BaseAWSLLM):
|
||||||
if isinstance(timeout, float) or isinstance(timeout, int):
|
if isinstance(timeout, float) or isinstance(timeout, int):
|
||||||
timeout = httpx.Timeout(timeout)
|
timeout = httpx.Timeout(timeout)
|
||||||
_params["timeout"] = timeout
|
_params["timeout"] = timeout
|
||||||
client = _get_async_httpx_client(_params) # type: ignore
|
client = get_async_httpx_client(_params) # type: ignore
|
||||||
else:
|
else:
|
||||||
client = client # type: ignore
|
client = client # type: ignore
|
||||||
|
|
||||||
|
|
|
@ -15,8 +15,8 @@ from litellm.llms.cohere.embed import embedding as cohere_embedding
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
HTTPHandler,
|
HTTPHandler,
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.secret_managers.main import get_secret
|
from litellm.secret_managers.main import get_secret
|
||||||
from litellm.types.llms.bedrock import AmazonEmbeddingRequest, CohereEmbeddingRequest
|
from litellm.types.llms.bedrock import AmazonEmbeddingRequest, CohereEmbeddingRequest
|
||||||
|
@ -130,7 +130,7 @@ class BedrockEmbedding(BaseAWSLLM):
|
||||||
if isinstance(timeout, float) or isinstance(timeout, int):
|
if isinstance(timeout, float) or isinstance(timeout, int):
|
||||||
timeout = httpx.Timeout(timeout)
|
timeout = httpx.Timeout(timeout)
|
||||||
_params["timeout"] = timeout
|
_params["timeout"] = timeout
|
||||||
client = _get_async_httpx_client(_params) # type: ignore
|
client = get_async_httpx_client(_params) # type: ignore
|
||||||
else:
|
else:
|
||||||
client = client
|
client = client
|
||||||
|
|
||||||
|
|
|
@ -11,8 +11,8 @@ from pydantic import BaseModel
|
||||||
|
|
||||||
from litellm.llms.base import BaseLLM
|
from litellm.llms.base import BaseLLM
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.rerank_api.types import RerankRequest, RerankResponse
|
from litellm.rerank_api.types import RerankRequest, RerankResponse
|
||||||
|
|
||||||
|
@ -65,7 +65,7 @@ class CohereRerank(BaseLLM):
|
||||||
api_key: str,
|
api_key: str,
|
||||||
api_base: str,
|
api_base: str,
|
||||||
) -> RerankResponse:
|
) -> RerankResponse:
|
||||||
client = _get_async_httpx_client()
|
client = get_async_httpx_client()
|
||||||
|
|
||||||
response = await client.post(
|
response = await client.post(
|
||||||
api_base,
|
api_base,
|
||||||
|
|
|
@ -378,7 +378,7 @@ class HTTPHandler:
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
def _get_async_httpx_client(params: Optional[dict] = None) -> AsyncHTTPHandler:
|
def get_async_httpx_client(params: Optional[dict] = None) -> AsyncHTTPHandler:
|
||||||
"""
|
"""
|
||||||
Retrieves the async HTTP client from the cache
|
Retrieves the async HTTP client from the cache
|
||||||
If not present, creates a new client
|
If not present, creates a new client
|
||||||
|
|
|
@ -19,8 +19,8 @@ from litellm.litellm_core_utils.asyncify import asyncify
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
HTTPHandler,
|
HTTPHandler,
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.types.llms.openai import (
|
from litellm.types.llms.openai import (
|
||||||
ChatCompletionToolCallChunk,
|
ChatCompletionToolCallChunk,
|
||||||
|
@ -566,7 +566,7 @@ class SagemakerLLM(BaseAWSLLM):
|
||||||
try:
|
try:
|
||||||
if client is None:
|
if client is None:
|
||||||
client = (
|
client = (
|
||||||
_get_async_httpx_client()
|
get_async_httpx_client()
|
||||||
) # Create a new client if none provided
|
) # Create a new client if none provided
|
||||||
response = await client.post(
|
response = await client.post(
|
||||||
api_base,
|
api_base,
|
||||||
|
@ -673,7 +673,7 @@ class SagemakerLLM(BaseAWSLLM):
|
||||||
model_id: Optional[str],
|
model_id: Optional[str],
|
||||||
):
|
):
|
||||||
timeout = 300.0
|
timeout = 300.0
|
||||||
async_handler = _get_async_httpx_client()
|
async_handler = get_async_httpx_client()
|
||||||
|
|
||||||
async_transform_prompt = asyncify(self._transform_prompt)
|
async_transform_prompt = asyncify(self._transform_prompt)
|
||||||
|
|
||||||
|
|
|
@ -11,8 +11,8 @@ from pydantic import BaseModel
|
||||||
|
|
||||||
from litellm.llms.base import BaseLLM
|
from litellm.llms.base import BaseLLM
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.rerank_api.types import RerankRequest, RerankResponse
|
from litellm.rerank_api.types import RerankRequest, RerankResponse
|
||||||
|
|
||||||
|
@ -77,7 +77,7 @@ class TogetherAIRerank(BaseLLM):
|
||||||
request_data_dict: Dict[str, Any],
|
request_data_dict: Dict[str, Any],
|
||||||
api_key: str,
|
api_key: str,
|
||||||
) -> RerankResponse:
|
) -> RerankResponse:
|
||||||
client = _get_async_httpx_client() # Use async client
|
client = get_async_httpx_client() # Use async client
|
||||||
|
|
||||||
response = await client.post(
|
response = await client.post(
|
||||||
"https://api.together.xyz/v1/rerank",
|
"https://api.together.xyz/v1/rerank",
|
||||||
|
|
|
@ -22,7 +22,7 @@ from litellm.litellm_core_utils.core_helpers import map_finish_reason
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
HTTPHandler,
|
HTTPHandler,
|
||||||
_get_async_httpx_client,
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.llms.prompt_templates.factory import (
|
from litellm.llms.prompt_templates.factory import (
|
||||||
convert_url_to_base64,
|
convert_url_to_base64,
|
||||||
|
@ -1293,7 +1293,7 @@ class VertexLLM(BaseLLM):
|
||||||
_async_client_params = {}
|
_async_client_params = {}
|
||||||
if timeout:
|
if timeout:
|
||||||
_async_client_params["timeout"] = timeout
|
_async_client_params["timeout"] = timeout
|
||||||
client = _get_async_httpx_client(params=_async_client_params)
|
client = get_async_httpx_client(params=_async_client_params)
|
||||||
## LOGGING
|
## LOGGING
|
||||||
logging_obj.pre_call(
|
logging_obj.pre_call(
|
||||||
input=messages,
|
input=messages,
|
||||||
|
|
|
@ -9,8 +9,8 @@ from litellm.llms.base import BaseLLM
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
HTTPHandler,
|
HTTPHandler,
|
||||||
_get_async_httpx_client,
|
|
||||||
_get_httpx_client,
|
_get_httpx_client,
|
||||||
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.llms.OpenAI.openai import HttpxBinaryResponseContent
|
from litellm.llms.OpenAI.openai import HttpxBinaryResponseContent
|
||||||
from litellm.llms.vertex_ai_and_google_ai_studio.gemini.vertex_and_google_ai_studio_gemini import (
|
from litellm.llms.vertex_ai_and_google_ai_studio.gemini.vertex_and_google_ai_studio_gemini import (
|
||||||
|
@ -178,7 +178,7 @@ class VertexTextToSpeechAPI(VertexLLM):
|
||||||
) -> HttpxBinaryResponseContent:
|
) -> HttpxBinaryResponseContent:
|
||||||
import base64
|
import base64
|
||||||
|
|
||||||
async_handler = _get_async_httpx_client()
|
async_handler = get_async_httpx_client()
|
||||||
|
|
||||||
response = await async_handler.post(
|
response = await async_handler.post(
|
||||||
url=url,
|
url=url,
|
||||||
|
|
|
@ -20,7 +20,7 @@ async def check_oauth2_token(token: str) -> UserAPIKeyAuth:
|
||||||
import httpx
|
import httpx
|
||||||
|
|
||||||
from litellm._logging import verbose_proxy_logger
|
from litellm._logging import verbose_proxy_logger
|
||||||
from litellm.llms.custom_httpx.http_handler import _get_async_httpx_client
|
from litellm.llms.custom_httpx.http_handler import get_async_httpx_client
|
||||||
from litellm.proxy._types import CommonProxyErrors
|
from litellm.proxy._types import CommonProxyErrors
|
||||||
from litellm.proxy.proxy_server import premium_user
|
from litellm.proxy.proxy_server import premium_user
|
||||||
|
|
||||||
|
@ -40,7 +40,7 @@ async def check_oauth2_token(token: str) -> UserAPIKeyAuth:
|
||||||
if not token_info_endpoint:
|
if not token_info_endpoint:
|
||||||
raise ValueError("OAUTH_TOKEN_INFO_ENDPOINT environment variable is not set")
|
raise ValueError("OAUTH_TOKEN_INFO_ENDPOINT environment variable is not set")
|
||||||
|
|
||||||
client = _get_async_httpx_client()
|
client = get_async_httpx_client()
|
||||||
headers = {"Authorization": f"Bearer {token}", "Content-Type": "application/json"}
|
headers = {"Authorization": f"Bearer {token}", "Content-Type": "application/json"}
|
||||||
|
|
||||||
try:
|
try:
|
||||||
|
|
|
@ -33,7 +33,7 @@ from litellm.litellm_core_utils.logging_utils import (
|
||||||
from litellm.llms.base_aws_llm import BaseAWSLLM
|
from litellm.llms.base_aws_llm import BaseAWSLLM
|
||||||
from litellm.llms.custom_httpx.http_handler import (
|
from litellm.llms.custom_httpx.http_handler import (
|
||||||
AsyncHTTPHandler,
|
AsyncHTTPHandler,
|
||||||
_get_async_httpx_client,
|
get_async_httpx_client,
|
||||||
)
|
)
|
||||||
from litellm.proxy._types import UserAPIKeyAuth
|
from litellm.proxy._types import UserAPIKeyAuth
|
||||||
from litellm.proxy.guardrails.guardrail_helpers import should_proceed_based_on_metadata
|
from litellm.proxy.guardrails.guardrail_helpers import should_proceed_based_on_metadata
|
||||||
|
@ -55,7 +55,7 @@ class BedrockGuardrail(CustomGuardrail, BaseAWSLLM):
|
||||||
guardrailVersion: Optional[str] = None,
|
guardrailVersion: Optional[str] = None,
|
||||||
**kwargs,
|
**kwargs,
|
||||||
):
|
):
|
||||||
self.async_handler = _get_async_httpx_client()
|
self.async_handler = get_async_httpx_client()
|
||||||
self.guardrailIdentifier = guardrailIdentifier
|
self.guardrailIdentifier = guardrailIdentifier
|
||||||
self.guardrailVersion = guardrailVersion
|
self.guardrailVersion = guardrailVersion
|
||||||
|
|
||||||
|
|
|
@ -22,7 +22,7 @@ import litellm # noqa: E401
|
||||||
from litellm._logging import verbose_proxy_logger
|
from litellm._logging import verbose_proxy_logger
|
||||||
from litellm.caching import DualCache
|
from litellm.caching import DualCache
|
||||||
from litellm.integrations.custom_guardrail import CustomGuardrail
|
from litellm.integrations.custom_guardrail import CustomGuardrail
|
||||||
from litellm.llms.custom_httpx.http_handler import _get_async_httpx_client
|
from litellm.llms.custom_httpx.http_handler import get_async_httpx_client
|
||||||
from litellm.proxy._types import UserAPIKeyAuth
|
from litellm.proxy._types import UserAPIKeyAuth
|
||||||
from litellm.utils import (
|
from litellm.utils import (
|
||||||
EmbeddingResponse,
|
EmbeddingResponse,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue