mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 18:54:30 +00:00
pass llm provider when creating async httpx clients
This commit is contained in:
parent
87bac7c026
commit
421b857714
14 changed files with 57 additions and 19 deletions
|
@ -565,8 +565,8 @@ class SagemakerLLM(BaseAWSLLM):
|
|||
):
|
||||
try:
|
||||
if client is None:
|
||||
client = (
|
||||
get_async_httpx_client()
|
||||
client = get_async_httpx_client(
|
||||
llm_provider=litellm.LlmProviders.SAGEMAKER
|
||||
) # Create a new client if none provided
|
||||
response = await client.post(
|
||||
api_base,
|
||||
|
@ -673,7 +673,9 @@ class SagemakerLLM(BaseAWSLLM):
|
|||
model_id: Optional[str],
|
||||
):
|
||||
timeout = 300.0
|
||||
async_handler = get_async_httpx_client()
|
||||
async_handler = get_async_httpx_client(
|
||||
llm_provider=litellm.LlmProviders.SAGEMAKER
|
||||
)
|
||||
|
||||
async_transform_prompt = asyncify(self._transform_prompt)
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue