mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 03:34:10 +00:00
pass llm provider when creating async httpx clients
This commit is contained in:
parent
9d102e7054
commit
0d6081e370
14 changed files with 57 additions and 19 deletions
|
@ -1244,6 +1244,7 @@ class QdrantSemanticCache(BaseCache):
|
|||
from litellm.llms.custom_httpx.http_handler import (
|
||||
_get_httpx_client,
|
||||
get_async_httpx_client,
|
||||
httpxSpecialProvider,
|
||||
)
|
||||
|
||||
if collection_name is None:
|
||||
|
@ -1290,7 +1291,9 @@ class QdrantSemanticCache(BaseCache):
|
|||
self.headers = headers
|
||||
|
||||
self.sync_client = _get_httpx_client()
|
||||
self.async_client = get_async_httpx_client()
|
||||
self.async_client = get_async_httpx_client(
|
||||
llm_provider=httpxSpecialProvider.Caching
|
||||
)
|
||||
|
||||
if quantization_config is None:
|
||||
print_verbose(
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue