mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 03:34:10 +00:00
fix(openai.py): fix client caching logic
This commit is contained in:
parent
466e7dee5b
commit
5d3a0ace4b
2 changed files with 4 additions and 1 deletions
|
@ -14,6 +14,7 @@ from functools import partial
|
|||
import dotenv, traceback, random, asyncio, time, contextvars
|
||||
from copy import deepcopy
|
||||
import httpx
|
||||
|
||||
import litellm
|
||||
from ._logging import verbose_logger
|
||||
from litellm import ( # type: ignore
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue