forked from phoenix/litellm-mirror
fix(bedrock_httpx.py): cache response stream shape
Fixes https://github.com/BerriAI/litellm/issues/4774
This commit is contained in:
parent
30d6c1d171
commit
e3ca2789df
1 changed files with 13 additions and 6 deletions
|
@ -77,7 +77,9 @@ BEDROCK_CONVERSE_MODELS = [
|
|||
"anthropic.claude-instant-v1",
|
||||
]
|
||||
|
||||
|
||||
iam_cache = DualCache()
|
||||
_response_stream_shape_cache = None
|
||||
|
||||
|
||||
class AmazonCohereChatConfig:
|
||||
|
@ -1991,13 +1993,18 @@ class BedrockConverseLLM(BaseLLM):
|
|||
|
||||
|
||||
def get_response_stream_shape():
|
||||
from botocore.loaders import Loader
|
||||
from botocore.model import ServiceModel
|
||||
global _response_stream_shape_cache
|
||||
if _response_stream_shape_cache is None:
|
||||
|
||||
loader = Loader()
|
||||
bedrock_service_dict = loader.load_service_model("bedrock-runtime", "service-2")
|
||||
bedrock_service_model = ServiceModel(bedrock_service_dict)
|
||||
return bedrock_service_model.shape_for("ResponseStream")
|
||||
from botocore.loaders import Loader
|
||||
from botocore.model import ServiceModel
|
||||
|
||||
loader = Loader()
|
||||
bedrock_service_dict = loader.load_service_model("bedrock-runtime", "service-2")
|
||||
bedrock_service_model = ServiceModel(bedrock_service_dict)
|
||||
_response_stream_shape_cache = bedrock_service_model.shape_for("ResponseStream")
|
||||
|
||||
return _response_stream_shape_cache
|
||||
|
||||
|
||||
class AWSEventStreamDecoder:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue