mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-30 07:39:38 +00:00
fix return type and remove debug print
This commit is contained in:
parent
0e7e4bfb35
commit
09d5986ed8
1 changed files with 3 additions and 3 deletions
|
@ -295,7 +295,9 @@ class BedrockInferenceAdapter(ModelRegistryHelper, Inference):
|
||||||
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
|
tool_prompt_format: Optional[ToolPromptFormat] = ToolPromptFormat.json,
|
||||||
stream: Optional[bool] = False,
|
stream: Optional[bool] = False,
|
||||||
logprobs: Optional[LogProbConfig] = None,
|
logprobs: Optional[LogProbConfig] = None,
|
||||||
) -> AsyncGenerator:
|
) -> Union[
|
||||||
|
ChatCompletionResponse, AsyncIterator[ChatCompletionResponseStreamChunk]
|
||||||
|
]:
|
||||||
request = ChatCompletionRequest(
|
request = ChatCompletionRequest(
|
||||||
model=model,
|
model=model,
|
||||||
messages=messages,
|
messages=messages,
|
||||||
|
@ -316,7 +318,6 @@ class BedrockInferenceAdapter(ModelRegistryHelper, Inference):
|
||||||
async def _nonstream_chat_completion(
|
async def _nonstream_chat_completion(
|
||||||
self, request: ChatCompletionRequest
|
self, request: ChatCompletionRequest
|
||||||
) -> ChatCompletionResponse:
|
) -> ChatCompletionResponse:
|
||||||
print("non-streaming chat completion")
|
|
||||||
params = self._get_params_for_chat_completion(request)
|
params = self._get_params_for_chat_completion(request)
|
||||||
converse_api_res = self.client.converse(**params)
|
converse_api_res = self.client.converse(**params)
|
||||||
|
|
||||||
|
@ -332,7 +333,6 @@ class BedrockInferenceAdapter(ModelRegistryHelper, Inference):
|
||||||
async def _stream_chat_completion(
|
async def _stream_chat_completion(
|
||||||
self, request: ChatCompletionRequest
|
self, request: ChatCompletionRequest
|
||||||
) -> AsyncGenerator:
|
) -> AsyncGenerator:
|
||||||
print("streaming chat completion")
|
|
||||||
params = self._get_params_for_chat_completion(request)
|
params = self._get_params_for_chat_completion(request)
|
||||||
converse_stream_api_res = self.client.converse_stream(**params)
|
converse_stream_api_res = self.client.converse_stream(**params)
|
||||||
event_stream = converse_stream_api_res["stream"]
|
event_stream = converse_stream_api_res["stream"]
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue