mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 07:14:20 +00:00
Test fixes
This commit is contained in:
parent
8eee5b9adc
commit
f40cd62306
3 changed files with 1 additions and 3 deletions
|
@ -84,7 +84,7 @@ class InferenceClient(Inference):
|
|||
|
||||
response.raise_for_status()
|
||||
j = response.json()
|
||||
yield ChatCompletionResponse(**j)
|
||||
return ChatCompletionResponse(**j)
|
||||
|
||||
async def _stream_chat_completion(
|
||||
self, request: ChatCompletionRequest
|
||||
|
|
|
@ -68,7 +68,6 @@ async def inference_settings(request):
|
|||
ModelDef(
|
||||
identifier=model,
|
||||
llama_model=model,
|
||||
provider_id="",
|
||||
)
|
||||
],
|
||||
)
|
||||
|
|
|
@ -67,7 +67,6 @@ def sample_documents():
|
|||
async def register_memory_bank(memory_impl: Memory):
|
||||
bank = VectorMemoryBankDef(
|
||||
identifier="test_bank",
|
||||
provider_id="",
|
||||
embedding_model="all-MiniLM-L6-v2",
|
||||
chunk_size_in_tokens=512,
|
||||
overlap_size_in_tokens=64,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue