mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 07:14:20 +00:00
Test fixes
This commit is contained in:
parent
8eee5b9adc
commit
f40cd62306
3 changed files with 1 additions and 3 deletions
|
@ -84,7 +84,7 @@ class InferenceClient(Inference):
|
||||||
|
|
||||||
response.raise_for_status()
|
response.raise_for_status()
|
||||||
j = response.json()
|
j = response.json()
|
||||||
yield ChatCompletionResponse(**j)
|
return ChatCompletionResponse(**j)
|
||||||
|
|
||||||
async def _stream_chat_completion(
|
async def _stream_chat_completion(
|
||||||
self, request: ChatCompletionRequest
|
self, request: ChatCompletionRequest
|
||||||
|
|
|
@ -68,7 +68,6 @@ async def inference_settings(request):
|
||||||
ModelDef(
|
ModelDef(
|
||||||
identifier=model,
|
identifier=model,
|
||||||
llama_model=model,
|
llama_model=model,
|
||||||
provider_id="",
|
|
||||||
)
|
)
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
|
|
|
@ -67,7 +67,6 @@ def sample_documents():
|
||||||
async def register_memory_bank(memory_impl: Memory):
|
async def register_memory_bank(memory_impl: Memory):
|
||||||
bank = VectorMemoryBankDef(
|
bank = VectorMemoryBankDef(
|
||||||
identifier="test_bank",
|
identifier="test_bank",
|
||||||
provider_id="",
|
|
||||||
embedding_model="all-MiniLM-L6-v2",
|
embedding_model="all-MiniLM-L6-v2",
|
||||||
chunk_size_in_tokens=512,
|
chunk_size_in_tokens=512,
|
||||||
overlap_size_in_tokens=64,
|
overlap_size_in_tokens=64,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue