mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-18 19:39:47 +00:00
Tests pass with Ollama now
This commit is contained in:
parent
a9a041a1de
commit
e51154964f
27 changed files with 83 additions and 65 deletions
|
|
@ -17,7 +17,7 @@ from llama_stack.apis.inference import * # noqa: F403
|
|||
@json_schema_type
|
||||
class BatchCompletionRequest(BaseModel):
|
||||
model: str
|
||||
content_batch: List[InterleavedTextMedia]
|
||||
content_batch: List[InterleavedContent]
|
||||
sampling_params: Optional[SamplingParams] = SamplingParams()
|
||||
logprobs: Optional[LogProbConfig] = None
|
||||
|
||||
|
|
@ -53,7 +53,7 @@ class BatchInference(Protocol):
|
|||
async def batch_completion(
|
||||
self,
|
||||
model: str,
|
||||
content_batch: List[InterleavedTextMedia],
|
||||
content_batch: List[InterleavedContent],
|
||||
sampling_params: Optional[SamplingParams] = SamplingParams(),
|
||||
logprobs: Optional[LogProbConfig] = None,
|
||||
) -> BatchCompletionResponse: ...
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue