mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-09 21:18:38 +00:00
test
# What does this PR do? ## Test Plan
This commit is contained in:
parent
96886afaca
commit
1e891489a8
13 changed files with 170 additions and 4 deletions
|
@ -153,6 +153,7 @@ class PassthroughInferenceAdapter(Inference):
|
|||
top_logprobs: int | None = None,
|
||||
top_p: float | None = None,
|
||||
user: str | None = None,
|
||||
**kwargs: Any,
|
||||
) -> OpenAIChatCompletion | AsyncIterator[OpenAIChatCompletionChunk]:
|
||||
client = self._get_client()
|
||||
model_obj = await self.model_store.get_model(model)
|
||||
|
@ -181,6 +182,7 @@ class PassthroughInferenceAdapter(Inference):
|
|||
top_logprobs=top_logprobs,
|
||||
top_p=top_p,
|
||||
user=user,
|
||||
**kwargs,
|
||||
)
|
||||
|
||||
return await client.inference.openai_chat_completion(**params)
|
||||
|
|
|
@ -57,6 +57,7 @@ class RunpodInferenceAdapter(OpenAIMixin):
|
|||
top_logprobs: int | None = None,
|
||||
top_p: float | None = None,
|
||||
user: str | None = None,
|
||||
**kwargs: Any,
|
||||
):
|
||||
"""Override to add RunPod-specific stream_options requirement."""
|
||||
if stream and not stream_options:
|
||||
|
@ -86,4 +87,5 @@ class RunpodInferenceAdapter(OpenAIMixin):
|
|||
top_logprobs=top_logprobs,
|
||||
top_p=top_p,
|
||||
user=user,
|
||||
**kwargs,
|
||||
)
|
||||
|
|
|
@ -102,6 +102,7 @@ class VLLMInferenceAdapter(OpenAIMixin):
|
|||
top_logprobs: int | None = None,
|
||||
top_p: float | None = None,
|
||||
user: str | None = None,
|
||||
**kwargs: Any,
|
||||
) -> OpenAIChatCompletion | AsyncIterator[OpenAIChatCompletionChunk]:
|
||||
max_tokens = max_tokens or self.config.max_tokens
|
||||
|
||||
|
@ -136,4 +137,5 @@ class VLLMInferenceAdapter(OpenAIMixin):
|
|||
top_logprobs=top_logprobs,
|
||||
top_p=top_p,
|
||||
user=user,
|
||||
**kwargs,
|
||||
)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue