From ffae192540d5bfbeccfaf43735ca95e5016b2ca3 Mon Sep 17 00:00:00 2001 From: Ben Browning Date: Thu, 10 Apr 2025 14:19:48 -0400 Subject: [PATCH] Bug fixes for together.ai OpenAI endpoints After actually running the test_openai_completion.py tests against together.ai, turns out there were a couple of bugs in the initial implementation. This fixes those. Signed-off-by: Ben Browning --- llama_stack/providers/remote/inference/together/together.py | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/llama_stack/providers/remote/inference/together/together.py b/llama_stack/providers/remote/inference/together/together.py index 2c9a7ec03..1615b8cd1 100644 --- a/llama_stack/providers/remote/inference/together/together.py +++ b/llama_stack/providers/remote/inference/together/together.py @@ -279,7 +279,7 @@ class TogetherInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProvi guided_choice: Optional[List[str]] = None, prompt_logprobs: Optional[int] = None, ) -> OpenAICompletion: - model_obj = await self._get_model(model) + model_obj = await self.model_store.get_model(model) params = await prepare_openai_completion_params( model=model_obj.provider_resource_id, prompt=prompt, @@ -298,8 +298,6 @@ class TogetherInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProvi temperature=temperature, top_p=top_p, user=user, - guided_choice=guided_choice, - prompt_logprobs=prompt_logprobs, ) return await self._get_openai_client().completions.create(**params) # type: ignore @@ -329,7 +327,7 @@ class TogetherInferenceAdapter(ModelRegistryHelper, Inference, NeedsRequestProvi top_p: Optional[float] = None, user: Optional[str] = None, ) -> OpenAIChatCompletion: - model_obj = await self._get_model(model) + model_obj = await self.model_store.get_model(model) params = await prepare_openai_completion_params( model=model_obj.provider_resource_id, messages=messages,