mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-07 04:45:44 +00:00
clean up
This commit is contained in:
parent
a6baa7b3d4
commit
0b9480a0e3
5 changed files with 0 additions and 18 deletions
|
@ -590,7 +590,6 @@ class InferenceRouter(Inference):
|
|||
|
||||
async def _nonstream_openai_chat_completion(self, provider: Inference, params: dict) -> OpenAIChatCompletion:
|
||||
response = await provider.openai_chat_completion(**params)
|
||||
|
||||
for choice in response.choices:
|
||||
# some providers return an empty list for no tool calls in non-streaming responses
|
||||
# but the OpenAI API returns None. So, set tool_calls to None if it's empty
|
||||
|
|
|
@ -61,7 +61,6 @@ MODEL_ENTRIES = [
|
|||
),
|
||||
ProviderModelEntry(
|
||||
provider_model_id="nomic-ai/nomic-embed-text-v1.5",
|
||||
aliases=["nomic-ai/nomic-embed-text-v1.5"],
|
||||
model_type=ModelType.embedding,
|
||||
metadata={
|
||||
"embedding_dimension": 768,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue