diff --git a/distributions/sambanova/build.yaml b/distributions/sambanova/build.yaml index 1915a0acd..dbf013d2d 100644 --- a/distributions/sambanova/build.yaml +++ b/distributions/sambanova/build.yaml @@ -1 +1 @@ -../../llama_stack/templates/sambanova/build.yaml \ No newline at end of file +../../llama_stack/templates/sambanova/build.yaml diff --git a/distributions/sambanova/run.yaml b/distributions/sambanova/run.yaml index 3cbd6b4ac..385282c67 100644 --- a/distributions/sambanova/run.yaml +++ b/distributions/sambanova/run.yaml @@ -1 +1 @@ -../../llama_stack/templates/sambanova/run.yaml \ No newline at end of file +../../llama_stack/templates/sambanova/run.yaml diff --git a/llama_stack/providers/remote/inference/sambanova/sambanova.py b/llama_stack/providers/remote/inference/sambanova/sambanova.py index a1e85baf5..da446567a 100644 --- a/llama_stack/providers/remote/inference/sambanova/sambanova.py +++ b/llama_stack/providers/remote/inference/sambanova/sambanova.py @@ -7,7 +7,12 @@ import json from typing import AsyncGenerator -from llama_models.datatypes import CoreModelId, TopPSamplingStrategy, TopKSamplingStrategy, GreedySamplingStrategy +from llama_models.datatypes import ( + CoreModelId, + GreedySamplingStrategy, + TopKSamplingStrategy, + TopPSamplingStrategy, +) from llama_models.llama3.api.chat_format import ChatFormat from llama_models.llama3.api.tokenizer import Tokenizer from openai import OpenAI diff --git a/llama_stack/templates/sambanova/build.yaml b/llama_stack/templates/sambanova/build.yaml index c6c01848a..0966bfdd9 100644 --- a/llama_stack/templates/sambanova/build.yaml +++ b/llama_stack/templates/sambanova/build.yaml @@ -30,4 +30,3 @@ distribution_spec: - inline::rag-runtime - remote::model-context-protocol image_type: conda -