mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-19 04:19:40 +00:00
Implement additional functionality supported by Sambanova.
This commit is contained in:
parent
b6a79d6291
commit
8920c4216f
9 changed files with 565 additions and 203 deletions
|
|
@ -22,6 +22,7 @@ from llama_stack.providers.remote.inference.nvidia import NVIDIAConfig
|
|||
from llama_stack.providers.remote.inference.ollama import OllamaImplConfig
|
||||
from llama_stack.providers.remote.inference.together import TogetherImplConfig
|
||||
from llama_stack.providers.remote.inference.vllm import VLLMInferenceAdapterConfig
|
||||
from llama_stack.providers.remote.inference.sambanova import SambanovaImplConfig
|
||||
from llama_stack.providers.tests.resolver import construct_stack_for_test
|
||||
|
||||
from ..conftest import ProviderFixture, remote_stack_fixture
|
||||
|
|
@ -155,6 +156,21 @@ def inference_nvidia() -> ProviderFixture:
|
|||
],
|
||||
)
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def inference_sambanova() -> ProviderFixture:
|
||||
return ProviderFixture(
|
||||
providers=[
|
||||
Provider(
|
||||
provider_id="sambanova",
|
||||
provider_type="remote::sambanova",
|
||||
config=SambanovaImplConfig().model_dump(),
|
||||
)
|
||||
],
|
||||
provider_data=dict(
|
||||
sambanova_api_key=get_env_or_fail("SAMBANOVA_API_KEY"),
|
||||
),
|
||||
)
|
||||
|
||||
|
||||
def get_model_short_name(model_name: str) -> str:
|
||||
"""Convert model name to a short test identifier.
|
||||
|
|
@ -190,6 +206,7 @@ INFERENCE_FIXTURES = [
|
|||
"remote",
|
||||
"bedrock",
|
||||
"nvidia",
|
||||
"sambanova",
|
||||
]
|
||||
|
||||
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue