mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-18 07:18:53 +00:00
init
This commit is contained in:
parent
7ee9f8d8ac
commit
d960f9b60f
16 changed files with 140 additions and 105 deletions
|
@ -10,6 +10,7 @@ import pytest
|
|||
import pytest_asyncio
|
||||
|
||||
from llama_stack.distribution.datatypes import Api, Provider
|
||||
from llama_stack.providers.adapters.inference.bedrock import BedrockConfig
|
||||
from llama_stack.providers.inline.inference.meta_reference import (
|
||||
MetaReferenceInferenceConfig,
|
||||
)
|
||||
|
@ -127,13 +128,33 @@ def inference_together() -> ProviderFixture:
|
|||
)
|
||||
|
||||
|
||||
@pytest.fixture(scope="session")
|
||||
def inference_bedrock() -> ProviderFixture:
|
||||
return ProviderFixture(
|
||||
providers=[
|
||||
Provider(
|
||||
provider_id="bedrock",
|
||||
provider_type="remote::bedrock",
|
||||
config=BedrockConfig().model_dump(),
|
||||
)
|
||||
],
|
||||
)
|
||||
|
||||
|
||||
INFERENCE_FIXTURES = [
|
||||
|
||||
"meta_reference",
|
||||
|
||||
"ollama",
|
||||
|
||||
"fireworks",
|
||||
|
||||
"together",
|
||||
|
||||
"vllm_remote",
|
||||
"remote",
|
||||
"bedrock",
|
||||
,
|
||||
]
|
||||
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue