mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-18 15:27:16 +00:00
bedrock test for inference fixes
This commit is contained in:
parent
e0f227f23c
commit
98c09323a9
3 changed files with 4 additions and 10 deletions
|
@ -10,10 +10,10 @@ import pytest
|
|||
import pytest_asyncio
|
||||
|
||||
from llama_stack.distribution.datatypes import Api, Provider
|
||||
from llama_stack.providers.adapters.inference.bedrock import BedrockConfig
|
||||
from llama_stack.providers.inline.inference.meta_reference import (
|
||||
MetaReferenceInferenceConfig,
|
||||
)
|
||||
from llama_stack.providers.remote.inference.bedrock import BedrockConfig
|
||||
|
||||
from llama_stack.providers.remote.inference.fireworks import FireworksImplConfig
|
||||
from llama_stack.providers.remote.inference.ollama import OllamaImplConfig
|
||||
|
@ -142,19 +142,13 @@ def inference_bedrock() -> ProviderFixture:
|
|||
|
||||
|
||||
INFERENCE_FIXTURES = [
|
||||
|
||||
"meta_reference",
|
||||
|
||||
"ollama",
|
||||
|
||||
"fireworks",
|
||||
|
||||
"together",
|
||||
|
||||
"vllm_remote",
|
||||
"remote",
|
||||
"bedrock",
|
||||
,
|
||||
]
|
||||
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue