mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
# What does this PR do? Extract API definitions and provider specifications into a standalone llama-stack-api package that can be published to PyPI independently of the main llama-stack server. see: https://github.com/llamastack/llama-stack/pull/2978 and https://github.com/llamastack/llama-stack/pull/2978#issuecomment-3145115942 Motivation External providers currently import from llama-stack, which overrides the installed version and causes dependency conflicts. This separation allows external providers to: - Install only the type definitions they need without server dependencies - Avoid version conflicts with the installed llama-stack package - Be versioned and released independently This enables us to re-enable external provider module tests that were previously blocked by these import conflicts. Changes - Created llama-stack-api package with minimal dependencies (pydantic, jsonschema) - Moved APIs, providers datatypes, strong_typing, and schema_utils - Updated all imports from llama_stack.* to llama_stack_api.* - Configured local editable install for development workflow - Updated linting and type-checking configuration for both packages Next Steps - Publish llama-stack-api to PyPI - Update external provider dependencies - Re-enable external provider module tests Pre-cursor PRs to this one: - #4093 - #3954 - #4064 These PRs moved key pieces _out_ of the Api pkg, limiting the scope of change here. relates to #3237 ## Test Plan Package builds successfully and can be imported independently. All pre-commit hooks pass with expected exclusions maintained. --------- Signed-off-by: Charlie Doern <cdoern@redhat.com>
78 lines
3.3 KiB
Python
78 lines
3.3 KiB
Python
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
from types import SimpleNamespace
|
|
from unittest.mock import AsyncMock, MagicMock
|
|
|
|
import pytest
|
|
from llama_stack_api import OpenAIChatCompletionRequestWithExtraBody
|
|
from openai import AuthenticationError
|
|
|
|
from llama_stack.providers.remote.inference.bedrock.bedrock import BedrockInferenceAdapter
|
|
from llama_stack.providers.remote.inference.bedrock.config import BedrockConfig
|
|
|
|
|
|
def test_adapter_initialization():
|
|
config = BedrockConfig(api_key="test-key", region_name="us-east-1")
|
|
adapter = BedrockInferenceAdapter(config=config)
|
|
|
|
assert adapter.config.auth_credential.get_secret_value() == "test-key"
|
|
assert adapter.config.region_name == "us-east-1"
|
|
|
|
|
|
def test_client_url_construction():
|
|
config = BedrockConfig(api_key="test-key", region_name="us-west-2")
|
|
adapter = BedrockInferenceAdapter(config=config)
|
|
|
|
assert adapter.get_base_url() == "https://bedrock-runtime.us-west-2.amazonaws.com/openai/v1"
|
|
|
|
|
|
def test_api_key_from_config():
|
|
config = BedrockConfig(api_key="config-key", region_name="us-east-1")
|
|
adapter = BedrockInferenceAdapter(config=config)
|
|
assert adapter.config.auth_credential.get_secret_value() == "config-key"
|
|
|
|
|
|
def test_api_key_from_header_overrides_config():
|
|
"""Test API key from request header overrides config via client property"""
|
|
config = BedrockConfig(api_key="config-key", region_name="us-east-1")
|
|
adapter = BedrockInferenceAdapter(config=config)
|
|
adapter.provider_data_api_key_field = "aws_bedrock_api_key"
|
|
adapter.get_request_provider_data = MagicMock(return_value=SimpleNamespace(aws_bedrock_api_key="header-key"))
|
|
|
|
# The client property is where header override happens (in OpenAIMixin)
|
|
assert adapter.client.api_key == "header-key"
|
|
|
|
|
|
async def test_authentication_error_handling():
|
|
"""Test that AuthenticationError from OpenAI client is converted to ValueError with helpful message"""
|
|
config = BedrockConfig(api_key="invalid-key", region_name="us-east-1")
|
|
adapter = BedrockInferenceAdapter(config=config)
|
|
|
|
# Mock the parent class method to raise AuthenticationError
|
|
mock_response = MagicMock()
|
|
mock_response.message = "Invalid authentication credentials"
|
|
auth_error = AuthenticationError(message="Invalid authentication credentials", response=mock_response, body=None)
|
|
|
|
# Create a mock that raises the error
|
|
mock_super = AsyncMock(side_effect=auth_error)
|
|
|
|
# Patch the parent class method
|
|
original_method = BedrockInferenceAdapter.__bases__[0].openai_chat_completion
|
|
BedrockInferenceAdapter.__bases__[0].openai_chat_completion = mock_super
|
|
|
|
try:
|
|
with pytest.raises(ValueError) as exc_info:
|
|
params = OpenAIChatCompletionRequestWithExtraBody(
|
|
model="test-model", messages=[{"role": "user", "content": "test"}]
|
|
)
|
|
await adapter.openai_chat_completion(params=params)
|
|
|
|
assert "AWS Bedrock authentication failed" in str(exc_info.value)
|
|
assert "Please verify your API key" in str(exc_info.value)
|
|
finally:
|
|
# Restore original method
|
|
BedrockInferenceAdapter.__bases__[0].openai_chat_completion = original_method
|