mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-17 10:28:11 +00:00
Some checks failed
Integration Tests / discover-tests (push) Successful in 3s
Vector IO Integration Tests / test-matrix (3.12, remote::pgvector) (push) Failing after 4s
Python Package Build Test / build (3.12) (push) Failing after 3s
Integration Auth Tests / test-matrix (oauth2_token) (push) Failing after 9s
Vector IO Integration Tests / test-matrix (3.13, inline::faiss) (push) Failing after 8s
Integration Tests / test-matrix (push) Failing after 4s
Python Package Build Test / build (3.13) (push) Failing after 7s
Vector IO Integration Tests / test-matrix (3.12, inline::faiss) (push) Failing after 13s
Vector IO Integration Tests / test-matrix (3.13, remote::chromadb) (push) Failing after 10s
Vector IO Integration Tests / test-matrix (3.12, inline::milvus) (push) Failing after 12s
Vector IO Integration Tests / test-matrix (3.12, inline::sqlite-vec) (push) Failing after 12s
Unit Tests / unit-tests (3.12) (push) Failing after 8s
Vector IO Integration Tests / test-matrix (3.12, remote::chromadb) (push) Failing after 13s
Test External Providers / test-external-providers (venv) (push) Failing after 9s
Vector IO Integration Tests / test-matrix (3.13, inline::milvus) (push) Failing after 13s
Vector IO Integration Tests / test-matrix (3.13, inline::sqlite-vec) (push) Failing after 12s
Vector IO Integration Tests / test-matrix (3.13, remote::pgvector) (push) Failing after 12s
Unit Tests / unit-tests (3.13) (push) Failing after 10s
SqlStore Integration Tests / test-postgres (3.12) (push) Failing after 18s
Update ReadTheDocs / update-readthedocs (push) Failing after 40s
SqlStore Integration Tests / test-postgres (3.13) (push) Failing after 51s
Pre-commit / pre-commit (push) Successful in 2m1s
# What does this PR do? https://github.com/meta-llama/llama-stack/pull/2490 introduced a new function for type conversion of strings. However, a side effect of this is that it will cast any string that can be cast to an integer if possible, which for something like `image_name` is not desired as we only accept strings for this field in the `StackRunConfig` This PR introduces logic to ensure that `image_name` remains a string Closes #2749 ## Test Plan You can run the original step to reproduce from the bug to verify this manually ```bash OPENAI_API_KEY=bogus llama stack build --image-type venv --image-name 2745 --providers inference=remote::openai --run ``` I have also added an additional unit test to prevent any future regression here Signed-off-by: Nathan Weinberg <nweinber@redhat.com>
163 lines
4.9 KiB
Python
163 lines
4.9 KiB
Python
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
from datetime import datetime
|
|
|
|
import pytest
|
|
import yaml
|
|
|
|
from llama_stack.distribution.configure import (
|
|
LLAMA_STACK_RUN_CONFIG_VERSION,
|
|
parse_and_maybe_upgrade_config,
|
|
)
|
|
|
|
|
|
@pytest.fixture
|
|
def config_with_image_name_int():
|
|
return yaml.safe_load(
|
|
f"""
|
|
version: {LLAMA_STACK_RUN_CONFIG_VERSION}
|
|
image_name: 1234
|
|
apis_to_serve: []
|
|
built_at: {datetime.now().isoformat()}
|
|
providers:
|
|
inference:
|
|
- provider_id: provider1
|
|
provider_type: inline::meta-reference
|
|
config: {{}}
|
|
safety:
|
|
- provider_id: provider1
|
|
provider_type: inline::meta-reference
|
|
config:
|
|
llama_guard_shield:
|
|
model: Llama-Guard-3-1B
|
|
excluded_categories: []
|
|
disable_input_check: false
|
|
disable_output_check: false
|
|
enable_prompt_guard: false
|
|
memory:
|
|
- provider_id: provider1
|
|
provider_type: inline::meta-reference
|
|
config: {{}}
|
|
"""
|
|
)
|
|
|
|
|
|
@pytest.fixture
|
|
def up_to_date_config():
|
|
return yaml.safe_load(
|
|
f"""
|
|
version: {LLAMA_STACK_RUN_CONFIG_VERSION}
|
|
image_name: foo
|
|
apis_to_serve: []
|
|
built_at: {datetime.now().isoformat()}
|
|
providers:
|
|
inference:
|
|
- provider_id: provider1
|
|
provider_type: inline::meta-reference
|
|
config: {{}}
|
|
safety:
|
|
- provider_id: provider1
|
|
provider_type: inline::meta-reference
|
|
config:
|
|
llama_guard_shield:
|
|
model: Llama-Guard-3-1B
|
|
excluded_categories: []
|
|
disable_input_check: false
|
|
disable_output_check: false
|
|
enable_prompt_guard: false
|
|
memory:
|
|
- provider_id: provider1
|
|
provider_type: inline::meta-reference
|
|
config: {{}}
|
|
"""
|
|
)
|
|
|
|
|
|
@pytest.fixture
|
|
def old_config():
|
|
return yaml.safe_load(
|
|
f"""
|
|
image_name: foo
|
|
built_at: {datetime.now().isoformat()}
|
|
apis_to_serve: []
|
|
routing_table:
|
|
inference:
|
|
- provider_type: remote::ollama
|
|
config:
|
|
host: localhost
|
|
port: 11434
|
|
routing_key: Llama3.2-1B-Instruct
|
|
- provider_type: inline::meta-reference
|
|
config:
|
|
model: Llama3.1-8B-Instruct
|
|
routing_key: Llama3.1-8B-Instruct
|
|
safety:
|
|
- routing_key: ["shield1", "shield2"]
|
|
provider_type: inline::meta-reference
|
|
config:
|
|
llama_guard_shield:
|
|
model: Llama-Guard-3-1B
|
|
excluded_categories: []
|
|
disable_input_check: false
|
|
disable_output_check: false
|
|
enable_prompt_guard: false
|
|
memory:
|
|
- routing_key: vector
|
|
provider_type: inline::meta-reference
|
|
config: {{}}
|
|
api_providers:
|
|
telemetry:
|
|
provider_type: noop
|
|
config: {{}}
|
|
"""
|
|
)
|
|
|
|
|
|
@pytest.fixture
|
|
def invalid_config():
|
|
return yaml.safe_load(
|
|
"""
|
|
routing_table: {}
|
|
api_providers: {}
|
|
"""
|
|
)
|
|
|
|
|
|
def test_parse_and_maybe_upgrade_config_up_to_date(up_to_date_config):
|
|
result = parse_and_maybe_upgrade_config(up_to_date_config)
|
|
assert result.version == LLAMA_STACK_RUN_CONFIG_VERSION
|
|
assert "inference" in result.providers
|
|
|
|
|
|
def test_parse_and_maybe_upgrade_config_old_format(old_config):
|
|
result = parse_and_maybe_upgrade_config(old_config)
|
|
assert result.version == LLAMA_STACK_RUN_CONFIG_VERSION
|
|
assert all(api in result.providers for api in ["inference", "safety", "memory", "telemetry"])
|
|
safety_provider = result.providers["safety"][0]
|
|
assert safety_provider.provider_type == "inline::meta-reference"
|
|
assert "llama_guard_shield" in safety_provider.config
|
|
|
|
inference_providers = result.providers["inference"]
|
|
assert len(inference_providers) == 2
|
|
assert {x.provider_id for x in inference_providers} == {
|
|
"remote::ollama-00",
|
|
"inline::meta-reference-01",
|
|
}
|
|
|
|
ollama = inference_providers[0]
|
|
assert ollama.provider_type == "remote::ollama"
|
|
assert ollama.config["port"] == 11434
|
|
|
|
|
|
def test_parse_and_maybe_upgrade_config_invalid(invalid_config):
|
|
with pytest.raises(KeyError):
|
|
parse_and_maybe_upgrade_config(invalid_config)
|
|
|
|
|
|
def test_parse_and_maybe_upgrade_config_image_name_int(config_with_image_name_int):
|
|
result = parse_and_maybe_upgrade_config(config_with_image_name_int)
|
|
assert isinstance(result.image_name, str)
|