From e8d20b9c505d41078fbb435302a926080828e9ed Mon Sep 17 00:00:00 2001 From: Emilio Garcia Date: Thu, 30 Oct 2025 14:31:38 -0400 Subject: [PATCH] fix(scripts): add metrics interval to env variables --- scripts/integration-tests.sh | 17 +++++++++-------- tests/integration/telemetry/test_completions.py | 3 +-- 2 files changed, 10 insertions(+), 10 deletions(-) diff --git a/scripts/integration-tests.sh b/scripts/integration-tests.sh index 6d2b272b4..372e97d8c 100755 --- a/scripts/integration-tests.sh +++ b/scripts/integration-tests.sh @@ -227,14 +227,15 @@ if [[ "$STACK_CONFIG" == *"server:"* && "$COLLECT_ONLY" == false ]]; then echo "=== Starting Llama Stack Server ===" export LLAMA_STACK_LOG_WIDTH=120 - # Configure telemetry collector for server mode - # Use a fixed port for the OTEL collector so the server can connect to it - COLLECTOR_PORT=4317 - export LLAMA_STACK_TEST_COLLECTOR_PORT="${COLLECTOR_PORT}" - export OTEL_EXPORTER_OTLP_ENDPOINT="http://127.0.0.1:${COLLECTOR_PORT}" - export OTEL_EXPORTER_OTLP_PROTOCOL="http/protobuf" - export OTEL_BSP_SCHEDULE_DELAY="200" - export OTEL_BSP_EXPORT_TIMEOUT="2000" + # Configure telemetry collector for server mode + # Use a fixed port for the OTEL collector so the server can connect to it + COLLECTOR_PORT=4317 + export LLAMA_STACK_TEST_COLLECTOR_PORT="${COLLECTOR_PORT}" + export OTEL_EXPORTER_OTLP_ENDPOINT="http://127.0.0.1:${COLLECTOR_PORT}" + export OTEL_EXPORTER_OTLP_PROTOCOL="http/protobuf" + export OTEL_BSP_SCHEDULE_DELAY="200" + export OTEL_BSP_EXPORT_TIMEOUT="2000" + export OTEL_METRIC_EXPORT_INTERVAL="200" # remove "server:" from STACK_CONFIG stack_config=$(echo "$STACK_CONFIG" | sed 's/^server://') diff --git a/tests/integration/telemetry/test_completions.py b/tests/integration/telemetry/test_completions.py index 695f0c036..2b8835f6c 100644 --- a/tests/integration/telemetry/test_completions.py +++ b/tests/integration/telemetry/test_completions.py @@ -113,8 +113,7 @@ def test_telemetry_format_completeness(mock_otlp_collector, llama_stack_client, assert len(metrics) > 0, "No metrics found within timeout" # Filter metrics to only those from the specific model used in the request - # This prevents issues when multiple metrics with the same name exist from different models - # (e.g., when safety models like llama-guard are also called) + # Multiple metrics with the same name can exist (e.g., from safety models) inference_model_metrics = {} all_model_ids = set()