mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
Some checks failed
Integration Auth Tests / test-matrix (oauth2_token) (push) Failing after 1s
SqlStore Integration Tests / test-postgres (3.13) (push) Failing after 0s
Integration Tests (Replay) / generate-matrix (push) Successful in 3s
Test External Providers Installed via Module / test-external-providers-from-module (venv) (push) Has been skipped
SqlStore Integration Tests / test-postgres (3.12) (push) Failing after 4s
Test Llama Stack Build / generate-matrix (push) Successful in 4s
API Conformance Tests / check-schema-compatibility (push) Successful in 11s
Python Package Build Test / build (3.12) (push) Successful in 17s
Python Package Build Test / build (3.13) (push) Successful in 21s
Test Llama Stack Build / build-single-provider (push) Successful in 27s
Test External API and Providers / test-external (venv) (push) Failing after 28s
Vector IO Integration Tests / test-matrix (push) Failing after 37s
Test Llama Stack Build / build (push) Successful in 40s
UI Tests / ui-tests (22) (push) Successful in 1m18s
Unit Tests / unit-tests (3.12) (push) Failing after 1m50s
Unit Tests / unit-tests (3.13) (push) Failing after 2m9s
Test Llama Stack Build / build-custom-container-distribution (push) Successful in 2m41s
Test Llama Stack Build / build-ubi9-container-distribution (push) Successful in 2m51s
Pre-commit / pre-commit (push) Successful in 2m54s
Integration Tests (Replay) / Integration Tests (, , , client=, ) (push) Failing after 2m42s
# What does this PR do? This allows llama-stack users of the Docker image to use OpenTelemetry like previous versions. #4127 migrated to automatic instrumentation, but unless we add those libraries to the image, everyone needs to build a custom image to enable otel. Also, unless we establish a convention for enabling it, users who formerly just set config now need to override the entrypoint. This PR bootstraps OTEL packages, so they are available (only +10MB). It also prefixes `llama stack run` with `opentelemetry-instrument` when any `OTEL_*` environment variable is set. The result is implicit tracing like before, where you don't need a custom image to use traces or metrics. ## Test Plan ```bash # Build image docker build -f containers/Containerfile \ --build-arg DISTRO_NAME=starter \ --build-arg INSTALL_MODE=editable \ --tag llamastack/distribution-starter:otel-test . # Run with OTEL env to implicitly use `opentelemetry-instrument`. The # Settings below ensure inbound traces are honored, but no # "junk traces" like SQL connects are created. docker run -p 8321:8321 \ -e OTEL_EXPORTER_OTLP_ENDPOINT=http://host.docker.internal:4318 \ -e OTEL_SERVICE_NAME=llama-stack \ -e OTEL_TRACES_SAMPLER=parentbased_traceidratio \ -e OTEL_TRACES_SAMPLER_ARG=0.0 \ llamastack/distribution-starter:otel-test ``` Ran a sample flight search agent which is instrumented on the client side. This and llama-stack target [otel-tui](https://github.com/ymtdzzz/otel-tui) I verified no root database spans, yet database spans are attached to incoming traces. <img width="1608" height="742" alt="screenshot" src="https://github.com/user-attachments/assets/69f59b74-3054-42cd-947d-a6c0d9472a7c" /> Signed-off-by: Adrian Cole <adrian@tetrate.io>
163 lines
5.8 KiB
Docker
163 lines
5.8 KiB
Docker
# syntax=docker/dockerfile:1.6
|
|
#
|
|
# This Dockerfile is used to build the Llama Stack container image.
|
|
# Example:
|
|
# docker build \
|
|
# -f containers/Containerfile \
|
|
# --build-arg DISTRO_NAME=starter \
|
|
# --tag llama-stack:starter .
|
|
|
|
ARG BASE_IMAGE=python:3.12-slim
|
|
FROM ${BASE_IMAGE}
|
|
|
|
ARG INSTALL_MODE="pypi"
|
|
ARG LLAMA_STACK_DIR="/workspace"
|
|
ARG LLAMA_STACK_CLIENT_DIR=""
|
|
ARG PYPI_VERSION=""
|
|
ARG TEST_PYPI_VERSION=""
|
|
ARG KEEP_WORKSPACE=""
|
|
ARG DISTRO_NAME="starter"
|
|
ARG RUN_CONFIG_PATH=""
|
|
ARG UV_HTTP_TIMEOUT=500
|
|
ARG UV_EXTRA_INDEX_URL=""
|
|
ARG UV_INDEX_STRATEGY=""
|
|
ENV UV_HTTP_TIMEOUT=${UV_HTTP_TIMEOUT}
|
|
ENV PYTHONDONTWRITEBYTECODE=1
|
|
ENV PIP_DISABLE_PIP_VERSION_CHECK=1
|
|
WORKDIR /app
|
|
|
|
RUN set -eux; \
|
|
if command -v dnf >/dev/null 2>&1; then \
|
|
dnf -y update && \
|
|
dnf install -y iputils git net-tools wget \
|
|
vim-minimal python3.12 python3.12-pip python3.12-wheel \
|
|
python3.12-setuptools python3.12-devel gcc gcc-c++ make && \
|
|
ln -sf /usr/bin/pip3.12 /usr/local/bin/pip && \
|
|
ln -sf /usr/bin/python3.12 /usr/local/bin/python && \
|
|
dnf clean all; \
|
|
elif command -v apt-get >/dev/null 2>&1; then \
|
|
apt-get update && \
|
|
apt-get install -y --no-install-recommends \
|
|
iputils-ping net-tools iproute2 dnsutils telnet \
|
|
curl wget git procps psmisc lsof traceroute bubblewrap \
|
|
gcc g++ && \
|
|
rm -rf /var/lib/apt/lists/*; \
|
|
else \
|
|
echo "Unsupported base image: expected dnf or apt-get" >&2; \
|
|
exit 1; \
|
|
fi
|
|
|
|
RUN pip install --no-cache uv
|
|
ENV UV_SYSTEM_PYTHON=1
|
|
|
|
ENV INSTALL_MODE=${INSTALL_MODE}
|
|
ENV LLAMA_STACK_DIR=${LLAMA_STACK_DIR}
|
|
ENV LLAMA_STACK_CLIENT_DIR=${LLAMA_STACK_CLIENT_DIR}
|
|
ENV PYPI_VERSION=${PYPI_VERSION}
|
|
ENV TEST_PYPI_VERSION=${TEST_PYPI_VERSION}
|
|
ENV KEEP_WORKSPACE=${KEEP_WORKSPACE}
|
|
ENV DISTRO_NAME=${DISTRO_NAME}
|
|
ENV RUN_CONFIG_PATH=${RUN_CONFIG_PATH}
|
|
|
|
# Copy the repository so editable installs and run configurations are available.
|
|
COPY . /workspace
|
|
|
|
# Install the client package if it is provided
|
|
# NOTE: this is installed before llama-stack since llama-stack depends on llama-stack-client-python
|
|
# Unset UV index env vars to ensure we only use PyPI for the client
|
|
RUN set -eux; \
|
|
unset UV_EXTRA_INDEX_URL UV_INDEX_STRATEGY; \
|
|
if [ -n "$LLAMA_STACK_CLIENT_DIR" ]; then \
|
|
if [ ! -d "$LLAMA_STACK_CLIENT_DIR" ]; then \
|
|
echo "LLAMA_STACK_CLIENT_DIR is set but $LLAMA_STACK_CLIENT_DIR does not exist" >&2; \
|
|
exit 1; \
|
|
fi; \
|
|
uv pip install --no-cache -e "$LLAMA_STACK_CLIENT_DIR"; \
|
|
fi;
|
|
|
|
# Install llama-stack
|
|
# Use UV_EXTRA_INDEX_URL inline only for editable install with RC dependencies
|
|
RUN set -eux; \
|
|
SAVED_UV_EXTRA_INDEX_URL="${UV_EXTRA_INDEX_URL:-}"; \
|
|
SAVED_UV_INDEX_STRATEGY="${UV_INDEX_STRATEGY:-}"; \
|
|
unset UV_EXTRA_INDEX_URL UV_INDEX_STRATEGY; \
|
|
if [ "$INSTALL_MODE" = "editable" ]; then \
|
|
if [ ! -d "$LLAMA_STACK_DIR" ]; then \
|
|
echo "INSTALL_MODE=editable requires LLAMA_STACK_DIR to point to a directory inside the build context" >&2; \
|
|
exit 1; \
|
|
fi; \
|
|
if [ -n "$SAVED_UV_EXTRA_INDEX_URL" ] && [ -n "$SAVED_UV_INDEX_STRATEGY" ]; then \
|
|
UV_EXTRA_INDEX_URL="$SAVED_UV_EXTRA_INDEX_URL" UV_INDEX_STRATEGY="$SAVED_UV_INDEX_STRATEGY" \
|
|
uv pip install --no-cache -e "$LLAMA_STACK_DIR"; \
|
|
else \
|
|
uv pip install --no-cache -e "$LLAMA_STACK_DIR"; \
|
|
fi; \
|
|
elif [ "$INSTALL_MODE" = "test-pypi" ]; then \
|
|
uv pip install --no-cache fastapi libcst; \
|
|
if [ -n "$TEST_PYPI_VERSION" ]; then \
|
|
uv pip install --no-cache --extra-index-url https://test.pypi.org/simple/ --index-strategy unsafe-best-match "llama-stack==$TEST_PYPI_VERSION"; \
|
|
else \
|
|
uv pip install --no-cache --extra-index-url https://test.pypi.org/simple/ --index-strategy unsafe-best-match llama-stack; \
|
|
fi; \
|
|
else \
|
|
if [ -n "$PYPI_VERSION" ]; then \
|
|
uv pip install --no-cache "llama-stack==$PYPI_VERSION"; \
|
|
else \
|
|
uv pip install --no-cache llama-stack; \
|
|
fi; \
|
|
fi;
|
|
|
|
# Install the dependencies for the distribution
|
|
# Explicitly unset UV index env vars to ensure we only use PyPI for distribution deps
|
|
RUN set -eux; \
|
|
unset UV_EXTRA_INDEX_URL UV_INDEX_STRATEGY; \
|
|
if [ -z "$DISTRO_NAME" ]; then \
|
|
echo "DISTRO_NAME must be provided" >&2; \
|
|
exit 1; \
|
|
fi; \
|
|
deps="$(llama stack list-deps "$DISTRO_NAME")"; \
|
|
if [ -n "$deps" ]; then \
|
|
printf '%s\n' "$deps" | xargs -L1 uv pip install --no-cache; \
|
|
fi
|
|
|
|
# Install OpenTelemetry auto-instrumentation support
|
|
RUN set -eux; \
|
|
pip install --no-cache opentelemetry-distro opentelemetry-exporter-otlp; \
|
|
opentelemetry-bootstrap -a install
|
|
|
|
# Cleanup
|
|
RUN set -eux; \
|
|
pip uninstall -y uv; \
|
|
should_remove=1; \
|
|
if [ -n "$KEEP_WORKSPACE" ]; then should_remove=0; fi; \
|
|
if [ "$INSTALL_MODE" = "editable" ]; then should_remove=0; fi; \
|
|
case "$RUN_CONFIG_PATH" in \
|
|
/workspace*) should_remove=0 ;; \
|
|
esac; \
|
|
if [ "$should_remove" -eq 1 ] && [ -d /workspace ]; then rm -rf /workspace; fi
|
|
|
|
RUN cat <<'EOF' >/usr/local/bin/llama-stack-entrypoint.sh
|
|
#!/bin/sh
|
|
set -e
|
|
|
|
# Enable OpenTelemetry auto-instrumentation if any OTEL_* variable is set
|
|
CMD_PREFIX=""
|
|
if env | grep -q '^OTEL_'; then
|
|
CMD_PREFIX="opentelemetry-instrument"
|
|
fi
|
|
|
|
if [ -n "$RUN_CONFIG_PATH" ] && [ -f "$RUN_CONFIG_PATH" ]; then
|
|
exec $CMD_PREFIX llama stack run "$RUN_CONFIG_PATH" "$@"
|
|
fi
|
|
|
|
if [ -n "$DISTRO_NAME" ]; then
|
|
exec $CMD_PREFIX llama stack run "$DISTRO_NAME" "$@"
|
|
fi
|
|
|
|
exec $CMD_PREFIX llama stack run "$@"
|
|
EOF
|
|
RUN chmod +x /usr/local/bin/llama-stack-entrypoint.sh
|
|
|
|
RUN mkdir -p /.llama /.cache && chmod -R g+rw /app /.llama /.cache
|
|
|
|
ENTRYPOINT ["/usr/local/bin/llama-stack-entrypoint.sh"]
|