mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-27 18:50:41 +00:00
# What does this PR do?
- Add Github workflow for publishing docker images.
- Manual Inputs
- We can use a (1) TestPyPi version / (2) build via released PyPi
version
**Notes**
- Keep this workflow manually triggered as we don't want to publish
nightly docker images
**Additional Changes**
- Resolve issue with running llama stack build in non-terminal device
```
File "/home/runner/.local/lib/python3.12/site-packages/llama_stack/distribution/utils/exec.py", line 25, in run_with_pty
old_settings = termios.tcgetattr(sys.stdin)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^
termios.error: (25, 'Inappropriate ioctl for device')
```
- Modified build_container.sh to work in non-terminal environment
## Test Plan
- Triggered workflow:
3562217878
<img width="1076" alt="image"
src="https://github.com/user-attachments/assets/f1b5cef6-05ab-49c7-b405-53abc9264734"
/>
- Tested published docker image
<img width="702" alt="image"
src="https://github.com/user-attachments/assets/e7135189-65c8-45d8-86f9-9f3be70e380b"
/>
- /tools API endpoints are served so that docker is correctly using the
TestPyPi package
<img width="296" alt="image"
src="https://github.com/user-attachments/assets/bbcaa7fe-c0a4-4d22-b600-90e3c254bbfd"
/>
- Published tagged images:
https://hub.docker.com/repositories/llamastack
<img width="947" alt="image"
src="https://github.com/user-attachments/assets/2a0a0494-4d45-4643-bc29-72154ecc54a5"
/>
## Sources
Please link relevant resources if necessary.
## Before submitting
- [ ] This PR fixes a typo or improves the docs (you can dismiss the
other checks if that's the case).
- [ ] Ran pre-commit to handle lint / formatting issues.
- [ ] Read the [contributor
guideline](https://github.com/meta-llama/llama-stack/blob/main/CONTRIBUTING.md),
Pull Request section?
- [ ] Updated relevant documentation.
- [ ] Wrote necessary unit or integration tests.
208 lines
5.5 KiB
Bash
Executable file
208 lines
5.5 KiB
Bash
Executable file
#!/bin/bash
|
|
|
|
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
LLAMA_MODELS_DIR=${LLAMA_MODELS_DIR:-}
|
|
LLAMA_STACK_DIR=${LLAMA_STACK_DIR:-}
|
|
TEST_PYPI_VERSION=${TEST_PYPI_VERSION:-}
|
|
PYPI_VERSION=${PYPI_VERSION:-}
|
|
BUILD_PLATFORM=${BUILD_PLATFORM:-}
|
|
|
|
if [ "$#" -lt 4 ]; then
|
|
echo "Usage: $0 <build_name> <docker_base> <pip_dependencies> [<special_pip_deps>]" >&2
|
|
echo "Example: $0 my-fastapi-app python:3.9-slim 'fastapi uvicorn' " >&2
|
|
exit 1
|
|
fi
|
|
|
|
special_pip_deps="$6"
|
|
|
|
set -euo pipefail
|
|
|
|
build_name="$1"
|
|
image_name="distribution-$build_name"
|
|
docker_base=$2
|
|
build_file_path=$3
|
|
host_build_dir=$4
|
|
pip_dependencies=$5
|
|
|
|
# Define color codes
|
|
RED='\033[0;31m'
|
|
GREEN='\033[0;32m'
|
|
NC='\033[0m' # No Color
|
|
|
|
SCRIPT_DIR=$(dirname "$(readlink -f "$0")")
|
|
REPO_DIR=$(dirname $(dirname "$SCRIPT_DIR"))
|
|
DOCKER_BINARY=${DOCKER_BINARY:-docker}
|
|
DOCKER_OPTS=${DOCKER_OPTS:-}
|
|
|
|
TEMP_DIR=$(mktemp -d)
|
|
|
|
add_to_docker() {
|
|
local input
|
|
output_file="$TEMP_DIR/Dockerfile"
|
|
if [ -t 0 ]; then
|
|
printf '%s\n' "$1" >>"$output_file"
|
|
else
|
|
# If stdin is not a terminal, read from it (heredoc)
|
|
cat >>"$output_file"
|
|
fi
|
|
}
|
|
|
|
# Update and install UBI9 components if UBI9 base image is used
|
|
if [[ $docker_base == *"registry.access.redhat.com/ubi9"* ]]; then
|
|
add_to_docker << EOF
|
|
FROM $docker_base
|
|
WORKDIR /app
|
|
|
|
RUN microdnf -y update && microdnf install -y iputils net-tools wget \
|
|
vim-minimal python3.11 python3.11-pip python3.11-wheel \
|
|
python3.11-setuptools && ln -s /bin/pip3.11 /bin/pip && ln -s /bin/python3.11 /bin/python && microdnf clean all
|
|
|
|
EOF
|
|
else
|
|
add_to_docker << EOF
|
|
FROM $docker_base
|
|
WORKDIR /app
|
|
|
|
RUN apt-get update && apt-get install -y \
|
|
iputils-ping net-tools iproute2 dnsutils telnet \
|
|
curl wget telnet \
|
|
procps psmisc lsof \
|
|
traceroute \
|
|
bubblewrap \
|
|
&& rm -rf /var/lib/apt/lists/*
|
|
|
|
EOF
|
|
fi
|
|
|
|
# Add pip dependencies first since llama-stack is what will change most often
|
|
# so we can reuse layers.
|
|
if [ -n "$pip_dependencies" ]; then
|
|
add_to_docker << EOF
|
|
RUN pip install --no-cache $pip_dependencies
|
|
EOF
|
|
fi
|
|
|
|
if [ -n "$special_pip_deps" ]; then
|
|
IFS='#' read -ra parts <<<"$special_pip_deps"
|
|
for part in "${parts[@]}"; do
|
|
add_to_docker <<EOF
|
|
RUN pip install --no-cache $part
|
|
EOF
|
|
done
|
|
fi
|
|
|
|
stack_mount="/app/llama-stack-source"
|
|
models_mount="/app/llama-models-source"
|
|
|
|
if [ -n "$LLAMA_STACK_DIR" ]; then
|
|
if [ ! -d "$LLAMA_STACK_DIR" ]; then
|
|
echo "${RED}Warning: LLAMA_STACK_DIR is set but directory does not exist: $LLAMA_STACK_DIR${NC}" >&2
|
|
exit 1
|
|
fi
|
|
|
|
# Install in editable format. We will mount the source code into the container
|
|
# so that changes will be reflected in the container without having to do a
|
|
# rebuild. This is just for development convenience.
|
|
add_to_docker << EOF
|
|
RUN pip install --no-cache -e $stack_mount
|
|
EOF
|
|
else
|
|
if [ -n "$TEST_PYPI_VERSION" ]; then
|
|
# these packages are damaged in test-pypi, so install them first
|
|
add_to_docker << EOF
|
|
RUN pip install fastapi libcst
|
|
EOF
|
|
add_to_docker << EOF
|
|
RUN pip install --no-cache --extra-index-url https://test.pypi.org/simple/ \
|
|
llama-models==$TEST_PYPI_VERSION llama-stack-client==$TEST_PYPI_VERSION llama-stack==$TEST_PYPI_VERSION
|
|
|
|
EOF
|
|
else
|
|
if [ -n "$PYPI_VERSION" ]; then
|
|
SPEC_VERSION="llama-stack==${PYPI_VERSION} llama-models==${PYPI_VERSION} llama-stack-client==${PYPI_VERSION}"
|
|
else
|
|
SPEC_VERSION="llama-stack"
|
|
fi
|
|
add_to_docker << EOF
|
|
RUN pip install --no-cache $SPEC_VERSION
|
|
EOF
|
|
fi
|
|
fi
|
|
|
|
if [ -n "$LLAMA_MODELS_DIR" ]; then
|
|
if [ ! -d "$LLAMA_MODELS_DIR" ]; then
|
|
echo "${RED}Warning: LLAMA_MODELS_DIR is set but directory does not exist: $LLAMA_MODELS_DIR${NC}" >&2
|
|
exit 1
|
|
fi
|
|
|
|
add_to_docker << EOF
|
|
RUN pip uninstall -y llama-models
|
|
RUN pip install --no-cache $models_mount
|
|
|
|
EOF
|
|
fi
|
|
|
|
add_to_docker << EOF
|
|
|
|
# This would be good in production but for debugging flexibility lets not add it right now
|
|
# We need a more solid production ready entrypoint.sh anyway
|
|
#
|
|
ENTRYPOINT ["python", "-m", "llama_stack.distribution.server.server", "--template", "$build_name"]
|
|
|
|
EOF
|
|
|
|
printf "Dockerfile created successfully in $TEMP_DIR/Dockerfile\n\n"
|
|
cat $TEMP_DIR/Dockerfile
|
|
printf "\n"
|
|
|
|
mounts=""
|
|
if [ -n "$LLAMA_STACK_DIR" ]; then
|
|
mounts="$mounts -v $(readlink -f $LLAMA_STACK_DIR):$stack_mount"
|
|
fi
|
|
if [ -n "$LLAMA_MODELS_DIR" ]; then
|
|
mounts="$mounts -v $(readlink -f $LLAMA_MODELS_DIR):$models_mount"
|
|
fi
|
|
|
|
if command -v selinuxenabled &>/dev/null && selinuxenabled; then
|
|
# Disable SELinux labels -- we don't want to relabel the llama-stack source dir
|
|
DOCKER_OPTS="$DOCKER_OPTS --security-opt label=disable"
|
|
fi
|
|
|
|
# Set version tag based on PyPI version
|
|
if [ -n "$TEST_PYPI_VERSION" ]; then
|
|
version_tag="test-$TEST_PYPI_VERSION"
|
|
elif [[ -n "$LLAMA_STACK_DIR" || -n "$LLAMA_MODELS_DIR" ]]; then
|
|
version_tag="dev"
|
|
else
|
|
URL="https://pypi.org/pypi/llama-stack/json"
|
|
version_tag=$(curl -s $URL | jq -r '.info.version')
|
|
fi
|
|
|
|
# Add version tag to image name
|
|
image_tag="$image_name:$version_tag"
|
|
|
|
# Detect platform architecture
|
|
ARCH=$(uname -m)
|
|
if [ -n "$BUILD_PLATFORM" ]; then
|
|
PLATFORM="--platform $BUILD_PLATFORM"
|
|
elif [ "$ARCH" = "arm64" ] || [ "$ARCH" = "aarch64" ]; then
|
|
PLATFORM="--platform linux/arm64"
|
|
elif [ "$ARCH" = "x86_64" ]; then
|
|
PLATFORM="--platform linux/amd64"
|
|
else
|
|
echo "Unsupported architecture: $ARCH"
|
|
exit 1
|
|
fi
|
|
|
|
set -x
|
|
$DOCKER_BINARY build $DOCKER_OPTS $PLATFORM -t $image_tag -f "$TEMP_DIR/Dockerfile" "$REPO_DIR" $mounts
|
|
|
|
# clean up tmp/configs
|
|
set +x
|
|
|
|
echo "Success!"
|