forked from phoenix-oss/llama-stack-mirror
* Add distribution CLI scaffolding * More progress towards `llama distribution install` * getting closer to a distro definition, distro install + configure works * Distribution server now functioning * read existing configuration, save enums properly * Remove inference uvicorn server entrypoint and llama inference CLI command * updated dependency and client model name * Improved exception handling * local imports for faster cli * undo a typo, add a passthrough distribution * implement full-passthrough in the server * add safety adapters, configuration handling, server + clients * cleanup, moving stuff to common, nuke utils * Add a Path() wrapper at the earliest place * fixes * Bring agentic system api to toolchain Add adapter dependencies and resolve adapters using a topological sort * refactor to reduce size of `agentic_system` * move straggler files and fix some important existing bugs * ApiSurface -> Api * refactor a method out * Adapter -> Provider * Make each inference provider into its own subdirectory * installation fixes * Rename Distribution -> DistributionSpec, simplify RemoteProviders * dict key instead of attr * update inference config to take model and not model_dir * Fix passthrough streaming, send headers properly not part of body :facepalm * update safety to use model sku ids and not model dirs * Update cli_reference.md * minor fixes * add DistributionConfig, fix a bug in model download * Make install + start scripts do proper configuration automatically * Update CLI_reference * Nuke fp8_requirements, fold fbgemm into common requirements * Update README, add newline between API surface configurations * Refactor download functionality out of the Command so can be reused * Add `llama model download` alias for `llama download` * Show message about checksum file so users can check themselves * Simpler intro statements * get ollama working * Reduce a bunch of dependencies from toolchain Some improvements to the distribution install script * Avoid using `conda run` since it buffers everything * update dependencies and rely on LLAMA_TOOLCHAIN_DIR for dev purposes * add validation for configuration input * resort imports * make optional subclasses default to yes for configuration * Remove additional_pip_packages; move deps to providers * for inline make 8b model the default * Add scripts to MANIFEST * allow installing from test.pypi.org * Fix #2 to help with testing packages * Must install llama-models at that same version first * fix PIP_ARGS --------- Co-authored-by: Hardik Shah <hjshah@fb.com> Co-authored-by: Hardik Shah <hjshah@meta.com>
112 lines
3.6 KiB
Bash
Executable file
112 lines
3.6 KiB
Bash
Executable file
#!/bin/bash
|
|
|
|
# Copyright (c) Meta Platforms, Inc. and affiliates.
|
|
# All rights reserved.
|
|
#
|
|
# This source code is licensed under the terms described in the LICENSE file in
|
|
# the root directory of this source tree.
|
|
|
|
LLAMA_MODELS_DIR=${LLAMA_MODELS_DIR:-}
|
|
LLAMA_TOOLCHAIN_DIR=${LLAMA_TOOLCHAIN_DIR:-}
|
|
TEST_PYPI_VERSION=${TEST_PYPI_VERSION:-}
|
|
|
|
set -euo pipefail
|
|
|
|
# Define color codes
|
|
RED='\033[0;31m'
|
|
GREEN='\033[0;32m'
|
|
NC='\033[0m' # No Color
|
|
|
|
error_handler() {
|
|
echo "Error occurred in script at line: ${1}" >&2
|
|
exit 1
|
|
}
|
|
|
|
# Set up the error trap
|
|
trap 'error_handler ${LINENO}' ERR
|
|
|
|
ensure_conda_env_python310() {
|
|
local env_name="$1"
|
|
local pip_dependencies="$2"
|
|
local python_version="3.10"
|
|
|
|
# Check if conda command is available
|
|
if ! command -v conda &>/dev/null; then
|
|
echo -e "${RED}Error: conda command not found. Is Conda installed and in your PATH?${NC}" >&2
|
|
exit 1
|
|
fi
|
|
|
|
# Check if the environment exists
|
|
if conda env list | grep -q "^${env_name} "; then
|
|
echo "Conda environment '${env_name}' exists. Checking Python version..."
|
|
|
|
# Check Python version in the environment
|
|
current_version=$(conda run -n "${env_name}" python --version 2>&1 | cut -d' ' -f2 | cut -d'.' -f1,2)
|
|
|
|
if [ "$current_version" = "$python_version" ]; then
|
|
echo "Environment '${env_name}' already has Python ${python_version}. No action needed."
|
|
else
|
|
echo "Updating environment '${env_name}' to Python ${python_version}..."
|
|
conda install -n "${env_name}" python="${python_version}" -y
|
|
fi
|
|
else
|
|
echo "Conda environment '${env_name}' does not exist. Creating with Python ${python_version}..."
|
|
conda create -n "${env_name}" python="${python_version}" -y
|
|
fi
|
|
|
|
eval "$(conda shell.bash hook)"
|
|
conda deactivate && conda activate "${env_name}"
|
|
|
|
if [ -n "$TEST_PYPI_VERSION" ]; then
|
|
# these packages are damaged in test-pypi, so install them first
|
|
pip install fastapi libcst
|
|
pip install --extra-index-url https://test.pypi.org/simple/ llama-models==$TEST_PYPI_VERSION llama-toolchain==$TEST_PYPI_VERSION $pip_dependencies
|
|
else
|
|
# Re-installing llama-toolchain in the new conda environment
|
|
if [ -n "$LLAMA_TOOLCHAIN_DIR" ]; then
|
|
if [ ! -d "$LLAMA_TOOLCHAIN_DIR" ]; then
|
|
echo -e "${RED}Warning: LLAMA_TOOLCHAIN_DIR is set but directory does not exist: $LLAMA_TOOLCHAIN_DIR${NC}" >&2
|
|
exit 1
|
|
fi
|
|
|
|
echo "Installing from LLAMA_TOOLCHAIN_DIR: $LLAMA_TOOLCHAIN_DIR"
|
|
pip install -e "$LLAMA_TOOLCHAIN_DIR"
|
|
else
|
|
pip install llama-toolchain
|
|
fi
|
|
|
|
if [ -n "$LLAMA_MODELS_DIR" ]; then
|
|
if [ ! -d "$LLAMA_MODELS_DIR" ]; then
|
|
echo -e "${RED}Warning: LLAMA_MODELS_DIR is set but directory does not exist: $LLAMA_MODELS_DIR${NC}" >&2
|
|
exit 1
|
|
fi
|
|
|
|
echo "Installing from LLAMA_MODELS_DIR: $LLAMA_MODELS_DIR"
|
|
pip uninstall -y llama-models
|
|
pip install -e "$LLAMA_MODELS_DIR"
|
|
fi
|
|
|
|
# Install pip dependencies
|
|
if [ -n "$pip_dependencies" ]; then
|
|
echo "Installing pip dependencies: $pip_dependencies"
|
|
pip install $pip_dependencies
|
|
fi
|
|
fi
|
|
}
|
|
|
|
if [ "$#" -ne 3 ]; then
|
|
echo "Usage: $0 <environment_name> <distribution_name> <pip_dependencies>" >&2
|
|
echo "Example: $0 my_env local-inline 'numpy pandas scipy'" >&2
|
|
exit 1
|
|
fi
|
|
|
|
env_name="$1"
|
|
distribution_name="$2"
|
|
pip_dependencies="$3"
|
|
|
|
ensure_conda_env_python310 "$env_name" "$pip_dependencies"
|
|
|
|
echo -e "${GREEN}Successfully setup distribution environment. Configuring...${NC}"
|
|
|
|
python_interp=$(conda run -n "$env_name" which python)
|
|
$python_interp -m llama_toolchain.cli.llama distribution configure --name "$distribution_name"
|