mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-05 12:21:52 +00:00
llama-toolchain -> llama-stack (hyphens)
This commit is contained in:
parent
6665d31cdf
commit
d1959e6889
9 changed files with 17 additions and 17 deletions
|
@ -106,7 +106,7 @@ def _hf_download(
|
|||
local_dir=output_dir,
|
||||
ignore_patterns=ignore_patterns,
|
||||
token=hf_token,
|
||||
library_name="llama-toolchain",
|
||||
library_name="llama-stack",
|
||||
)
|
||||
except GatedRepoError:
|
||||
parser.error(
|
||||
|
|
|
@ -11,7 +11,7 @@ LLAMA_TOOLCHAIN_DIR=${LLAMA_TOOLCHAIN_DIR:-}
|
|||
TEST_PYPI_VERSION=${TEST_PYPI_VERSION:-}
|
||||
|
||||
if [ -n "$LLAMA_TOOLCHAIN_DIR" ]; then
|
||||
echo "Using llama-toolchain-dir=$LLAMA_TOOLCHAIN_DIR"
|
||||
echo "Using llama-stack-dir=$LLAMA_TOOLCHAIN_DIR"
|
||||
fi
|
||||
if [ -n "$LLAMA_MODELS_DIR" ]; then
|
||||
echo "Using llama-models-dir=$LLAMA_MODELS_DIR"
|
||||
|
@ -78,9 +78,9 @@ ensure_conda_env_python310() {
|
|||
if [ -n "$TEST_PYPI_VERSION" ]; then
|
||||
# these packages are damaged in test-pypi, so install them first
|
||||
pip install fastapi libcst
|
||||
pip install --extra-index-url https://test.pypi.org/simple/ llama-models==$TEST_PYPI_VERSION llama-toolchain==$TEST_PYPI_VERSION $pip_dependencies
|
||||
pip install --extra-index-url https://test.pypi.org/simple/ llama-models==$TEST_PYPI_VERSION llama-stack==$TEST_PYPI_VERSION $pip_dependencies
|
||||
else
|
||||
# Re-installing llama-toolchain in the new conda environment
|
||||
# Re-installing llama-stack in the new conda environment
|
||||
if [ -n "$LLAMA_TOOLCHAIN_DIR" ]; then
|
||||
if [ ! -d "$LLAMA_TOOLCHAIN_DIR" ]; then
|
||||
printf "${RED}Warning: LLAMA_TOOLCHAIN_DIR is set but directory does not exist: $LLAMA_TOOLCHAIN_DIR${NC}\n" >&2
|
||||
|
@ -90,7 +90,7 @@ ensure_conda_env_python310() {
|
|||
printf "Installing from LLAMA_TOOLCHAIN_DIR: $LLAMA_TOOLCHAIN_DIR\n"
|
||||
pip install --no-cache-dir -e "$LLAMA_TOOLCHAIN_DIR"
|
||||
else
|
||||
pip install --no-cache-dir llama-toolchain
|
||||
pip install --no-cache-dir llama-stack
|
||||
fi
|
||||
|
||||
if [ -n "$LLAMA_MODELS_DIR" ]; then
|
||||
|
|
|
@ -55,7 +55,7 @@ RUN apt-get update && apt-get install -y \
|
|||
|
||||
EOF
|
||||
|
||||
toolchain_mount="/app/llama-toolchain-source"
|
||||
toolchain_mount="/app/llama-stack-source"
|
||||
models_mount="/app/llama-models-source"
|
||||
|
||||
if [ -n "$LLAMA_TOOLCHAIN_DIR" ]; then
|
||||
|
@ -65,7 +65,7 @@ if [ -n "$LLAMA_TOOLCHAIN_DIR" ]; then
|
|||
fi
|
||||
add_to_docker "RUN pip install $toolchain_mount"
|
||||
else
|
||||
add_to_docker "RUN pip install llama-toolchain"
|
||||
add_to_docker "RUN pip install llama-stack"
|
||||
fi
|
||||
|
||||
if [ -n "$LLAMA_MODELS_DIR" ]; then
|
||||
|
|
|
@ -17,7 +17,7 @@ from llama_stack.apis.telemetry import Telemetry
|
|||
from .datatypes import Api, ApiEndpoint, ProviderSpec, remote_provider_spec
|
||||
|
||||
# These are the dependencies needed by the distribution server.
|
||||
# `llama-toolchain` is automatically installed by the installation script.
|
||||
# `llama-stack` is automatically installed by the installation script.
|
||||
SERVER_DEPENDENCIES = [
|
||||
"fastapi",
|
||||
"uvicorn",
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue