mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 18:29:50 +00:00
48 lines
1.4 KiB
Docker
48 lines
1.4 KiB
Docker
# WARNING: This file is auto-generated. Do not modify it manually.
|
|
# Generated by: redhat-distribution/build.py
|
|
|
|
FROM registry.access.redhat.com/ubi9/python-312@sha256:95ec8d3ee9f875da011639213fd254256c29bc58861ac0b11f290a291fa04435
|
|
WORKDIR /opt/app-root
|
|
|
|
RUN pip install sqlalchemy # somehow sqlalchemy[asyncio] is not sufficient
|
|
RUN pip install \
|
|
aiosqlite \
|
|
autoevals \
|
|
chardet \
|
|
datasets \
|
|
fastapi \
|
|
fire \
|
|
httpx \
|
|
kubernetes \
|
|
llama_stack_provider_lmeval==0.1.7 \
|
|
llama_stack_provider_trustyai_fms==0.1.2 \
|
|
matplotlib \
|
|
mcp \
|
|
nltk \
|
|
numpy \
|
|
openai \
|
|
opentelemetry-exporter-otlp-proto-http \
|
|
opentelemetry-sdk \
|
|
pandas \
|
|
pillow \
|
|
psycopg2-binary \
|
|
pymilvus>=2.4.10 \
|
|
pymongo \
|
|
pypdf \
|
|
redis \
|
|
requests \
|
|
scikit-learn \
|
|
scipy \
|
|
sentencepiece \
|
|
sqlalchemy[asyncio] \
|
|
tqdm \
|
|
transformers \
|
|
uvicorn
|
|
RUN pip install --index-url https://download.pytorch.org/whl/cpu torch torchvision
|
|
RUN pip install --no-deps sentence-transformers
|
|
RUN pip install --no-cache llama-stack==0.2.14
|
|
RUN mkdir -p ${HOME}/.llama/providers.d ${HOME}/.cache
|
|
COPY redhat-distribution/run.yaml ${APP_ROOT}/run.yaml
|
|
COPY redhat-distribution/providers.d/ ${HOME}/.llama/providers.d/
|
|
|
|
ENTRYPOINT ["python", "-m", "llama_stack.distribution.server.server", "--config", "/opt/app-root/run.yaml"]
|