llama-stack-mirror/redhat-distribution/Containerfile

48 lines
1.4 KiB
Docker

# WARNING: This file is auto-generated. Do not modify it manually.
# Generated by: redhat-distribution/build.py
FROM registry.access.redhat.com/ubi9/python-312@sha256:95ec8d3ee9f875da011639213fd254256c29bc58861ac0b11f290a291fa04435
WORKDIR /opt/app-root
RUN pip install sqlalchemy # somehow sqlalchemy[asyncio] is not sufficient
RUN pip install \
aiosqlite \
autoevals \
chardet \
datasets \
fastapi \
fire \
httpx \
kubernetes \
llama_stack_provider_lmeval==0.1.7 \
llama_stack_provider_trustyai_fms==0.1.2 \
matplotlib \
mcp \
nltk \
numpy \
openai \
opentelemetry-exporter-otlp-proto-http \
opentelemetry-sdk \
pandas \
pillow \
psycopg2-binary \
pymilvus>=2.4.10 \
pymongo \
pypdf \
redis \
requests \
scikit-learn \
scipy \
sentencepiece \
sqlalchemy[asyncio] \
tqdm \
transformers \
uvicorn
RUN pip install --index-url https://download.pytorch.org/whl/cpu torch torchvision
RUN pip install --no-deps sentence-transformers
RUN pip install --no-cache llama-stack==0.2.14
RUN mkdir -p ${HOME}/.llama/providers.d ${HOME}/.cache
COPY redhat-distribution/run.yaml ${APP_ROOT}/run.yaml
COPY redhat-distribution/providers.d/ ${HOME}/.llama/providers.d/
ENTRYPOINT ["python", "-m", "llama_stack.distribution.server.server", "--config", "/opt/app-root/run.yaml"]