diff --git a/llama_stack/distribution/templates/docker/llamastack-local-cpu/Dockerfile b/llama_stack/distribution/templates/docker/llamastack-local-cpu/Dockerfile deleted file mode 100644 index e69de29bb..000000000 diff --git a/llama_stack/distribution/templates/docker/llamastack-local-gpu/Dockerfile b/llama_stack/distribution/templates/docker/llamastack-local-gpu/Dockerfile deleted file mode 100644 index 00d2fa140..000000000 --- a/llama_stack/distribution/templates/docker/llamastack-local-gpu/Dockerfile +++ /dev/null @@ -1,21 +0,0 @@ -FROM python:3.10-slim -WORKDIR /app - -RUN apt-get update && apt-get install -y iputils-ping net-tools iproute2 dnsutils telnet curl wget telnet procps psmisc lsof traceroute bubblewrap && rm -rf /var/lib/apt/lists/* - -RUN pip install /app/llama-stack-source -RUN pip uninstall -y llama-models -RUN pip install /app/llama-models-source - -RUN pip install fastapi fire httpx uvicorn accelerate blobfile fairscale fbgemm-gpu==0.8.0 torch torchvision transformers zmq codeshield transformers matplotlib pillow pandas scikit-learn aiosqlite psycopg2-binary redis blobfile chardet pypdf tqdm numpy scikit-learn scipy nltk sentencepiece transformers faiss-cpu -RUN pip install torch --index-url https://download.pytorch.org/whl/cpu -RUN pip install torch --index-url https://download.pytorch.org/whl/cpu -RUN pip install sentence-transformers --no-deps - -# This would be good in production but for debugging flexibility lets not add it right now -# We need a more solid production ready entrypoint.sh anyway -# -CMD ["python", "-m", "llama_stack.distribution.server.server", "./llamastack-run.yaml"] - -ADD tmp/configs/local-gpu-build.yaml ./llamastack-build.yaml -ADD tmp/configs/local-gpu-run.yaml ./llamastack-run.yaml