mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-30 11:50:14 +00:00
This fixes the issue when using `llama stack run` by correctly specifying entrypoint: ``` LLAMA_STACK_DIR=. llama stack run /home/yutang/.llama/distributions/llamastack-vllm/vllm-run.yaml Using config file: /home/yutang/.llama/distributions/llamastack-vllm/vllm-run.yaml + command -v selinuxenabled + selinuxenabled + DOCKER_OPTS=' --security-opt label=disable' + mounts= + '[' -n . ']' ++ readlink -f . + mounts=' -v /home/yutang/repos/llama-stack:/app/llama-stack-source' + '[' -n '' ']' + version_tag=latest + '[' -n '' ']' + '[' -n . ']' + version_tag=dev + podman run --security-opt label=disable -it -p 5000:5000 -v /home/yutang/.llama/distributions/llamastack-vllm/vllm-run.yaml:/app/config.yaml -v /home/yutang/repos/llama-stack:/app/llama-stack-source localhost/distribution-vllm:dev python -m llama_stack.distribution.server.server --yaml-config /app/config.yaml --port 5000 usage: server.py [-h] [--yaml-config YAML_CONFIG] [--template TEMPLATE] [--port PORT] [--disable-ipv6] [--env ENV] server.py: error: unrecognized arguments: python -m llama_stack.distribution.server.server ++ error_handler 88 ++ echo 'Error occurred in script at line: 88' Error occurred in script at line: 88 ++ exit 1 ``` --------- Signed-off-by: Yuan Tang <terrytangyuan@gmail.com> |
||
---|---|---|
.. | ||
routers | ||
server | ||
store | ||
tests | ||
ui | ||
utils | ||
__init__.py | ||
build.py | ||
build_conda_env.sh | ||
build_container.sh | ||
build_venv.sh | ||
client.py | ||
common.sh | ||
configure.py | ||
configure_container.sh | ||
datatypes.py | ||
distribution.py | ||
inspect.py | ||
library_client.py | ||
request_headers.py | ||
resolver.py | ||
stack.py | ||
start_conda_env.sh | ||
start_container.sh |