mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-29 15:23:51 +00:00
update compose file
This commit is contained in:
parent
542ffbee72
commit
dcac9e4874
1 changed files with 8 additions and 6 deletions
|
@ -26,7 +26,8 @@ services:
|
||||||
capabilities: [gpu]
|
capabilities: [gpu]
|
||||||
runtime: nvidia
|
runtime: nvidia
|
||||||
llamastack-local-cpu:
|
llamastack-local-cpu:
|
||||||
depends_on: ollama
|
depends_on:
|
||||||
|
- ollama
|
||||||
image: llamastack/llamastack-local-cpu
|
image: llamastack/llamastack-local-cpu
|
||||||
network_mode: "host"
|
network_mode: "host"
|
||||||
volumes:
|
volumes:
|
||||||
|
@ -37,10 +38,11 @@ services:
|
||||||
- "5000:5000"
|
- "5000:5000"
|
||||||
# Hack: wait for ollama server to start before starting docker
|
# Hack: wait for ollama server to start before starting docker
|
||||||
entrypoint: bash -c "sleep 60; python -m llama_stack.distribution.server.server --yaml_config /root/llamastack-run-ollama.yaml"
|
entrypoint: bash -c "sleep 60; python -m llama_stack.distribution.server.server --yaml_config /root/llamastack-run-ollama.yaml"
|
||||||
restart_policy:
|
deploy:
|
||||||
condition: on-failure
|
restart_policy:
|
||||||
delay: 3s
|
condition: on-failure
|
||||||
max_attempts: 5
|
delay: 3s
|
||||||
window: 60s
|
max_attempts: 5
|
||||||
|
window: 60s
|
||||||
volumes:
|
volumes:
|
||||||
ollama:
|
ollama:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue