services: llamastack: image: llamastack/distribution-fireworks network_mode: "host" volumes: - ~/.llama:/root/.llama - ./run.yaml:/root/llamastack-run-fireworks.yaml ports: - "8321:8321" entrypoint: bash -c "python -m llama_stack.distribution.server.server --yaml_config /root/llamastack-run-fireworks.yaml" deploy: restart_policy: condition: on-failure delay: 3s max_attempts: 5 window: 60s