Allow setting environment variables from llama stack run and fix ollama

This commit is contained in:
Ashwin Bharambe 2024-11-17 19:33:48 -08:00
parent a061f3f8c1
commit b1d119466e
19 changed files with 129 additions and 55 deletions

View file

@ -1,20 +1,20 @@
version: '2'
built_at: 2024-11-17 15:19:07.405618
built_at: 2024-11-17 19:33:00
image_name: ollama
docker_image: llamastack/distribution-ollama:test-0.0.52rc3
docker_image: null
conda_env: null
apis:
- telemetry
- agents
- memory
- inference
- agents
- safety
- inference
- telemetry
providers:
inference:
- provider_id: ollama
provider_type: remote::ollama
config:
port: ${env.OLLAMA_PORT}
url: ${env.OLLAMA_URL:http://localhost:11434}
memory:
- provider_id: faiss
provider_type: inline::faiss

View file

@ -1,20 +1,20 @@
version: '2'
built_at: 2024-11-17 15:19:07.395495
built_at: 2024-11-17 19:33:00
image_name: ollama
docker_image: llamastack/distribution-ollama:test-0.0.52rc3
docker_image: null
conda_env: null
apis:
- telemetry
- agents
- memory
- inference
- agents
- safety
- inference
- telemetry
providers:
inference:
- provider_id: ollama
provider_type: remote::ollama
config:
port: ${env.OLLAMA_PORT}
url: ${env.OLLAMA_URL:http://localhost:11434}
memory:
- provider_id: faiss
provider_type: inline::faiss

View file

@ -1,14 +1,14 @@
version: '2'
built_at: 2024-11-17 15:19:07.405727
built_at: 2024-11-17 19:33:00
image_name: remote-vllm
docker_image: llamastack/distribution-remote-vllm:test-0.0.52rc3
conda_env: null
apis:
- telemetry
- agents
- memory
- inference
- agents
- safety
- inference
- telemetry
providers:
inference:
- provider_id: vllm-inference

View file

@ -1,14 +1,14 @@
version: '2'
built_at: 2024-11-17 15:19:07.395327
built_at: 2024-11-17 19:33:00
image_name: remote-vllm
docker_image: llamastack/distribution-remote-vllm:test-0.0.52rc3
conda_env: null
apis:
- telemetry
- agents
- memory
- inference
- agents
- safety
- inference
- telemetry
providers:
inference:
- provider_id: vllm-inference

View file

@ -1,14 +1,14 @@
version: '2'
built_at: 2024-11-17 15:19:09.184709
built_at: 2024-11-17 19:33:00
image_name: tgi
docker_image: llamastack/distribution-tgi:test-0.0.52rc3
conda_env: null
apis:
- telemetry
- agents
- memory
- inference
- agents
- safety
- inference
- telemetry
providers:
inference:
- provider_id: tgi-inference

View file

@ -1,14 +1,14 @@
version: '2'
built_at: 2024-11-17 15:19:09.156305
built_at: 2024-11-17 19:33:00
image_name: tgi
docker_image: llamastack/distribution-tgi:test-0.0.52rc3
conda_env: null
apis:
- telemetry
- agents
- memory
- inference
- agents
- safety
- inference
- telemetry
providers:
inference:
- provider_id: tgi-inference