mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-21 19:52:27 +00:00
1.6 KiB
1.6 KiB
Portkey Distribution
The llamastack/distribution-{{ name }} distribution consists of the following provider configurations.
{{ providers_table }}
{% if run_config_env_vars %}
Environment Variables
The following environment variables can be configured:
{% for var, (default_value, description) in run_config_env_vars.items() %}
{{ var }}: {{ description }} (default:{{ default_value }}) {% endfor %} {% endif %}
{% if default_models %}
Models
The following models are available by default:
{% for model in default_models %}
{{ model.model_id }} ({{ model.provider_model_id }}){% endfor %} {% endif %}
Prerequisite: API Keys
Make sure you have access to a Portkey API Key and Virtual Key or Config ID. You can get these by visiting app.portkey.ai.
Running Llama Stack with Portkey
You can do this via Conda (build code) or Docker which has a pre-built image.
Via Docker
This method allows you to get started quickly without having to build the distribution code.
LLAMA_STACK_PORT=5001
docker run \
-it \
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
-v ./run.yaml:/root/my-run.yaml \
llamastack/distribution-{{ name }} \
--yaml-config /root/my-run.yaml \
--port $LLAMA_STACK_PORT \
--env PORTKEY_API_KEY=$PORTKEY_API_KEY
--env PORTKEY_VIRTUAL_KEY=$PORTKEY_VIRTUAL_KEY
--env PORTKEY_CONFIG_ID=$PORTKEY_CONFIG_ID
Via Conda
llama stack build --template portkey --image-type conda
llama stack run ./run.yaml \
--port 5001 \
--env PORTKEY_API_KEY=$PORTKEY_API_KEY
--env PORTKEY_VIRTUAL_KEY=$PORTKEY_VIRTUAL_KEY
--env PORTKEY_CONFIG_ID=$PORTKEY_CONFIG_ID