forked from phoenix-oss/llama-stack-mirror
update requiements
This commit is contained in:
parent
51816af52e
commit
a77b554bcf
2 changed files with 4 additions and 12 deletions
|
@ -1,12 +1,7 @@
|
|||
#!/usr/bin/env bash
|
||||
|
||||
export INFERENCE_MODEL="inference-llama4-maverick"
|
||||
export EMBEDDING_MODEL="inference-bge-m3"
|
||||
export EMBEDDING_DIMENSION="1024"
|
||||
export LLAMA_STACK_PORT=8321
|
||||
export OPENAI_BASE_URL=https://maas.ai-2.kvant.cloud/v1
|
||||
# OPENAI_API_KEY= env file
|
||||
export VLLM_MAX_TOKENS=125000
|
||||
# VLLM_API_TOKEN= env file
|
||||
# KEYCLOAK_CLIENT_SECRET= env file
|
||||
|
||||
|
||||
|
@ -19,7 +14,4 @@ docker run -it \
|
|||
distribution-kvant:dev \
|
||||
-m llama_stack.distribution.server.server --config /root/.llama/config.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env VLLM_URL=$OPENAI_BASE_URL \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env EMBEDDING_MODEL=$EMBEDDING_MODEL \
|
||||
--env EMBEDDING_DIMENSION=$EMBEDDING_DIMENSION \
|
||||
|
||||
|
|
|
@ -1,5 +1,5 @@
|
|||
llama-stack>=0.2.1
|
||||
llama-stack-client>=0.2.1
|
||||
llama-stack>=0.2.9
|
||||
llama-stack-client>=0.2.9
|
||||
pandas
|
||||
streamlit
|
||||
streamlit-option-menu
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue