forked from phoenix-oss/llama-stack-mirror
update requiements
This commit is contained in:
parent
51816af52e
commit
a77b554bcf
2 changed files with 4 additions and 12 deletions
|
@ -1,12 +1,7 @@
|
||||||
#!/usr/bin/env bash
|
#!/usr/bin/env bash
|
||||||
|
|
||||||
export INFERENCE_MODEL="inference-llama4-maverick"
|
|
||||||
export EMBEDDING_MODEL="inference-bge-m3"
|
|
||||||
export EMBEDDING_DIMENSION="1024"
|
|
||||||
export LLAMA_STACK_PORT=8321
|
export LLAMA_STACK_PORT=8321
|
||||||
export OPENAI_BASE_URL=https://maas.ai-2.kvant.cloud/v1
|
# VLLM_API_TOKEN= env file
|
||||||
# OPENAI_API_KEY= env file
|
|
||||||
export VLLM_MAX_TOKENS=125000
|
|
||||||
# KEYCLOAK_CLIENT_SECRET= env file
|
# KEYCLOAK_CLIENT_SECRET= env file
|
||||||
|
|
||||||
|
|
||||||
|
@ -19,7 +14,4 @@ docker run -it \
|
||||||
distribution-kvant:dev \
|
distribution-kvant:dev \
|
||||||
-m llama_stack.distribution.server.server --config /root/.llama/config.yaml \
|
-m llama_stack.distribution.server.server --config /root/.llama/config.yaml \
|
||||||
--port $LLAMA_STACK_PORT \
|
--port $LLAMA_STACK_PORT \
|
||||||
--env VLLM_URL=$OPENAI_BASE_URL \
|
|
||||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
|
||||||
--env EMBEDDING_MODEL=$EMBEDDING_MODEL \
|
|
||||||
--env EMBEDDING_DIMENSION=$EMBEDDING_DIMENSION \
|
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
llama-stack>=0.2.1
|
llama-stack>=0.2.9
|
||||||
llama-stack-client>=0.2.1
|
llama-stack-client>=0.2.9
|
||||||
pandas
|
pandas
|
||||||
streamlit
|
streamlit
|
||||||
streamlit-option-menu
|
streamlit-option-menu
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue