mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-14 12:12:45 +00:00
Improve VLLM model discovery error handling
• Add comprehensive error handling in check_model_availability method • Provide helpful error messages with actionable solutions for 404 errors • Warn when API token is set but model discovery is disabled
This commit is contained in:
parent
e9214f9004
commit
e28bc93635
15 changed files with 69 additions and 50 deletions
|
|
@ -31,7 +31,7 @@ providers:
|
|||
max_tokens: ${env.VLLM_MAX_TOKENS:=4096}
|
||||
api_token: ${env.VLLM_API_TOKEN:=fake}
|
||||
tls_verify: ${env.VLLM_TLS_VERIFY:=true}
|
||||
allow_listing_models: ${env.VLLM_ALLOW_LISTING_MODELS:=true}
|
||||
enable_model_discovery: ${env.VLLM_ENABLE_MODEL_DISCOVERY:=true}
|
||||
- provider_id: ${env.TGI_URL:+tgi}
|
||||
provider_type: remote::tgi
|
||||
config:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue