llama-stack-mirror/llama_stack/providers/remote/inference/vllm
Akram Ben Aissi 2b54b57145 feat: implement graceful model discovery for vLLM provider
- Attempt model discovery first for backward compatibility
- If discovery fails and refresh_models=false, continue without error
- If discovery fails and refresh_models=true, fail hard with ValueError
- Supports dynamic token authentication scenarios

Fixes OAuth authentication issues when vLLM service requires dynamic tokens
2025-10-03 21:32:15 +02:00
..
__init__.py feat: Add dynamic authentication token forwarding support for vLLM (#3388) 2025-09-18 11:13:55 +02:00
config.py chore: use remoteinferenceproviderconfig for remote inference providers (#3668) 2025-10-03 08:48:42 -07:00
vllm.py feat: implement graceful model discovery for vLLM provider 2025-10-03 21:32:15 +02:00