mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-14 04:42:41 +00:00
Improve VLLM model discovery error handling
• Add comprehensive error handling in check_model_availability method • Provide helpful error messages with actionable solutions for 404 errors • Warn when API token is set but model discovery is disabled
This commit is contained in:
parent
e9214f9004
commit
e28bc93635
15 changed files with 69 additions and 50 deletions
|
|
@ -100,7 +100,7 @@ class ModelRegistryHelper(ModelsProtocolPrivate):
|
|||
async def should_refresh_models(self) -> bool:
|
||||
return False
|
||||
|
||||
async def allow_listing_models(self) -> bool:
|
||||
async def enable_model_discovery(self) -> bool:
|
||||
return True
|
||||
|
||||
def get_provider_model_id(self, identifier: str) -> str | None:
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue