mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-16 21:52:36 +00:00
fix vllm registry
This commit is contained in:
parent
a019011326
commit
e272f8aa62
3 changed files with 10 additions and 4 deletions
|
|
@ -57,7 +57,7 @@ class ModelRegistryHelper(ModelsProtocolPrivate):
|
|||
if provider_model_id in self.provider_id_to_llama_model_map:
|
||||
return self.provider_id_to_llama_model_map[provider_model_id]
|
||||
else:
|
||||
None
|
||||
return None
|
||||
|
||||
async def register_model(self, model: Model) -> Model:
|
||||
model.provider_resource_id = self.get_provider_model_id(
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue