llama-stack/llama_stack/providers/inline/inference/vllm
Dinesh Yeduguru fdff24e77a
Inference to use provider resource id to register and validate (#428)
This PR changes the way model id gets translated to the final model name
that gets passed through the provider.
Major changes include:
1) Providers are responsible for registering an object and as part of
the registration returning the object with the correct provider specific
name of the model provider_resource_id
2) To help with the common look ups different names a new ModelLookup
class is created.



Tested all inference providers including together, fireworks, vllm,
ollama, meta reference and bedrock
2024-11-12 20:02:00 -08:00
..
__init__.py Add provider deprecation support; change directory structure (#397) 2024-11-07 13:04:53 -08:00
config.py precommit 2024-11-08 17:58:58 -08:00
vllm.py Inference to use provider resource id to register and validate (#428) 2024-11-12 20:02:00 -08:00