mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-16 14:57:20 +00:00
add inline-vllm details, fix things
This commit is contained in:
parent
02c66b49fc
commit
38cdbdec5a
12 changed files with 142 additions and 101 deletions
|
@ -45,7 +45,7 @@ def available_providers() -> List[ProviderSpec]:
|
|||
),
|
||||
InlineProviderSpec(
|
||||
api=Api.inference,
|
||||
provider_type="vllm",
|
||||
provider_type="inline::vllm",
|
||||
pip_packages=[
|
||||
"vllm",
|
||||
],
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue