mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 20:14:13 +00:00
instantiate inference models
This commit is contained in:
parent
d2ec822b12
commit
7071c46422
6 changed files with 40 additions and 20 deletions
|
@ -321,12 +321,10 @@ async def resolve_impls(
|
|||
inner_specs=inner_specs,
|
||||
)
|
||||
|
||||
for k, v in specs.items():
|
||||
cprint(k, "blue")
|
||||
cprint(v, "blue")
|
||||
|
||||
sorted_specs = topological_sort(specs.values())
|
||||
|
||||
cprint(f"sorted_specs={sorted_specs}", "red")
|
||||
|
||||
impls = {}
|
||||
for spec in sorted_specs:
|
||||
api = spec.api
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue