llama-stack-mirror/llama_stack/providers/inline/inference
Ashwin Bharambe 540fc4d717
Fix Meta reference GPU implementation (#663)
By performing in-place mutations, we lost. Never in life do that.
2024-12-19 14:09:45 -08:00
..
meta_reference Fix Meta reference GPU implementation (#663) 2024-12-19 14:09:45 -08:00
sentence_transformers add embedding model by default to distribution templates (#617) 2024-12-13 12:48:00 -08:00
vllm Fix conversion to RawMessage everywhere 2024-12-17 14:00:43 -08:00
__init__.py precommit 2024-11-08 17:58:58 -08:00