llama-stack-mirror/llama_stack/providers/impls
Ashwin Bharambe 70d59b0f5d Make vllm inference better
Tests still don't pass completely (some hang) so I think there are some
potential threading issues maybe
2024-10-24 22:52:47 -07:00
..
ios/inference Update iOS inference instructions for new quantization 2024-10-24 14:47:27 -04:00
meta_reference Fix for get_agents_session (#300) 2024-10-24 18:36:27 -07:00
vllm Make vllm inference better 2024-10-24 22:52:47 -07:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00