llama-stack-mirror/llama_stack/providers/impls
Ashwin Bharambe be3adb0964 Make vllm inference better
Tests still don't pass completely (some hang) so I think there are some
potential threading issues maybe
2024-10-25 12:03:42 -07:00
..
ios/inference Update iOS inference instructions for new quantization 2024-10-24 14:47:27 -04:00
meta_reference Fix for get_agents_session (#300) 2024-10-24 18:36:27 -07:00
vllm Make vllm inference better 2024-10-25 12:03:42 -07:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00