llama-stack-mirror/llama_stack/providers/impls
Ashwin Bharambe 4fa467731e Fix a bug in meta-reference inference when stream=False
Also introduce a gross hack (to cover grosser(?) hack) to ensure
non-stream requests don't send back responses in SSE format. Not sure
which of these hacks is grosser.
2024-10-08 17:23:02 -07:00
..
ios/inference Drop header from LocalInference.h 2024-09-25 11:27:37 -07:00
meta_reference Fix a bug in meta-reference inference when stream=False 2024-10-08 17:23:02 -07:00
vllm more fixes, plug shutdown handlers 2024-10-08 17:23:02 -07:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00