llama-stack-mirror/llama_stack/providers/impls
Dalton Flanagan 7a8aa775e5
JSON serialization for parallel processing queue (#232)
* send/recv pydantic json over socket

* fixup

* address feedback

* bidirectional wrapper

* second round of feedback
2024-10-09 17:24:12 -04:00
..
ios/inference Add function for stopping inference (#224) 2024-10-09 10:50:19 -04:00
meta_reference JSON serialization for parallel processing queue (#232) 2024-10-09 17:24:12 -04:00
vllm Inline vLLM inference provider (#181) 2024-10-05 23:34:16 -07:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00