llama-stack-mirror/llama_stack/providers/utils/inference
Jaideep Rao 66412ab12b convert blocking calls to async
Signed-off-by: Jaideep Rao <jrao@redhat.com>
2025-03-20 13:07:23 -04:00
..
__init__.py chore: move all Llama Stack types from llama-models to llama-stack (#1098) 2025-02-14 09:10:59 -08:00
embedding_mixin.py convert blocking calls to async 2025-03-20 13:07:23 -04:00
litellm_openai_mixin.py convert blocking calls to async 2025-03-20 13:07:23 -04:00
model_registry.py feat(providers): support non-llama models for inference providers (#1200) 2025-02-21 13:21:28 -08:00
openai_compat.py fix: Updating ToolCall.arguments to allow for json strings that can be decoded on client side (#1685) 2025-03-19 10:36:19 -07:00
prompt_adapter.py feat(logging): implement category-based logging (#1362) 2025-03-07 11:34:30 -08:00