llama-stack-mirror/llama_stack/providers/inline
2025-10-03 12:32:26 -07:00
..
agents fix: responses <> chat completion input conversion (#3645) 2025-10-02 16:01:08 -07:00
batches feat(batches, completions): add /v1/completions support to /v1/batches (#3309) 2025-09-05 11:59:57 -07:00
datasetio chore(misc): make tests and starter faster (#3042) 2025-08-05 14:55:05 -07:00
eval feat: update eval runner to use openai endpoints (#3588) 2025-09-29 13:13:53 -07:00
files/localfs fix(expires_after): make sure multipart/form-data is properly parsed (#3612) 2025-09-30 16:14:03 -04:00
inference chore: remove deprecated inference.chat_completion implementations (#3654) 2025-10-03 07:55:34 -04:00
ios/inference feat(tools)!: substantial clean up of "Tool" related datatypes (#3627) 2025-10-02 15:12:03 -07:00
post_training chore(pre-commit): add pre-commit hook to enforce llama_stack logger usage (#3061) 2025-08-20 07:15:35 -04:00
safety feat: use /v1/chat/completions for safety model inference (#3591) 2025-09-30 11:01:44 -07:00
scoring chore: use openai_chat_completion for llm as a judge scoring (#3635) 2025-10-01 09:44:31 -04:00
telemetry chore: Remove debug logging from telemetry adapter (#3643) 2025-10-01 15:16:23 -07:00
tool_runtime feat(tools)!: substantial clean up of "Tool" related datatypes (#3627) 2025-10-02 15:12:03 -07:00
vector_io add __init__ to the mixin 2025-10-03 12:32:26 -07:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00