llama-stack-mirror/llama_stack/providers/inline/inference
Matthew Farrellee f6d1867bf5 chore: remove batch-related APIs
APIs removed:
 - POST /v1/batch-inference/completion
 - POST /v1/batch-inference/chat-completion
 - POST /v1/inference/batch-completion
 - POST /v1/inference/batch-chat-completion

note -
 - batch-completion & batch-chat-completion were only implemented for inference=inline::meta-reference
 - batch-inference were not implemented
2025-08-26 19:18:16 -04:00
..
meta_reference chore: remove batch-related APIs 2025-08-26 19:18:16 -04:00
sentence_transformers chore: indicate to mypy that InferenceProvider.batch_completion/batch_chat_completion is concrete (#3239) 2025-08-22 14:17:30 -07:00
__init__.py precommit 2024-11-08 17:58:58 -08:00