llama-stack-mirror/docs/openapi_generator
Matthew Farrellee f6d1867bf5 chore: remove batch-related APIs
APIs removed:
 - POST /v1/batch-inference/completion
 - POST /v1/batch-inference/chat-completion
 - POST /v1/inference/batch-completion
 - POST /v1/inference/batch-chat-completion

note -
 - batch-completion & batch-chat-completion were only implemented for inference=inline::meta-reference
 - batch-inference were not implemented
2025-08-26 19:18:16 -04:00
..
pyopenapi chore: remove batch-related APIs 2025-08-26 19:18:16 -04:00
generate.py chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
README.md chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
run_openapi_generator.sh chore: remove straggler references to llama-models (#1345) 2025-03-01 14:26:03 -08:00

The RFC Specification (OpenAPI format) is generated from the set of API endpoints located in llama_stack.core/server/endpoints.py using the generate.py utility.