llama-stack-mirror/llama_stack/providers/utils/inference
Matthew Farrellee 3d119a86d4
chore: indicate to mypy that InferenceProvider.batch_completion/batch_chat_completion is concrete (#3239)
# What does this PR do?

closes https://github.com/llamastack/llama-stack/issues/3236

mypy considered our default implementations (raise NotImplementedError)
to be trivial. the result was we implemented the same stubs in
providers.

this change puts enough into the default impls so mypy considers them
non-trivial. this allows us to remove the duplicate implementations.
2025-08-22 14:17:30 -07:00
..
__init__.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
embedding_mixin.py refactor(logging): rename llama_stack logger categories (#3065) 2025-08-21 17:31:04 -07:00
inference_store.py chore(rename): move llama_stack.distribution to llama_stack.core (#2975) 2025-07-30 23:30:53 -07:00
litellm_openai_mixin.py chore: indicate to mypy that InferenceProvider.batch_completion/batch_chat_completion is concrete (#3239) 2025-08-22 14:17:30 -07:00
model_registry.py refactor(logging): rename llama_stack logger categories (#3065) 2025-08-21 17:31:04 -07:00
openai_compat.py refactor(logging): rename llama_stack logger categories (#3065) 2025-08-21 17:31:04 -07:00
openai_mixin.py refactor(logging): rename llama_stack logger categories (#3065) 2025-08-21 17:31:04 -07:00
prompt_adapter.py refactor(logging): rename llama_stack logger categories (#3065) 2025-08-21 17:31:04 -07:00