llama-stack/llama_stack/providers/utils/inference
Sébastien Han c223b1862b
fix: resolve type hint issues and import dependencies (#1176)
# What does this PR do?

- Fixed type hinting and missing imports across multiple modules.
- Improved compatibility by using `TYPE_CHECKING` for conditional
imports.
- Updated `pyproject.toml` to enforce stricter linting.

Signed-off-by: Sébastien Han <seb@redhat.com>

Signed-off-by: Sébastien Han <seb@redhat.com>
2025-02-25 11:06:47 -08:00
..
__init__.py chore: move all Llama Stack types from llama-models to llama-stack (#1098) 2025-02-14 09:10:59 -08:00
embedding_mixin.py fix: resolve type hint issues and import dependencies (#1176) 2025-02-25 11:06:47 -08:00
model_registry.py feat(providers): support non-llama models for inference providers (#1200) 2025-02-21 13:21:28 -08:00
openai_compat.py chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
prompt_adapter.py fix: set default tool_prompt_format in inference api (#1214) 2025-02-24 12:38:37 -08:00