llama-stack-mirror/llama_stack/providers/remote/inference
Ashwin Bharambe cdcbeb005b
chore: remove llama_models.llama3.api imports from providers (#1107)
There should be a choke-point for llama3.api imports -- this is the
prompt adapter. Creating a ChatFormat() object on demand is inexpensive.
The underlying Tokenizer is a singleton anyway.
2025-02-19 19:01:29 -08:00
..
bedrock chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
cerebras chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
databricks chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
fireworks chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
groq chore: move all Llama Stack types from llama-models to llama-stack (#1098) 2025-02-14 09:10:59 -08:00
nvidia fix: Get distro_codegen.py working with default deps and enabled in pre-commit hooks (#1123) 2025-02-19 18:39:20 -08:00
ollama chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
runpod chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
sambanova chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
sample build: format codebase imports using ruff linter (#1028) 2025-02-13 10:06:21 -08:00
tgi chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
together chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
vllm chore: remove llama_models.llama3.api imports from providers (#1107) 2025-02-19 19:01:29 -08:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00