llama-stack/llama_stack/providers/registry
Botao Chen 2b995c22eb
feat: inference passthrough provider (#1166)
##  What does this PR do?
In this PR, we implement a passthrough inference provider that works for
any endpoints that respect llama stack inference API definition.

## Test Plan
config some endpoint that respect llama stack inference API definition
and got the inference results successfully

<img width="1268" alt="Screenshot 2025-02-19 at 8 52 51 PM"
src="https://github.com/user-attachments/assets/447816e4-ea7a-4365-b90c-386dc7dcf4a1"
/>
2025-02-19 21:47:00 -08:00
..
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00
agents.py [memory refactor][2/n] Update faiss and make it pass tests (#830) 2025-01-22 10:02:15 -08:00
datasetio.py [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
eval.py [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
inference.py feat: inference passthrough provider (#1166) 2025-02-19 21:47:00 -08:00
post_training.py Pin torchtune pkg version (#791) 2025-01-16 16:31:13 -08:00
safety.py [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
scoring.py [remove import *] clean up import *'s (#689) 2024-12-27 15:45:44 -08:00
telemetry.py optional api dependencies (#793) 2025-01-17 15:26:53 -08:00
tool_runtime.py Move tool_runtime.memory -> tool_runtime.rag 2025-01-22 20:25:02 -08:00
vector_io.py fix: Update VectorIO config classes in registry (#1079) 2025-02-13 15:39:13 -08:00