llama-stack-mirror/llama_stack/providers/registry
Ashwin Bharambe 0a3999a9a4
Use inference APIs for executing Llama Guard (#121)
We should use Inference APIs to execute Llama Guard instead of directly needing to use HuggingFace modeling related code. The actual inference consideration is handled by Inference.
2024-09-28 15:40:06 -07:00
..
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00
agents.py [API Updates] Model / shield / memory-bank routing + agent persistence + support for private headers (#92) 2024-09-23 14:22:22 -07:00
inference.py Make TGI adapter compatible with HF Inference API (#97) 2024-09-25 14:08:31 -07:00
memory.py Make the "all-remote" distribution lightweight in dependencies and size 2024-09-24 14:18:57 -07:00
safety.py Use inference APIs for executing Llama Guard (#121) 2024-09-28 15:40:06 -07:00
telemetry.py [API Updates] Model / shield / memory-bank routing + agent persistence + support for private headers (#92) 2024-09-23 14:22:22 -07:00