llama-stack/llama_stack/providers/impls/meta_reference/inference
Ashwin Bharambe 0a3999a9a4
Use inference APIs for executing Llama Guard (#121)
We should use Inference APIs to execute Llama Guard instead of directly needing to use HuggingFace modeling related code. The actual inference consideration is handled by Inference.
2024-09-28 15:40:06 -07:00
..
quantization Add a test runner and 2 very simple tests for agents 2024-09-19 12:22:48 -07:00
__init__.py API Updates (#73) 2024-09-17 19:51:35 -07:00
config.py Use inference APIs for executing Llama Guard (#121) 2024-09-28 15:40:06 -07:00
generation.py Use inference APIs for executing Llama Guard (#121) 2024-09-28 15:40:06 -07:00
inference.py Support for Llama3.2 models and Swift SDK (#98) 2024-09-25 10:29:58 -07:00
model_parallel.py API Updates (#73) 2024-09-17 19:51:35 -07:00
parallel_utils.py API Updates (#73) 2024-09-17 19:51:35 -07:00