llama-stack-mirror/llama_stack/providers/inline/inference/meta_reference
Botao Chen d021983b0e refine
2024-12-17 20:43:20 -08:00
..
quantization use logging instead of prints (#499) 2024-11-21 11:32:53 -08:00
__init__.py temp commit 2024-12-16 19:04:47 -08:00
config.py refine 2024-12-17 13:38:19 -08:00
generation.py Merge branch 'main' into inference_refactor 2024-12-17 20:10:23 -08:00
inference.py refine 2024-12-17 20:43:20 -08:00
model_parallel.py refine 2024-12-17 13:38:19 -08:00
parallel_utils.py temp commit 2024-12-16 21:43:30 -08:00