llama-stack/llama_stack/providers/impls/meta_reference/inference
Ashwin Bharambe cde9bc1388
Enable vision models for (Together, Fireworks, Meta-Reference, Ollama) (#376)
* Enable vision models for Together and Fireworks

* Works with ollama 0.4.0 pre-release with the vision model

* localize media for meta_reference inference

* Fix
2024-11-05 16:22:33 -08:00
..
quantization Added hadamard transform for spinquant (#326) 2024-10-25 12:58:48 -07:00
__init__.py Split off meta-reference-quantized provider 2024-10-10 16:03:19 -07:00
config.py Allow overridding checkpoint_dir via config 2024-10-18 14:28:06 -07:00
generation.py Avoid warnings from pydantic for overriding schema 2024-10-28 21:39:48 -07:00
inference.py Enable vision models for (Together, Fireworks, Meta-Reference, Ollama) (#376) 2024-11-05 16:22:33 -08:00
model_parallel.py Make all methods async def again; add completion() for meta-reference (#270) 2024-10-18 20:50:59 -07:00
parallel_utils.py Make all methods async def again; add completion() for meta-reference (#270) 2024-10-18 20:50:59 -07:00