llama-stack-mirror/llama_stack/providers/inline
Dinesh Yeduguru 96e158eaac
Make embedding generation go through inference (#606)
This PR does the following:
1) adds the ability to generate embeddings in all supported inference
providers.
2) Moves all the memory providers to use the inference API and improved
the memory tests to setup the inference stack correctly and use the
embedding models

This is a merge from #589 and #598
2024-12-12 11:47:50 -08:00
..
agents add tracing back to the lib cli (#595) 2024-12-11 08:44:20 -08:00
datasetio Telemetry API redesign (#525) 2024-12-04 11:22:45 -08:00
eval Add ability to query and export spans to dataset (#574) 2024-12-05 21:07:30 -08:00
inference Make embedding generation go through inference (#606) 2024-12-12 11:47:50 -08:00
ios/inference impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00
memory Make embedding generation go through inference (#606) 2024-12-12 11:47:50 -08:00
meta_reference Telemetry API redesign (#525) 2024-12-04 11:22:45 -08:00
safety use logging instead of prints (#499) 2024-11-21 11:32:53 -08:00
scoring [/scoring] add ability to define aggregation functions for scoring functions & refactors (#597) 2024-12-11 10:03:42 -08:00
telemetry add tracing back to the lib cli (#595) 2024-12-11 08:44:20 -08:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00