llama-stack-mirror/llama_stack/templates
Ashwin Bharambe 63e6acd0c3
feat: add (openai, anthropic, gemini) providers via litellm (#1267)
# What does this PR do?

This PR introduces more non-llama model support to llama stack.
Providers introduced: openai, anthropic and gemini. All of these
providers use essentially the same piece of code -- the implementation
works via the `litellm` library.

We will expose only specific models for providers we enable making sure
they all work well and pass tests. This setup (instead of automatically
enabling _all_ providers and models allowed by LiteLLM) ensures we can
also perform any needed prompt tuning on a per-model basis as needed
(just like we do it for llama models.)

## Test Plan

```bash
#!/bin/bash

args=("$@")
for model in openai/gpt-4o anthropic/claude-3-5-sonnet-latest gemini/gemini-1.5-flash; do
    LLAMA_STACK_CONFIG=dev pytest -s -v tests/client-sdk/inference/test_text_inference.py \
        --embedding-model=all-MiniLM-L6-v2 \
        --vision-inference-model="" \
        --inference-model=$model "${args[@]}"
done
```
2025-02-25 22:07:33 -08:00
..
bedrock ModelAlias -> ProviderModelEntry 2025-02-20 14:02:36 -08:00
cerebras chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
ci-tests feat: add (openai, anthropic, gemini) providers via litellm (#1267) 2025-02-25 22:07:33 -08:00
dell chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
dev feat: add (openai, anthropic, gemini) providers via litellm (#1267) 2025-02-25 22:07:33 -08:00
experimental-post-training fix!: update eval-tasks -> benchmarks (#1032) 2025-02-13 16:40:58 -08:00
fireworks test: add a ci-tests distro template for running e2e tests (#1237) 2025-02-24 14:43:21 -08:00
groq feat: Add Groq distribution template (#1173) 2025-02-25 14:16:56 -08:00
hf-endpoint fix!: update eval-tasks -> benchmarks (#1032) 2025-02-13 16:40:58 -08:00
hf-serverless chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
meta-reference-gpu fix!: update eval-tasks -> benchmarks (#1032) 2025-02-13 16:40:58 -08:00
meta-reference-quantized-gpu fix!: update eval-tasks -> benchmarks (#1032) 2025-02-13 16:40:58 -08:00
nvidia test(client-sdk): Update embedding test types to use latest imports (#1203) 2025-02-21 08:09:17 -08:00
ollama chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
passthrough feat: inference passthrough provider (#1166) 2025-02-19 21:47:00 -08:00
remote-vllm chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
sambanova ModelAlias -> ProviderModelEntry 2025-02-20 14:02:36 -08:00
tgi chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
together chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
vllm-gpu chore: move embedding deps to RAG tool where they are needed (#1210) 2025-02-21 11:33:41 -08:00
__init__.py Auto-generate distro yamls + docs (#468) 2024-11-18 14:57:06 -08:00
template.py build: format codebase imports using ruff linter (#1028) 2025-02-13 10:06:21 -08:00