mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-27 06:28:50 +00:00
- Add new Vertex AI remote inference provider with litellm integration - Support for Gemini models through Google Cloud Vertex AI platform - Uses Google Cloud Application Default Credentials (ADC) for authentication - Added VertexAI models: gemini-2.5-flash, gemini-2.5-pro, gemini-2.0-flash. - Updated provider registry to include vertexai provider - Updated starter template to support Vertex AI configuration - Added comprehensive documentation and sample configuration Signed-off-by: Eran Cohen <eranco@redhat.com>
27 lines
No EOL
1.1 KiB
Markdown
27 lines
No EOL
1.1 KiB
Markdown
# Inference Providers
|
|
|
|
This section contains documentation for all available providers for the **inference** API.
|
|
|
|
- [inline::meta-reference](inline_meta-reference.md)
|
|
- [inline::sentence-transformers](inline_sentence-transformers.md)
|
|
- [remote::anthropic](remote_anthropic.md)
|
|
- [remote::bedrock](remote_bedrock.md)
|
|
- [remote::cerebras](remote_cerebras.md)
|
|
- [remote::databricks](remote_databricks.md)
|
|
- [remote::fireworks](remote_fireworks.md)
|
|
- [remote::gemini](remote_gemini.md)
|
|
- [remote::groq](remote_groq.md)
|
|
- [remote::hf::endpoint](remote_hf_endpoint.md)
|
|
- [remote::hf::serverless](remote_hf_serverless.md)
|
|
- [remote::llama-openai-compat](remote_llama-openai-compat.md)
|
|
- [remote::nvidia](remote_nvidia.md)
|
|
- [remote::ollama](remote_ollama.md)
|
|
- [remote::openai](remote_openai.md)
|
|
- [remote::passthrough](remote_passthrough.md)
|
|
- [remote::runpod](remote_runpod.md)
|
|
- [remote::sambanova](remote_sambanova.md)
|
|
- [remote::tgi](remote_tgi.md)
|
|
- [remote::together](remote_together.md)
|
|
- [remote::vertexai](remote_vertexai.md)
|
|
- [remote::vllm](remote_vllm.md)
|
|
- [remote::watsonx](remote_watsonx.md) |