mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-26 09:15:40 +00:00
docs: provider and distro codegen migration (#3531)
# What does this PR do? <!-- Provide a short summary of what this PR does and why. Link to relevant issues if applicable. --> <!-- If resolving an issue, uncomment and update the line below --> <!-- Closes #[issue-number] --> - Updates provider and distro codegen to handle the new format - Migrates provider and distro files to the new format ## Test Plan - Manual testing <!-- Describe the tests you ran to verify your changes with result summaries. *Provide clear instructions so the plan can be easily re-executed.* -->
This commit is contained in:
parent
45da31801c
commit
d23865757f
103 changed files with 1796 additions and 423 deletions
48
docs/docs/providers/inference/index.mdx
Normal file
48
docs/docs/providers/inference/index.mdx
Normal file
|
|
@ -0,0 +1,48 @@
|
|||
---
|
||||
description: "Llama Stack Inference API for generating completions, chat completions, and embeddings.
|
||||
|
||||
This API provides the raw interface to the underlying models. Two kinds of models are supported:
|
||||
- LLM models: these models generate \"raw\" and \"chat\" (conversational) completions.
|
||||
- Embedding models: these models generate embeddings to be used for semantic search."
|
||||
sidebar_label: Inference
|
||||
title: Inference
|
||||
---
|
||||
|
||||
# Inference
|
||||
|
||||
## Overview
|
||||
|
||||
Llama Stack Inference API for generating completions, chat completions, and embeddings.
|
||||
|
||||
This API provides the raw interface to the underlying models. Two kinds of models are supported:
|
||||
- LLM models: these models generate "raw" and "chat" (conversational) completions.
|
||||
- Embedding models: these models generate embeddings to be used for semantic search.
|
||||
|
||||
This section contains documentation for all available providers for the **inference** API.
|
||||
|
||||
## Providers
|
||||
|
||||
- [Meta-Reference](./inline_meta-reference)
|
||||
- [Sentence-Transformers](./inline_sentence-transformers)
|
||||
- [Remote - Anthropic](./remote_anthropic)
|
||||
- [Remote - Azure](./remote_azure)
|
||||
- [Remote - Bedrock](./remote_bedrock)
|
||||
- [Remote - Cerebras](./remote_cerebras)
|
||||
- [Remote - Databricks](./remote_databricks)
|
||||
- [Remote - Fireworks](./remote_fireworks)
|
||||
- [Remote - Gemini](./remote_gemini)
|
||||
- [Remote - Groq](./remote_groq)
|
||||
- [Remote - Hf - Endpoint](./remote_hf_endpoint)
|
||||
- [Remote - Hf - Serverless](./remote_hf_serverless)
|
||||
- [Remote - Llama-Openai-Compat](./remote_llama-openai-compat)
|
||||
- [Remote - Nvidia](./remote_nvidia)
|
||||
- [Remote - Ollama](./remote_ollama)
|
||||
- [Remote - Openai](./remote_openai)
|
||||
- [Remote - Passthrough](./remote_passthrough)
|
||||
- [Remote - Runpod](./remote_runpod)
|
||||
- [Remote - Sambanova](./remote_sambanova)
|
||||
- [Remote - Tgi](./remote_tgi)
|
||||
- [Remote - Together](./remote_together)
|
||||
- [Remote - Vertexai](./remote_vertexai)
|
||||
- [Remote - Vllm](./remote_vllm)
|
||||
- [Remote - Watsonx](./remote_watsonx)
|
||||
Loading…
Add table
Add a link
Reference in a new issue