forked from phoenix/litellm-mirror
407 lines
12 KiB
JavaScript
407 lines
12 KiB
JavaScript
/**
|
|
* Creating a sidebar enables you to:
|
|
- create an ordered group of docs
|
|
- render a sidebar for each doc of that group
|
|
- provide next/previous navigation
|
|
|
|
The sidebars can be generated from the filesystem, or explicitly defined here.
|
|
|
|
Create as many sidebars as you want.
|
|
*/
|
|
|
|
// @ts-check
|
|
|
|
/** @type {import('@docusaurus/plugin-content-docs').SidebarsConfig} */
|
|
const sidebars = {
|
|
// // By default, Docusaurus generates a sidebar from the docs folder structure
|
|
|
|
// But you can create a sidebar manually
|
|
tutorialSidebar: [
|
|
{ type: "doc", id: "index" }, // NEW
|
|
{
|
|
type: "category",
|
|
label: "LiteLLM Proxy Server",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "LiteLLM Proxy Server (LLM Gateway)",
|
|
description: `OpenAI Proxy Server (LLM Gateway) to call 100+ LLMs in a unified interface & track spend, set budgets per virtual key/user`,
|
|
slug: "/simple_proxy",
|
|
},
|
|
items: [
|
|
"proxy/docker_quick_start",
|
|
{
|
|
"type": "category",
|
|
"label": "Config.yaml",
|
|
"items": ["proxy/configs", "proxy/config_settings"]
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Setup & Deployment",
|
|
items: [
|
|
"proxy/deploy",
|
|
"proxy/prod",
|
|
"proxy/cli",
|
|
"proxy/model_management",
|
|
"proxy/health",
|
|
"proxy/debugging",
|
|
"proxy/pass_through",
|
|
],
|
|
},
|
|
"proxy/demo",
|
|
{
|
|
type: "category",
|
|
label: "Architecture",
|
|
items: ["proxy/architecture", "proxy/db_info", "router_architecture"],
|
|
},
|
|
{
|
|
type: "link",
|
|
label: "All Endpoints (Swagger)",
|
|
href: "https://litellm-api.up.railway.app/",
|
|
},
|
|
"proxy/enterprise",
|
|
{
|
|
type: "category",
|
|
label: "Making LLM Requests",
|
|
items: [
|
|
"proxy/user_keys",
|
|
"proxy/response_headers",
|
|
"pass_through/vertex_ai",
|
|
"pass_through/google_ai_studio",
|
|
"pass_through/cohere",
|
|
"pass_through/anthropic_completion",
|
|
"pass_through/bedrock",
|
|
"pass_through/langfuse"
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Authentication",
|
|
items: [
|
|
"proxy/virtual_keys",
|
|
"proxy/token_auth",
|
|
"proxy/service_accounts",
|
|
"proxy/access_control",
|
|
"proxy/ip_address",
|
|
"proxy/email",
|
|
"proxy/multiple_admins",
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Admin UI",
|
|
items: [
|
|
"proxy/ui",
|
|
"proxy/self_serve",
|
|
"proxy/custom_sso"
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Spend Tracking + Budgets",
|
|
items: ["proxy/cost_tracking", "proxy/users", "proxy/custom_pricing", "proxy/team_budgets", "proxy/billing", "proxy/customers"],
|
|
},
|
|
{
|
|
type: "link",
|
|
label: "Load Balancing, Routing, Fallbacks",
|
|
href: "https://docs.litellm.ai/docs/routing-load-balancing",
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Logging, Alerting, Metrics",
|
|
items: ["proxy/logging", "proxy/team_logging","proxy/alerting", "proxy/prometheus",],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "[Beta] Guardrails",
|
|
items: [
|
|
"proxy/guardrails/quick_start",
|
|
"proxy/guardrails/aporia_api",
|
|
"proxy/guardrails/guardrails_ai",
|
|
"proxy/guardrails/lakera_ai",
|
|
"proxy/guardrails/bedrock",
|
|
"proxy/guardrails/pii_masking_v2",
|
|
"proxy/guardrails/secret_detection",
|
|
"proxy/guardrails/custom_guardrail",
|
|
"prompt_injection"
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Secret Managers",
|
|
items: [
|
|
"secret",
|
|
"oidc"
|
|
]
|
|
},
|
|
"proxy/caching",
|
|
"proxy/call_hooks",
|
|
"proxy/rules",
|
|
]
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Supported Models & Providers",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Providers",
|
|
description:
|
|
"Learn how to deploy + call models from different providers on LiteLLM",
|
|
slug: "/providers",
|
|
},
|
|
items: [
|
|
"providers/openai",
|
|
"providers/text_completion_openai",
|
|
"providers/openai_compatible",
|
|
"providers/azure",
|
|
"providers/azure_ai",
|
|
"providers/vertex",
|
|
"providers/gemini",
|
|
"providers/anthropic",
|
|
"providers/aws_sagemaker",
|
|
"providers/bedrock",
|
|
"providers/litellm_proxy",
|
|
"providers/mistral",
|
|
"providers/codestral",
|
|
"providers/cohere",
|
|
"providers/anyscale",
|
|
"providers/huggingface",
|
|
"providers/databricks",
|
|
"providers/watsonx",
|
|
"providers/predibase",
|
|
"providers/nvidia_nim",
|
|
"providers/xai",
|
|
"providers/lm_studio",
|
|
"providers/cerebras",
|
|
"providers/volcano",
|
|
"providers/triton-inference-server",
|
|
"providers/ollama",
|
|
"providers/perplexity",
|
|
"providers/friendliai",
|
|
"providers/groq",
|
|
"providers/github",
|
|
"providers/deepseek",
|
|
"providers/fireworks_ai",
|
|
"providers/clarifai",
|
|
"providers/vllm",
|
|
"providers/xinference",
|
|
"providers/cloudflare_workers",
|
|
"providers/deepinfra",
|
|
"providers/ai21",
|
|
"providers/nlp_cloud",
|
|
"providers/replicate",
|
|
"providers/togetherai",
|
|
"providers/voyage",
|
|
"providers/jina_ai",
|
|
"providers/aleph_alpha",
|
|
"providers/baseten",
|
|
"providers/openrouter",
|
|
"providers/palm",
|
|
"providers/sambanova",
|
|
"providers/custom_llm_server",
|
|
"providers/petals",
|
|
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Guides",
|
|
items: [
|
|
"exception_mapping",
|
|
"completion/provider_specific_params",
|
|
"guides/finetuned_models",
|
|
"completion/audio",
|
|
"completion/vision",
|
|
"completion/json_mode",
|
|
"completion/prompt_caching",
|
|
"completion/predict_outputs",
|
|
"completion/prefix",
|
|
"completion/drop_params",
|
|
"completion/prompt_formatting",
|
|
"completion/stream",
|
|
"completion/message_trimming",
|
|
"completion/function_call",
|
|
"completion/model_alias",
|
|
"completion/batching",
|
|
"completion/mock_requests",
|
|
"completion/reliable_completions",
|
|
|
|
]
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Supported Endpoints",
|
|
items: [
|
|
{
|
|
type: "category",
|
|
label: "Chat",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Chat Completions",
|
|
description: "Details on the completion() function",
|
|
slug: "/completion",
|
|
},
|
|
items: [
|
|
"completion/input",
|
|
"completion/output",
|
|
"completion/usage",
|
|
],
|
|
},
|
|
"embedding/supported_embedding",
|
|
"image_generation",
|
|
{
|
|
type: "category",
|
|
label: "Audio",
|
|
"items": [
|
|
"audio_transcription",
|
|
"text_to_speech",
|
|
]
|
|
},
|
|
"rerank",
|
|
"assistants",
|
|
"batches",
|
|
"realtime",
|
|
"fine_tuning",
|
|
{
|
|
type: "link",
|
|
label: "Use LiteLLM Proxy with Vertex, Bedrock SDK",
|
|
href: "/docs/pass_through/vertex_ai",
|
|
},
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Routing, Loadbalancing & Fallbacks",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Routing, Loadbalancing & Fallbacks",
|
|
description: "Learn how to load balance, route, and set fallbacks for your LLM requests",
|
|
slug: "/routing-load-balancing",
|
|
},
|
|
items: ["routing", "scheduler", "proxy/load_balancing", "proxy/reliability", "proxy/tag_routing", "proxy/provider_budget_routing", "proxy/team_based_routing", "proxy/customer_routing"],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "LiteLLM Python SDK",
|
|
items: [
|
|
"set_keys",
|
|
"completion/token_usage",
|
|
"sdk_custom_pricing",
|
|
"embedding/async_embedding",
|
|
"embedding/moderation",
|
|
"budget_manager",
|
|
"caching/all_caches",
|
|
"migration",
|
|
{
|
|
type: "category",
|
|
label: "LangChain, LlamaIndex, Instructor Integration",
|
|
items: ["langchain/langchain", "tutorials/instructor"],
|
|
},
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Load Testing",
|
|
items: [
|
|
"benchmarks",
|
|
"load_test",
|
|
"load_test_advanced",
|
|
"load_test_sdk",
|
|
"load_test_rpm",
|
|
]
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Logging & Observability",
|
|
items: [
|
|
"observability/langfuse_integration",
|
|
"observability/gcs_bucket_integration",
|
|
"observability/langsmith_integration",
|
|
"observability/literalai_integration",
|
|
"observability/opentelemetry_integration",
|
|
"observability/logfire_integration",
|
|
"observability/argilla",
|
|
"observability/arize_integration",
|
|
"debugging/local_debugging",
|
|
"observability/raw_request_response",
|
|
"observability/custom_callback",
|
|
"observability/scrub_data",
|
|
"observability/braintrust",
|
|
"observability/sentry",
|
|
"observability/lago",
|
|
"observability/helicone_integration",
|
|
"observability/openmeter",
|
|
"observability/promptlayer_integration",
|
|
"observability/wandb_integration",
|
|
"observability/slack_integration",
|
|
"observability/athina_integration",
|
|
"observability/lunary_integration",
|
|
"observability/greenscale_integration",
|
|
"observability/supabase_integration",
|
|
`observability/telemetry`,
|
|
"observability/opik_integration",
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Tutorials",
|
|
items: [
|
|
'tutorials/litellm_proxy_aporia',
|
|
'tutorials/azure_openai',
|
|
'tutorials/instructor',
|
|
"tutorials/gradio_integration",
|
|
"tutorials/huggingface_codellama",
|
|
"tutorials/huggingface_tutorial",
|
|
"tutorials/TogetherAI_liteLLM",
|
|
"tutorials/finetuned_chat_gpt",
|
|
"tutorials/text_completion",
|
|
"tutorials/first_playground",
|
|
"tutorials/model_fallbacks",
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Extras",
|
|
items: [
|
|
"extras/contributing",
|
|
"data_security",
|
|
"migration_policy",
|
|
"contributing",
|
|
"proxy/pii_masking",
|
|
"extras/code_quality",
|
|
"rules",
|
|
"proxy_server",
|
|
{
|
|
type: "category",
|
|
label: "❤️ 🚅 Projects built on LiteLLM",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Projects built on LiteLLM",
|
|
description:
|
|
"Learn how to deploy + call models from different providers on LiteLLM",
|
|
slug: "/project",
|
|
},
|
|
items: [
|
|
"projects/Docq.AI",
|
|
"projects/OpenInterpreter",
|
|
"projects/dbally",
|
|
"projects/FastREPL",
|
|
"projects/PROMPTMETHEUS",
|
|
"projects/Codium PR Agent",
|
|
"projects/Prompt2Model",
|
|
"projects/SalesGPT",
|
|
"projects/Quivr",
|
|
"projects/Langstream",
|
|
"projects/Otter",
|
|
"projects/GPT Migrate",
|
|
"projects/YiVal",
|
|
"projects/LiteLLM Proxy",
|
|
"projects/llm_cord",
|
|
],
|
|
},
|
|
],
|
|
},
|
|
"troubleshoot",
|
|
],
|
|
};
|
|
|
|
module.exports = sidebars;
|