forked from phoenix/litellm-mirror
303 lines
8.6 KiB
JavaScript
303 lines
8.6 KiB
JavaScript
/**
|
|
* Creating a sidebar enables you to:
|
|
- create an ordered group of docs
|
|
- render a sidebar for each doc of that group
|
|
- provide next/previous navigation
|
|
|
|
The sidebars can be generated from the filesystem, or explicitly defined here.
|
|
|
|
Create as many sidebars as you want.
|
|
*/
|
|
|
|
// @ts-check
|
|
|
|
/** @type {import('@docusaurus/plugin-content-docs').SidebarsConfig} */
|
|
const sidebars = {
|
|
// // By default, Docusaurus generates a sidebar from the docs folder structure
|
|
|
|
// But you can create a sidebar manually
|
|
tutorialSidebar: [
|
|
{ type: "doc", id: "index" }, // NEW
|
|
{
|
|
type: "category",
|
|
label: "💥 LiteLLM Proxy Server",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "💥 LiteLLM Proxy Server",
|
|
description: `Proxy Server to call 100+ LLMs in a unified interface & track spend, set budgets per virtual key/user`,
|
|
slug: "/simple_proxy",
|
|
},
|
|
items: [
|
|
"proxy/quick_start",
|
|
"proxy/deploy",
|
|
"proxy/prod",
|
|
{
|
|
type: "link",
|
|
label: "📖 All Endpoints (Swagger)",
|
|
href: "https://litellm-api.up.railway.app/",
|
|
},
|
|
"proxy/enterprise",
|
|
"proxy/user_keys",
|
|
"proxy/demo",
|
|
"proxy/configs",
|
|
"proxy/reliability",
|
|
"proxy/cost_tracking",
|
|
"proxy/custom_pricing",
|
|
"proxy/self_serve",
|
|
"proxy/virtual_keys",
|
|
{
|
|
type: "category",
|
|
label: "🪢 Logging",
|
|
items: ["proxy/logging", "proxy/bucket", "proxy/streaming_logging"],
|
|
},
|
|
"proxy/team_logging",
|
|
"proxy/guardrails",
|
|
"proxy/tag_routing",
|
|
"proxy/users",
|
|
"proxy/team_budgets",
|
|
"proxy/customers",
|
|
"proxy/billing",
|
|
"proxy/token_auth",
|
|
"proxy/alerting",
|
|
"proxy/ui",
|
|
"proxy/prometheus",
|
|
"proxy/pass_through",
|
|
"proxy/email",
|
|
"proxy/multiple_admins",
|
|
"proxy/team_based_routing",
|
|
"proxy/customer_routing",
|
|
{
|
|
type: "category",
|
|
label: "Extra Load Balancing",
|
|
items: ["proxy/load_balancing"],
|
|
},
|
|
"proxy/model_management",
|
|
"proxy/health",
|
|
"proxy/debugging",
|
|
"proxy/pii_masking",
|
|
"proxy/prompt_injection",
|
|
"proxy/caching",
|
|
"proxy/call_hooks",
|
|
"proxy/rules",
|
|
"proxy/cli",
|
|
]
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "💯 Supported Models & Providers",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Providers",
|
|
description:
|
|
"Learn how to deploy + call models from different providers on LiteLLM",
|
|
slug: "/providers",
|
|
},
|
|
items: [
|
|
"providers/openai",
|
|
"providers/text_completion_openai",
|
|
"providers/openai_compatible",
|
|
"providers/azure",
|
|
"providers/azure_ai",
|
|
"providers/vertex",
|
|
"providers/palm",
|
|
"providers/gemini",
|
|
"providers/anthropic",
|
|
"providers/aws_sagemaker",
|
|
"providers/bedrock",
|
|
"providers/mistral",
|
|
"providers/codestral",
|
|
"providers/cohere",
|
|
"providers/anyscale",
|
|
"providers/huggingface",
|
|
"providers/databricks",
|
|
"providers/watsonx",
|
|
"providers/predibase",
|
|
"providers/nvidia_nim",
|
|
"providers/volcano",
|
|
"providers/triton-inference-server",
|
|
"providers/ollama",
|
|
"providers/perplexity",
|
|
"providers/friendliai",
|
|
"providers/groq",
|
|
"providers/github",
|
|
"providers/deepseek",
|
|
"providers/fireworks_ai",
|
|
"providers/clarifai",
|
|
"providers/vllm",
|
|
"providers/xinference",
|
|
"providers/cloudflare_workers",
|
|
"providers/deepinfra",
|
|
"providers/ai21",
|
|
"providers/nlp_cloud",
|
|
"providers/replicate",
|
|
"providers/togetherai",
|
|
"providers/voyage",
|
|
"providers/aleph_alpha",
|
|
"providers/baseten",
|
|
"providers/openrouter",
|
|
// "providers/custom_openai_proxy",
|
|
"providers/custom_llm_server",
|
|
"providers/petals",
|
|
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "litellm.completion()",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Completion()",
|
|
description: "Details on the completion() function",
|
|
slug: "/completion",
|
|
},
|
|
items: [
|
|
"completion/input",
|
|
"completion/provider_specific_params",
|
|
"completion/json_mode",
|
|
"completion/drop_params",
|
|
"completion/prompt_formatting",
|
|
"completion/output",
|
|
"exception_mapping",
|
|
"completion/stream",
|
|
"completion/message_trimming",
|
|
"completion/function_call",
|
|
"completion/vision",
|
|
"completion/model_alias",
|
|
"completion/batching",
|
|
"completion/mock_requests",
|
|
"completion/reliable_completions",
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Embedding(), Image Generation(), Assistants(), Moderation(), Audio Transcriptions(), TTS(), Batches(), Fine-Tuning()",
|
|
items: [
|
|
"embedding/supported_embedding",
|
|
"embedding/async_embedding",
|
|
"embedding/moderation",
|
|
"image_generation",
|
|
"audio_transcription",
|
|
"text_to_speech",
|
|
"assistants",
|
|
"batches",
|
|
"fine_tuning",
|
|
"anthropic_completion"
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Secret Manager",
|
|
items: [
|
|
"secret",
|
|
"oidc"
|
|
]
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "🚅 LiteLLM Python SDK",
|
|
items: [
|
|
"routing",
|
|
"scheduler",
|
|
"set_keys",
|
|
"completion/token_usage",
|
|
"sdk_custom_pricing",
|
|
"budget_manager",
|
|
"caching/all_caches",
|
|
{
|
|
type: "category",
|
|
label: "LangChain, LlamaIndex, Instructor Integration",
|
|
items: ["langchain/langchain", "tutorials/instructor"],
|
|
},
|
|
],
|
|
},
|
|
"load_test",
|
|
{
|
|
type: "category",
|
|
label: "Logging & Observability",
|
|
items: [
|
|
"observability/langfuse_integration",
|
|
"observability/logfire_integration",
|
|
"observability/gcs_bucket_integration",
|
|
"observability/langsmith_integration",
|
|
"observability/arize_integration",
|
|
"debugging/local_debugging",
|
|
"observability/raw_request_response",
|
|
"observability/custom_callback",
|
|
"observability/scrub_data",
|
|
"observability/braintrust",
|
|
"observability/sentry",
|
|
"observability/lago",
|
|
"observability/helicone_integration",
|
|
"observability/openmeter",
|
|
"observability/promptlayer_integration",
|
|
"observability/wandb_integration",
|
|
"observability/slack_integration",
|
|
"observability/traceloop_integration",
|
|
"observability/athina_integration",
|
|
"observability/lunary_integration",
|
|
"observability/greenscale_integration",
|
|
"observability/supabase_integration",
|
|
`observability/telemetry`,
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Tutorials",
|
|
items: [
|
|
'tutorials/azure_openai',
|
|
'tutorials/instructor',
|
|
"tutorials/gradio_integration",
|
|
"tutorials/huggingface_codellama",
|
|
"tutorials/huggingface_tutorial",
|
|
"tutorials/TogetherAI_liteLLM",
|
|
"tutorials/finetuned_chat_gpt",
|
|
"tutorials/sagemaker_llms",
|
|
"tutorials/text_completion",
|
|
"tutorials/first_playground",
|
|
"tutorials/model_fallbacks",
|
|
],
|
|
},
|
|
{
|
|
type: "category",
|
|
label: "Extras",
|
|
items: [
|
|
"extras/contributing",
|
|
"data_security",
|
|
"contributing",
|
|
"rules",
|
|
"proxy_server",
|
|
{
|
|
type: "category",
|
|
label: "❤️ 🚅 Projects built on LiteLLM",
|
|
link: {
|
|
type: "generated-index",
|
|
title: "Projects built on LiteLLM",
|
|
description:
|
|
"Learn how to deploy + call models from different providers on LiteLLM",
|
|
slug: "/project",
|
|
},
|
|
items: [
|
|
"projects/Docq.AI",
|
|
"projects/OpenInterpreter",
|
|
"projects/FastREPL",
|
|
"projects/PROMPTMETHEUS",
|
|
"projects/Codium PR Agent",
|
|
"projects/Prompt2Model",
|
|
"projects/SalesGPT",
|
|
"projects/Quivr",
|
|
"projects/Langstream",
|
|
"projects/Otter",
|
|
"projects/GPT Migrate",
|
|
"projects/YiVal",
|
|
"projects/LiteLLM Proxy",
|
|
"projects/llm_cord",
|
|
],
|
|
},
|
|
],
|
|
},
|
|
"migration",
|
|
"troubleshoot",
|
|
],
|
|
};
|
|
|
|
module.exports = sidebars;
|