client.agents.turn.create(session_id, \*, agent_id, \*\*params) -> TurnCreateResponse
- client.agents.turn.retrieve(turn_id, \*, agent_id, session_id) -> Turn
-## BatchInference
-Types:
-
-```python
-from llama_stack_client.types import BatchInferenceChatCompletionResponse
-```
-
-Methods:
-
-- client.batch_inference.chat_completion(\*\*params) -> BatchInferenceChatCompletionResponse
-- client.batch_inference.completion(\*\*params) -> BatchCompletion
## Datasets
diff --git a/docs/docusaurus.config.ts b/docs/docusaurus.config.ts
new file mode 100644
index 000000000..937aa4ddf
--- /dev/null
+++ b/docs/docusaurus.config.ts
@@ -0,0 +1,225 @@
+// @ts-check
+// Note: type annotations allow type checking and IDEs autocompletion
+
+import type * as Preset from "@docusaurus/preset-classic";
+import type { Config } from "@docusaurus/types";
+import type * as Plugin from "@docusaurus/types/src/plugin";
+import type * as OpenApiPlugin from "docusaurus-plugin-openapi-docs";
+
+const config: Config = {
+ title: 'Llama Stack',
+ tagline: 'The open-source framework for building generative AI applications',
+ url: 'https://llamastack.github.io',
+ baseUrl: '/',
+ onBrokenLinks: "warn",
+ onBrokenMarkdownLinks: "warn",
+ favicon: "img/favicon.ico",
+
+ // GitHub pages deployment config.
+ organizationName: 'reluctantfuturist',
+ projectName: 'llama-stack',
+ trailingSlash: false,
+
+ presets: [
+ [
+ "classic",
+ {
+ docs: {
+ sidebarPath: require.resolve("./sidebars.ts"),
+ // Please change this to your repo.
+ // Remove this to remove the "edit this page" links.
+ editUrl: 'https://github.com/meta-llama/llama-stack/tree/main/docs/',
+ docItemComponent: "@theme/ApiItem", // Derived from docusaurus-theme-openapi
+ },
+ blog: false,
+ theme: {
+ customCss: require.resolve("./src/css/custom.css"),
+ },
+ } satisfies Preset.Options,
+ ],
+ ],
+
+ themeConfig: {
+ image: 'img/llama-stack.png',
+ navbar: {
+ title: 'Llama Stack',
+ logo: {
+ alt: 'Llama Stack Logo',
+ src: 'img/llama-stack-logo.png',
+ },
+ items: [
+ {
+ type: 'docSidebar',
+ sidebarId: 'tutorialSidebar',
+ position: 'left',
+ label: 'Docs',
+ },
+ {
+ type: 'docSidebar',
+ sidebarId: 'apiSidebar',
+ position: 'left',
+ label: 'API Reference',
+ },
+ {
+ href: 'https://github.com/llamastack/llama-stack',
+ label: 'GitHub',
+ position: 'right',
+ },
+ ],
+ },
+ footer: {
+ style: 'dark',
+ links: [
+ {
+ title: 'Docs',
+ items: [
+ {
+ label: 'Getting Started',
+ to: '/docs/getting_started/quickstart',
+ },
+ {
+ label: 'Concepts',
+ to: '/docs/concepts',
+ },
+ {
+ label: 'API Reference',
+ to: '/docs/api/llama-stack-specification',
+ },
+ ],
+ },
+ {
+ title: 'Community',
+ items: [
+ {
+ label: 'Discord',
+ href: 'https://discord.gg/llama-stack',
+ },
+ {
+ label: 'GitHub Discussions',
+ href: 'https://github.com/llamastack/llama-stack/discussions',
+ },
+ {
+ label: 'Issues',
+ href: 'https://github.com/llamastack/llama-stack/issues',
+ },
+ ],
+ },
+ {
+ title: 'More',
+ items: [
+ {
+ label: 'GitHub',
+ href: 'https://github.com/llamastack/llama-stack',
+ },
+ {
+ label: 'PyPI',
+ href: 'https://pypi.org/project/llama-stack/',
+ },
+ ],
+ },
+ ],
+ copyright: `Copyright © ${new Date().getFullYear()} Meta Platforms, Inc. Built with Docusaurus.`,
+ },
+ prism: {
+ additionalLanguages: [
+ 'ruby',
+ 'csharp',
+ 'php',
+ 'java',
+ 'powershell',
+ 'json',
+ 'bash',
+ 'python',
+ 'yaml',
+ ],
+ },
+ docs: {
+ sidebar: {
+ hideable: true,
+ },
+ },
+ // Language tabs for API documentation
+ languageTabs: [
+ {
+ highlight: "python",
+ language: "python",
+ logoClass: "python",
+ },
+ {
+ highlight: "bash",
+ language: "curl",
+ logoClass: "curl",
+ },
+ {
+ highlight: "javascript",
+ language: "nodejs",
+ logoClass: "nodejs",
+ },
+ {
+ highlight: "java",
+ language: "java",
+ logoClass: "java",
+ },
+ ],
+ } satisfies Preset.ThemeConfig,
+
+ plugins: [
+ [
+ "docusaurus-plugin-openapi-docs",
+ {
+ id: "openapi",
+ docsPluginId: "classic",
+ config: {
+ llamastack: {
+ specPath: "static/llama-stack-spec.yaml",
+ outputDir: "docs/api",
+ downloadUrl: "https://raw.githubusercontent.com/meta-llama/llama-stack/main/docs/static/llama-stack-spec.yaml",
+ sidebarOptions: {
+ groupPathsBy: "tag",
+ categoryLinkSource: "tag",
+ },
+ } satisfies OpenApiPlugin.Options,
+ } satisfies Plugin.PluginOptions,
+ },
+ ],
+ ],
+
+ themes: [
+ "docusaurus-theme-openapi-docs",
+ [
+ require.resolve("@easyops-cn/docusaurus-search-local"),
+ {
+ // Optimization for production
+ hashed: true,
+
+ // Language settings
+ language: ["en"],
+
+ // Content indexing settings
+ indexDocs: true,
+ indexBlog: false, // No blog in Llama Stack
+ indexPages: true,
+
+ // Route configuration
+ docsRouteBasePath: '/docs',
+
+ // Search behavior optimization for technical docs
+ searchResultLimits: 8,
+ searchResultContextMaxLength: 50,
+ explicitSearchResultPath: true,
+
+ // User experience enhancements
+ searchBarShortcut: true,
+ searchBarShortcutHint: true,
+ searchBarPosition: "right",
+
+ // Performance optimizations
+ ignoreFiles: [
+ "node_modules/**/*",
+ ],
+ },
+ ],
+ ],
+};
+
+export default config;
diff --git a/docs/getting_started.ipynb b/docs/getting_started.ipynb
index eeebf12d9..641cf4224 100644
--- a/docs/getting_started.ipynb
+++ b/docs/getting_started.ipynb
@@ -11,11 +11,11 @@
"\n",
"# Llama Stack - Building AI Applications\n",
"\n",
- "docs
directory.
+ >
+ ),
+ },
+ {
+ title: 'Powered by React',
+ Svg: require('@site/static/img/undraw_docusaurus_react.svg').default,
+ description: (
+ <>
+ Extend or customize your website layout by reusing React. Docusaurus can
+ be extended while reusing the same header and footer.
+ >
+ ),
+ },
+];
+
+function Feature({Svg, title, description}) {
+ return (
+ {description}
++ Unified APIs for Inference, RAG, Agents, Tools, Safety, and Telemetry +
++ Get up and running with Llama Stack in just a few commands. Build your first RAG application locally. +
+{`# Install uv and start Ollama
+ollama run llama3.2:3b --keepalive 60m
+
+# Run Llama Stack server
+OLLAMA_URL=http://localhost:11434 \\
+ uv run --with llama-stack \\
+ llama stack build --distro starter \\
+ --image-type venv --run
+
+# Try the Python SDK
+from llama_stack_client import LlamaStackClient
+
+client = LlamaStackClient(
+ base_url="http://localhost:8321"
+)
+
+response = client.inference.chat_completion(
+ model="Llama3.2-3B-Instruct",
+ messages=[{
+ "role": "user",
+ "content": "What is machine learning?"
+ }]
+)`}
+ One consistent interface for all your AI needs - inference, safety, agents, and more.
+Swap between providers without code changes. Start local, deploy anywhere.
+Built-in safety, monitoring, and evaluation tools for enterprise applications.
+SDKs for Python, Node.js, iOS, Android, and REST APIs for any language.
++ Connect with developers building the future of AI applications +
+ +