From 4487b88ffe7521cd167aca48bf6b3ab9eea3bfd7 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Fri, 26 Sep 2025 21:11:51 +0000 Subject: [PATCH] build: Bump version to 0.2.23 --- llama_stack/ui/package-lock.json | 8 ++++---- llama_stack/ui/package.json | 2 +- pyproject.toml | 6 +++--- uv.lock | 12 ++++++------ 4 files changed, 14 insertions(+), 14 deletions(-) diff --git a/llama_stack/ui/package-lock.json b/llama_stack/ui/package-lock.json index 638655639..21dec59c3 100644 --- a/llama_stack/ui/package-lock.json +++ b/llama_stack/ui/package-lock.json @@ -18,7 +18,7 @@ "class-variance-authority": "^0.7.1", "clsx": "^2.1.1", "framer-motion": "^12.23.12", - "llama-stack-client": "^0.2.22", + "llama-stack-client": "^0.2.23", "lucide-react": "^0.542.0", "next": "15.5.3", "next-auth": "^4.24.11", @@ -10172,9 +10172,9 @@ "license": "MIT" }, "node_modules/llama-stack-client": { - "version": "0.2.22", - "resolved": "https://registry.npmjs.org/llama-stack-client/-/llama-stack-client-0.2.22.tgz", - "integrity": "sha512-7aW3UQj5MwjV73Brd+yQ1e4W1W33nhozyeHM5tzOgbsVZ88tL78JNiNvyFqDR5w6V9XO4/uSGGiQVG6v83yR4w==", + "version": "0.2.23", + "resolved": "https://registry.npmjs.org/llama-stack-client/-/llama-stack-client-0.2.23.tgz", + "integrity": "sha512-J3YFH1HW2K70capejQxGlCyTgKdfx+sQf8Ab+HFi1j2Q00KtpHXB79RxejvBxjWC3X2E++P9iU57KdU2Tp/rIQ==", "license": "MIT", "dependencies": { "@types/node": "^18.11.18", diff --git a/llama_stack/ui/package.json b/llama_stack/ui/package.json index 2ba81ea84..70462b534 100644 --- a/llama_stack/ui/package.json +++ b/llama_stack/ui/package.json @@ -23,7 +23,7 @@ "class-variance-authority": "^0.7.1", "clsx": "^2.1.1", "framer-motion": "^12.23.12", - "llama-stack-client": "^0.2.22", + "llama-stack-client": "^0.2.23", "lucide-react": "^0.542.0", "next": "15.5.3", "next-auth": "^4.24.11", diff --git a/pyproject.toml b/pyproject.toml index a26c4d645..98bae47c5 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -7,7 +7,7 @@ required-version = ">=0.7.0" [project] name = "llama_stack" -version = "0.2.22" +version = "0.2.23" authors = [{ name = "Meta Llama", email = "llama-oss@meta.com" }] description = "Llama Stack" readme = "README.md" @@ -31,7 +31,7 @@ dependencies = [ "huggingface-hub>=0.34.0,<1.0", "jinja2>=3.1.6", "jsonschema", - "llama-stack-client>=0.2.22", + "llama-stack-client>=0.2.23", "openai>=1.100.0", # for expires_after support "prompt-toolkit", "python-dotenv", @@ -55,7 +55,7 @@ dependencies = [ ui = [ "streamlit", "pandas", - "llama-stack-client>=0.2.22", + "llama-stack-client>=0.2.23", "streamlit-option-menu", ] diff --git a/uv.lock b/uv.lock index 0833a9d77..63639ee4a 100644 --- a/uv.lock +++ b/uv.lock @@ -1749,7 +1749,7 @@ wheels = [ [[package]] name = "llama-stack" -version = "0.2.22" +version = "0.2.23" source = { editable = "." } dependencies = [ { name = "aiohttp" }, @@ -1885,8 +1885,8 @@ requires-dist = [ { name = "huggingface-hub", specifier = ">=0.34.0,<1.0" }, { name = "jinja2", specifier = ">=3.1.6" }, { name = "jsonschema" }, - { name = "llama-stack-client", specifier = ">=0.2.22" }, - { name = "llama-stack-client", marker = "extra == 'ui'", specifier = ">=0.2.22" }, + { name = "llama-stack-client", specifier = ">=0.2.23" }, + { name = "llama-stack-client", marker = "extra == 'ui'", specifier = ">=0.2.23" }, { name = "openai", specifier = ">=1.100.0" }, { name = "opentelemetry-exporter-otlp-proto-http", specifier = ">=1.30.0" }, { name = "opentelemetry-sdk", specifier = ">=1.30.0" }, @@ -1993,7 +1993,7 @@ unit = [ [[package]] name = "llama-stack-client" -version = "0.2.22" +version = "0.2.23" source = { registry = "https://pypi.org/simple" } dependencies = [ { name = "anyio" }, @@ -2012,9 +2012,9 @@ dependencies = [ { name = "tqdm" }, { name = "typing-extensions" }, ] -sdist = { url = "https://files.pythonhosted.org/packages/60/80/4260816bfaaa889d515206c9df4906d08d405bf94c9b4d1be399b1923e46/llama_stack_client-0.2.22.tar.gz", hash = "sha256:9a0bc756b91ebd539858eeaf1f231c5e5c6900e1ea4fcced726c6717f3d27ca7", size = 318309, upload-time = "2025-09-16T19:43:33.212Z" } +sdist = { url = "https://files.pythonhosted.org/packages/9f/8f/306d5fcf2f97b3a6251219b03c194836a2ff4e0fcc8146c9970e50a72cd3/llama_stack_client-0.2.23.tar.gz", hash = "sha256:68f34e8ac8eea6a73ed9d4977d849992b2d8bd835804d770a11843431cd5bf74", size = 322288, upload-time = "2025-09-26T21:11:08.342Z" } wheels = [ - { url = "https://files.pythonhosted.org/packages/d1/8e/1ebf6ac0dbb62b81038e856ed00768e283d927b14fcd614e3018a227092b/llama_stack_client-0.2.22-py3-none-any.whl", hash = "sha256:b260d73aec56fcfd8fa601b3b34c2f83c4fbcfb7261a246b02bbdf6c2da184fe", size = 369901, upload-time = "2025-09-16T19:43:32.089Z" }, + { url = "https://files.pythonhosted.org/packages/fa/75/3eb58e092a681804013dbec7b7f549d18f55acf6fd6e6b27de7e249766d8/llama_stack_client-0.2.23-py3-none-any.whl", hash = "sha256:eee42c74eee8f218f9455e5a06d5d4be43f8a8c82a7937ef51ce367f916df847", size = 379809, upload-time = "2025-09-26T21:11:06.856Z" }, ] [[package]]