From 2424a3d3b209f34317b7d91793e6f436bddb9879 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 24 Nov 2025 21:12:52 +0000 Subject: [PATCH] build: Bump version to 0.3.3 --- llama_stack/ui/package-lock.json | 8 ++++---- llama_stack/ui/package.json | 2 +- pyproject.toml | 6 +++--- 3 files changed, 8 insertions(+), 8 deletions(-) diff --git a/llama_stack/ui/package-lock.json b/llama_stack/ui/package-lock.json index dc388025a..5efbe9ca7 100644 --- a/llama_stack/ui/package-lock.json +++ b/llama_stack/ui/package-lock.json @@ -18,7 +18,7 @@ "class-variance-authority": "^0.7.1", "clsx": "^2.1.1", "framer-motion": "^12.23.24", - "llama-stack-client": "^0.3.2", + "llama-stack-client": "^0.3.3", "lucide-react": "^0.545.0", "next": "15.5.4", "next-auth": "^4.24.11", @@ -9635,9 +9635,9 @@ "license": "MIT" }, "node_modules/llama-stack-client": { - "version": "0.3.2", - "resolved": "https://registry.npmjs.org/llama-stack-client/-/llama-stack-client-0.3.2.tgz", - "integrity": "sha512-vzcnIN6k3sp7dhMXSnyrPSd82ACH/H3snj2uF6DgZwZCacKQNp2Y5XIT5qZZgoM1EUXbaxdVYFCeWD9yNCwatw==", + "version": "0.3.3", + "resolved": "https://registry.npmjs.org/llama-stack-client/-/llama-stack-client-0.3.3.tgz", + "integrity": "sha512-eESw7ousfilyD3vQZrSVe51cuFn32VHXuFlrsVtopwLdjW6AOzuzLlfJLO2jroRqTGGI/vqaykqDvZSyxJHr9A==", "license": "MIT", "dependencies": { "@types/node": "^18.11.18", diff --git a/llama_stack/ui/package.json b/llama_stack/ui/package.json index a8e1ba15f..a5e044066 100644 --- a/llama_stack/ui/package.json +++ b/llama_stack/ui/package.json @@ -23,7 +23,7 @@ "class-variance-authority": "^0.7.1", "clsx": "^2.1.1", "framer-motion": "^12.23.24", - "llama-stack-client": "^0.3.2", + "llama-stack-client": "^0.3.3", "lucide-react": "^0.545.0", "next": "15.5.4", "next-auth": "^4.24.11", diff --git a/pyproject.toml b/pyproject.toml index 08d89beaa..49ea7290a 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -7,7 +7,7 @@ required-version = ">=0.7.0" [project] name = "llama_stack" -version = "0.3.3rc1" +version = "0.3.3" authors = [{ name = "Meta Llama", email = "llama-oss@meta.com" }] description = "Llama Stack" readme = "README.md" @@ -30,7 +30,7 @@ dependencies = [ "httpx", "jinja2>=3.1.6", "jsonschema", - "llama-stack-client>=0.3.3rc1", + "llama-stack-client>=0.3.3", "openai>=1.107", # for expires_after support "prompt-toolkit", "python-dotenv", @@ -55,7 +55,7 @@ dependencies = [ ui = [ "streamlit", "pandas", - "llama-stack-client>=0.3.3rc1", + "llama-stack-client>=0.3.3", "streamlit-option-menu", ]