refactor: remove lama-api-client from pyproject.toml (#3299)
Some checks failed
Integration Auth Tests / test-matrix (oauth2_token) (push) Failing after 0s
Vector IO Integration Tests / test-matrix (push) Failing after 1s
Integration Tests (Replay) / Integration Tests (, , , client=, vision=) (push) Failing after 1s
Pre-commit / pre-commit (push) Failing after 1s
Test Llama Stack Build / generate-matrix (push) Failing after 0s
Test Llama Stack Build / build-ubi9-container-distribution (push) Failing after 1s
Test Llama Stack Build / build (push) Has been skipped
Test External Providers Installed via Module / test-external-providers-from-module (venv) (push) Has been skipped
Python Package Build Test / build (3.12) (push) Failing after 1s
Python Package Build Test / build (3.13) (push) Failing after 1s
Unit Tests / unit-tests (3.12) (push) Failing after 1s
Test External API and Providers / test-external (venv) (push) Failing after 1s
Unit Tests / unit-tests (3.13) (push) Failing after 1s
Update ReadTheDocs / update-readthedocs (push) Failing after 1s
UI Tests / ui-tests (22) (push) Failing after 2s
Test Llama Stack Build / build-custom-container-distribution (push) Has started running
Test Llama Stack Build / build-single-provider (push) Has started running
SqlStore Integration Tests / test-postgres (3.12) (push) Failing after 7s
SqlStore Integration Tests / test-postgres (3.13) (push) Failing after 8s

# What does this PR do?
<!-- Provide a short summary of what this PR does and why. Link to
relevant issues if applicable. -->
This PR is eliminating `lama-api-client` dependency at `pyproject.toml`
because it's not used in Llama Stack codebase

<!-- If resolving an issue, uncomment and update the line below -->
<!-- Closes #[issue-number] -->

## Test Plan
<!-- Describe the tests you ran to verify your changes with result
summaries. *Provide clear instructions so the plan can be easily
re-executed.* -->
` ./scripts/unit-tests.sh`
This commit is contained in:
IAN MILLER 2025-09-01 15:50:50 +01:00 committed by GitHub
parent 9625ac6d02
commit 4a59961a6c
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
2 changed files with 0 additions and 19 deletions

View file

@ -32,7 +32,6 @@ dependencies = [
"jinja2>=3.1.6", "jinja2>=3.1.6",
"jsonschema", "jsonschema",
"llama-stack-client>=0.2.20", "llama-stack-client>=0.2.20",
"llama-api-client>=0.1.2",
"openai>=1.99.6", "openai>=1.99.6",
"prompt-toolkit", "prompt-toolkit",
"python-dotenv", "python-dotenv",

18
uv.lock generated
View file

@ -1748,22 +1748,6 @@ wheels = [
{ url = "https://files.pythonhosted.org/packages/5f/e4/f1546746049c99c6b8b247e2f34485b9eae36faa9322b84e2a17262e6712/litellm-1.74.9-py3-none-any.whl", hash = "sha256:ab8f8a6e4d8689d3c7c4f9c3bbc7e46212cc3ebc74ddd0f3c0c921bb459c9874", size = 8740449, upload-time = "2025-07-28T16:42:36.8Z" }, { url = "https://files.pythonhosted.org/packages/5f/e4/f1546746049c99c6b8b247e2f34485b9eae36faa9322b84e2a17262e6712/litellm-1.74.9-py3-none-any.whl", hash = "sha256:ab8f8a6e4d8689d3c7c4f9c3bbc7e46212cc3ebc74ddd0f3c0c921bb459c9874", size = 8740449, upload-time = "2025-07-28T16:42:36.8Z" },
] ]
[[package]]
name = "llama-api-client"
version = "0.2.0"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "anyio" },
{ name = "distro" },
{ name = "httpx" },
{ name = "pydantic" },
{ name = "sniffio" },
{ name = "typing-extensions" },
]
sdist = { url = "https://files.pythonhosted.org/packages/59/41/fa8521a0faff96bf5f810e2ab5b78c638f5ba44afd09aa86f94b6a1226ad/llama_api_client-0.2.0.tar.gz", hash = "sha256:b9bd5f5ad332b9133f0775a105f0940f057cbb311891f1d4487247d001c31f17", size = 117108, upload-time = "2025-08-12T17:07:07.734Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/1d/11/198e65c1a50d9e839b4e3d346b4bd0f624e532446e468d1aba6c74ed7484/llama_api_client-0.2.0-py3-none-any.whl", hash = "sha256:50614ed991e1a72439e6a624a97e6000615ada1b9e2046ecc026fe62f107663c", size = 85002, upload-time = "2025-08-12T17:07:06.293Z" },
]
[[package]] [[package]]
name = "llama-stack" name = "llama-stack"
@ -1780,7 +1764,6 @@ dependencies = [
{ name = "huggingface-hub" }, { name = "huggingface-hub" },
{ name = "jinja2" }, { name = "jinja2" },
{ name = "jsonschema" }, { name = "jsonschema" },
{ name = "llama-api-client" },
{ name = "llama-stack-client" }, { name = "llama-stack-client" },
{ name = "openai" }, { name = "openai" },
{ name = "opentelemetry-exporter-otlp-proto-http" }, { name = "opentelemetry-exporter-otlp-proto-http" },
@ -1906,7 +1889,6 @@ requires-dist = [
{ name = "huggingface-hub", specifier = ">=0.34.0,<1.0" }, { name = "huggingface-hub", specifier = ">=0.34.0,<1.0" },
{ name = "jinja2", specifier = ">=3.1.6" }, { name = "jinja2", specifier = ">=3.1.6" },
{ name = "jsonschema" }, { name = "jsonschema" },
{ name = "llama-api-client", specifier = ">=0.1.2" },
{ name = "llama-stack-client", specifier = ">=0.2.20" }, { name = "llama-stack-client", specifier = ">=0.2.20" },
{ name = "llama-stack-client", marker = "extra == 'ui'", specifier = ">=0.2.20" }, { name = "llama-stack-client", marker = "extra == 'ui'", specifier = ">=0.2.20" },
{ name = "openai", specifier = ">=1.99.6" }, { name = "openai", specifier = ">=1.99.6" },