mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-24 21:29:53 +00:00
build: add missing dev dependencies for unit tests (#1004)
# What does this PR do? Added necessary dependencies to ensure successful execution of unit tests. Without these, the following command would fail due to missing imports: ``` uv run pytest -v -k "ollama" \ --inference-model=llama3.2:3b-instruct-fp16 llama_stack/providers/tests/inference/test_model_registration.py ``` Signed-off-by: Sébastien Han <seb@redhat.com> [//]: # (If resolving an issue, uncomment and update the line below) [//]: # (Closes #[issue-number]) ## Test Plan Run: ``` ollama run llama3.2:3b-instruct-fp16 --keepalive 2m & uv run pytest -v -k "ollama" --inference-model=llama3.2:3b-instruct-fp16 llama_stack/providers/tests/inference/test_model_registration.py ``` You can observe that some tests pass while others fail, but the test runs successfully. [//]: # (## Documentation) [//]: # (- [ ] Added a Changelog entry if the change is significant) Signed-off-by: Sébastien Han <seb@redhat.com> Co-authored-by: Ashwin Bharambe <ashwin.bharambe@gmail.com>
This commit is contained in:
parent
61f43b8677
commit
69eebaf5bf
4 changed files with 787 additions and 263 deletions
|
@ -41,6 +41,7 @@ dependencies = [
|
|||
dev = [
|
||||
"pytest",
|
||||
"pytest-asyncio",
|
||||
"pytest-html",
|
||||
"nbval", # For notebook testing
|
||||
"black",
|
||||
"ruff",
|
||||
|
@ -49,7 +50,19 @@ dev = [
|
|||
"pre-commit",
|
||||
"uvicorn",
|
||||
"fastapi",
|
||||
"ruamel.yaml", # needed for openai generator
|
||||
"ruamel.yaml", # needed for openapi generator
|
||||
]
|
||||
test = [
|
||||
"openai",
|
||||
"aiosqlite",
|
||||
"ollama",
|
||||
"torch>=2.6.0",
|
||||
"fairscale>=0.4.13",
|
||||
"torchvision>=0.21.0",
|
||||
"lm-format-enforcer>=0.10.9",
|
||||
"groq",
|
||||
"opentelemetry-sdk",
|
||||
"opentelemetry-exporter-otlp-proto-http",
|
||||
]
|
||||
docs = [
|
||||
"sphinx-autobuild",
|
||||
|
@ -79,6 +92,10 @@ name = "pytorch-cpu"
|
|||
url = "https://download.pytorch.org/whl/cpu"
|
||||
explicit = true
|
||||
|
||||
[tool.uv.sources]
|
||||
torch = [{ index = "pytorch-cpu" }]
|
||||
torchvision = [{ index = "pytorch-cpu" }]
|
||||
|
||||
[tool.ruff]
|
||||
line-length = 120
|
||||
exclude = [
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue