mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
added a client extra, updated message
This commit is contained in:
parent
7d115d3b5c
commit
f6e15558bb
3 changed files with 13 additions and 2 deletions
|
|
@ -51,6 +51,11 @@ dependencies = [
|
||||||
"sqlalchemy[asyncio]>=2.0.41", # server - for conversations
|
"sqlalchemy[asyncio]>=2.0.41", # server - for conversations
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[project.optional-dependencies]
|
||||||
|
client = [
|
||||||
|
"llama-stack-client>=0.3.0", # Optional for library-only usage
|
||||||
|
]
|
||||||
|
|
||||||
[dependency-groups]
|
[dependency-groups]
|
||||||
dev = [
|
dev = [
|
||||||
"pytest>=8.4",
|
"pytest>=8.4",
|
||||||
|
|
|
||||||
|
|
@ -30,7 +30,7 @@ try:
|
||||||
)
|
)
|
||||||
except ImportError as e:
|
except ImportError as e:
|
||||||
raise ImportError(
|
raise ImportError(
|
||||||
"llama-stack-client is not installed. Please install it with `pip install llama-stack-client`."
|
"llama-stack-client is not installed. Please install it with `uv pip install llama-stack[client]`."
|
||||||
) from e
|
) from e
|
||||||
|
|
||||||
from pydantic import BaseModel, TypeAdapter
|
from pydantic import BaseModel, TypeAdapter
|
||||||
|
|
|
||||||
8
uv.lock
generated
8
uv.lock
generated
|
|
@ -1963,6 +1963,11 @@ dependencies = [
|
||||||
{ name = "uvicorn" },
|
{ name = "uvicorn" },
|
||||||
]
|
]
|
||||||
|
|
||||||
|
[package.optional-dependencies]
|
||||||
|
client = [
|
||||||
|
{ name = "llama-stack-client" },
|
||||||
|
]
|
||||||
|
|
||||||
[package.dev-dependencies]
|
[package.dev-dependencies]
|
||||||
benchmark = [
|
benchmark = [
|
||||||
{ name = "locust" },
|
{ name = "locust" },
|
||||||
|
|
@ -2089,7 +2094,7 @@ requires-dist = [
|
||||||
{ name = "httpx" },
|
{ name = "httpx" },
|
||||||
{ name = "jinja2", specifier = ">=3.1.6" },
|
{ name = "jinja2", specifier = ">=3.1.6" },
|
||||||
{ name = "jsonschema" },
|
{ name = "jsonschema" },
|
||||||
{ name = "llama-stack-client", specifier = ">=0.3.0" },
|
{ name = "llama-stack-client", marker = "extra == 'client'", specifier = ">=0.3.0" },
|
||||||
{ name = "openai", specifier = ">=2.5.0" },
|
{ name = "openai", specifier = ">=2.5.0" },
|
||||||
{ name = "opentelemetry-exporter-otlp-proto-http", specifier = ">=1.30.0" },
|
{ name = "opentelemetry-exporter-otlp-proto-http", specifier = ">=1.30.0" },
|
||||||
{ name = "opentelemetry-sdk", specifier = ">=1.30.0" },
|
{ name = "opentelemetry-sdk", specifier = ">=1.30.0" },
|
||||||
|
|
@ -2107,6 +2112,7 @@ requires-dist = [
|
||||||
{ name = "tiktoken" },
|
{ name = "tiktoken" },
|
||||||
{ name = "uvicorn", specifier = ">=0.34.0" },
|
{ name = "uvicorn", specifier = ">=0.34.0" },
|
||||||
]
|
]
|
||||||
|
provides-extras = ["client"]
|
||||||
|
|
||||||
[package.metadata.requires-dev]
|
[package.metadata.requires-dev]
|
||||||
benchmark = [{ name = "locust", specifier = ">=2.39.1" }]
|
benchmark = [{ name = "locust", specifier = ">=2.39.1" }]
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue