llama-stack/pyproject.toml
2025-01-31 21:28:08 -08:00

61 lines
1.4 KiB
TOML

[build-system]
requires = ["setuptools>=61.0"]
build-backend = "setuptools.build_meta"
[project]
name = "llama_stack"
version = "0.1.0"
authors = [
{ name = "Meta Llama", email = "llama-oss@meta.com" },
]
description = "Llama Stack"
readme = "README.md"
requires-python = ">=3.10"
license = { text = "MIT" }
classifiers = [
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Operating System :: OS Independent",
"Intended Audience :: Developers",
"Intended Audience :: Information Technology",
"Intended Audience :: Science/Research",
"Topic :: Scientific/Engineering :: Artificial Intelligence",
"Topic :: Scientific/Engineering :: Information Analysis",
]
dependencies = [
"blobfile",
"fire",
"httpx",
"huggingface-hub",
"llama-models>=0.1.0",
"llama-stack-client>=0.1.0",
"prompt-toolkit",
"python-dotenv",
"pydantic>=2",
"requests",
"rich",
"setuptools",
"termcolor",
]
[project.optional-dependencies]
dev = [
"pytest",
"pytest-asyncio",
"pytest-nbval", # For notebook testing
"black",
"ruff",
"mypy",
"types-requests",
"types-setuptools",
]
[project.urls]
Homepage = "https://github.com/meta-llama/llama-stack"
[project.scripts]
llama = "llama_stack.cli.llama:main"
install-wheel-from-presigned = "llama_stack.cli.scripts.run:install_wheel_from_presigned"
[tool.setuptools]
packages = ["llama_stack"]