[tool.poetry] name = "litellm" version = "1.15.7" description = "Library to easily interface with LLM API providers" authors = ["BerriAI"] license = "MIT License" readme = "README.md" [tool.poetry.dependencies] python = "^3.8" openai = ">=1.0.0" python-dotenv = ">=0.2.0" tiktoken = ">=0.4.0" importlib-metadata = ">=6.8.0" tokenizers = "*" click = "*" jinja2 = "^3.1.2" certifi = "^2023.7.22" appdirs = "^1.4.4" aiohttp = "*" uvicorn = {version = "^0.24.0.post1", optional = true} fastapi = {version = "^0.104.1", optional = true} backoff = {version = "*", optional = true} rq = {version = "*", optional = true} [tool.poetry.extras] proxy = [ "uvicorn", "fastapi", "backoff", "rq", "orjson", "importlib", ] extra_proxy = [ "prisma", "azure-identity", "azure-keyvault-secrets" ] proxy_otel = [ "opentelemetry-api", "opentelemetry-sdk", "opentelemetry-exporter-otlp", "opentelemetry-sdk-resources" ] [tool.poetry.scripts] litellm = 'litellm:run_server' [build-system] requires = ["poetry-core", "wheel"] build-backend = "poetry.core.masonry.api" [tool.commitizen] version = "1.15.7" version_files = [ "pyproject.toml:^version" ]