forked from phoenix/litellm-mirror
* fix(factory.py): ensure tool call converts image url Fixes https://github.com/BerriAI/litellm/issues/6953 * fix(transformation.py): support mp4 + pdf url's for vertex ai Fixes https://github.com/BerriAI/litellm/issues/6936 * fix(http_handler.py): mask gemini api key in error logs Fixes https://github.com/BerriAI/litellm/issues/6963 * docs(prometheus.md): update prometheus FAQs * feat(auth_checks.py): ensure specific model access > wildcard model access if wildcard model is in access group, but specific model is not - deny access * fix(auth_checks.py): handle auth checks for team based model access groups handles scenario where model access group used for wildcard models * fix(internal_user_endpoints.py): support adding guardrails on `/user/update` Fixes https://github.com/BerriAI/litellm/issues/6942 * fix(key_management_endpoints.py): fix prepare_metadata_fields helper * fix: fix tests * build(requirements.txt): bump openai dep version fixes proxies argument * test: fix tests * fix(http_handler.py): fix error message masking * fix(bedrock_guardrails.py): pass in prepped data * test: fix test * test: fix nvidia nim test * fix(http_handler.py): return original response headers * fix: revert maskedhttpstatuserror * test: update tests * test: cleanup test * fix(key_management_endpoints.py): fix metadata field update logic * fix(key_management_endpoints.py): maintain initial order of guardrails in key update * fix(key_management_endpoints.py): handle prepare metadata * fix: fix linting errors * fix: fix linting errors * fix: fix linting errors * fix: fix key management errors * fix(key_management_endpoints.py): update metadata * test: update test * refactor: add more debug statements * test: skip flaky test * test: fix test * fix: fix test * fix: fix update metadata logic * fix: fix test * ci(config.yml): change db url for e2e ui testing
50 lines
No EOL
1.8 KiB
Text
50 lines
No EOL
1.8 KiB
Text
# LITELLM PROXY DEPENDENCIES #
|
|
anyio==4.4.0 # openai + http req.
|
|
openai==1.55.3 # openai req.
|
|
fastapi==0.111.0 # server dep
|
|
backoff==2.2.1 # server dep
|
|
pyyaml==6.0.0 # server dep
|
|
uvicorn==0.29.0 # server dep
|
|
gunicorn==22.0.0 # server dep
|
|
boto3==1.34.34 # aws bedrock/sagemaker calls
|
|
redis==5.0.0 # caching
|
|
numpy==1.24.3 # semantic caching
|
|
prisma==0.11.0 # for db
|
|
mangum==0.17.0 # for aws lambda functions
|
|
pynacl==1.5.0 # for encrypting keys
|
|
google-cloud-aiplatform==1.47.0 # for vertex ai calls
|
|
anthropic[vertex]==0.21.3
|
|
google-generativeai==0.5.0 # for vertex ai calls
|
|
async_generator==1.10.0 # for async ollama calls
|
|
langfuse==2.45.0 # for langfuse self-hosted logging
|
|
prometheus_client==0.20.0 # for /metrics endpoint on proxy
|
|
orjson==3.9.15 # fast /embedding responses
|
|
apscheduler==3.10.4 # for resetting budget in background
|
|
fastapi-sso==0.10.0 # admin UI, SSO
|
|
pyjwt[crypto]==2.9.0
|
|
python-multipart==0.0.9 # admin UI
|
|
Pillow==10.3.0
|
|
azure-ai-contentsafety==1.0.0 # for azure content safety
|
|
azure-identity==1.16.1 # for azure content safety
|
|
opentelemetry-api==1.25.0
|
|
opentelemetry-sdk==1.25.0
|
|
opentelemetry-exporter-otlp==1.25.0
|
|
sentry_sdk==2.2.1 # for sentry error handling
|
|
detect-secrets==1.5.0 # Enterprise - secret detection / masking in LLM requests
|
|
cryptography==42.0.7
|
|
|
|
### LITELLM PACKAGE DEPENDENCIES
|
|
python-dotenv==1.0.0 # for env
|
|
tiktoken==0.7.0 # for calculating usage
|
|
importlib-metadata==6.8.0 # for random utils
|
|
tokenizers==0.14.0 # for calculating usage
|
|
click==8.1.7 # for proxy cli
|
|
jinja2==3.1.4 # for prompt templates
|
|
certifi==2024.7.4 # [TODO] clean up
|
|
aiohttp==3.10.2 # for network calls
|
|
aioboto3==12.3.0 # for async sagemaker calls
|
|
tenacity==8.2.3 # for retrying requests, when litellm.num_retries set
|
|
pydantic==2.7.1 # proxy + openai req.
|
|
jsonschema==4.22.0 # validating json schema
|
|
websockets==10.4 # for realtime API
|
|
#### |