Commit graph

3565 commits

Author SHA1 Message Date
Ishaan Jaff
feb8c3c5b4
Merge pull request #5259 from BerriAI/litellm_return_remaining_tokens_in_header
[Feat] return `x-litellm-key-remaining-requests-{model}`: 1, `x-litellm-key-remaining-tokens-{model}: None` in response headers
2024-08-17 12:41:16 -07:00
Ishaan Jaff
ee0f772b5c feat return rmng tokens for model for api key 2024-08-17 12:35:10 -07:00
Krrish Dholakia
bc0023a409 feat(google_ai_studio_endpoints.py): support pass-through endpoint for all google ai studio requests
New Feature
2024-08-17 10:46:59 -07:00
Ishaan Jaff
5985c7e933 feat - use commong helper for getting model group 2024-08-17 10:46:04 -07:00
Ishaan Jaff
412d30d362 add litellm-key-remaining-tokens on prometheus 2024-08-17 10:02:20 -07:00
Ishaan Jaff
785482f023 feat add settings for rpm/tpm limits for a model 2024-08-17 09:16:01 -07:00
Krrish Dholakia
b56ecd7e02 fix(pass_through_endpoints.py): fix returned response headers for pass-through endpoitns 2024-08-17 09:00:00 -07:00
Krrish Dholakia
08411f37b4 docs(vertex_ai.md): cleanup docs 2024-08-17 08:38:01 -07:00
Krish Dholakia
f3e17cd692
Merge branch 'main' into litellm_log_model_price_information 2024-08-16 19:34:16 -07:00
Krish Dholakia
a8dd2b6910
Merge pull request #5244 from BerriAI/litellm_better_error_logging_sentry
refactor: replace .error() with .exception() logging for better debugging on sentry
2024-08-16 19:16:20 -07:00
Ishaan Jaff
1ee33478c9 track rpm/tpm usage per key+model 2024-08-16 18:28:58 -07:00
Ishaan Jaff
a6a4b944ad user api key auth rpm_limit_per_model 2024-08-16 18:22:35 -07:00
Krrish Dholakia
7fce6b0163 fix(health_check.py): return 'missing mode' error message, if error with health check, and mode is missing 2024-08-16 17:24:29 -07:00
Krrish Dholakia
178139f18d feat(litellm_logging.py): support logging model price information to s3 logs 2024-08-16 16:21:34 -07:00
Ishaan Jaff
ac833f415d docs oauh 2.0 enterprise feature 2024-08-16 14:00:24 -07:00
Ishaan Jaff
cd28b6607e add debugging for oauth2.0 2024-08-16 13:40:32 -07:00
Ishaan Jaff
d4b33cf87c allow using oauth2 checks for logging into proxy 2024-08-16 13:36:29 -07:00
Ishaan Jaff
0c0b835c3f add init commit for oauth 2 checks 2024-08-16 13:30:22 -07:00
Ishaan Jaff
9c3124c5a7 ui new build 2024-08-16 12:53:23 -07:00
Krrish Dholakia
cbdaecb5a8 fix(__init__.py): fix models_by_provider to include cohere_chat models
Fixes https://github.com/BerriAI/litellm/issues/5201
2024-08-16 11:33:23 -07:00
Krrish Dholakia
61f4b71ef7 refactor: replace .error() with .exception() logging for better debugging on sentry 2024-08-16 09:22:47 -07:00
Krrish Dholakia
1510daba4f bump: version 1.43.15 → 1.43.16 2024-08-15 23:04:30 -07:00
Krish Dholakia
6c3f37f8b4
Merge pull request #5235 from BerriAI/litellm_fix_s3_logs
fix(s3.py): fix s3 logging payload to have valid json values
2024-08-15 23:00:18 -07:00
Krish Dholakia
b3d15ace89
Merge branch 'main' into litellm_pass_through_endpoints_api 2024-08-15 22:39:19 -07:00
Krrish Dholakia
6fc6df134f feat(ui): for adding pass-through endpoints 2024-08-15 21:58:11 -07:00
Krrish Dholakia
589da45c24 feat(pass_through_endpoints.py): initial working CRUD endpoints for /pass_through_endoints 2024-08-15 21:23:26 -07:00
Ishaan Jaff
6cb3675a06 fix using prompt caching on proxy 2024-08-15 20:12:11 -07:00
Krrish Dholakia
cda50e5d47 fix(s3.py): fix s3 logging payload to have valid json values
Previously pydantic objects were being stringified, making them unparsable
2024-08-15 17:09:02 -07:00
Ishaan Jaff
5c1c9f7616 fix ImportError 2024-08-15 13:02:44 -07:00
Ishaan Jaff
5f693971f7 fix - don't require boto3 on the cli 2024-08-15 12:36:38 -07:00
Ishaan Jaff
7a17b2132f fix /moderations endpoint 2024-08-15 08:58:28 -07:00
Ishaan Jaff
c50a60004f fix test proxy exception mapping 2024-08-15 08:52:28 -07:00
Ishaan Jaff
58828403ea refactor use 1 util for llm routing 2024-08-15 08:42:20 -07:00
Ishaan Jaff
d50f26d73d simplify logic for routing llm request 2024-08-15 08:29:28 -07:00
Ishaan Jaff
fdd6664420 use route_request for making llm call 2024-08-15 08:16:44 -07:00
Krrish Dholakia
28faafadb1 feat(pass_through_endpoints.py): initial commit of crud endpoints for pass through endpoints 2024-08-14 21:36:07 -07:00
Krrish Dholakia
a020563149 feat(proxy_server.py): support returning available fields for pass_through_endpoints via `/config/field/list 2024-08-14 19:07:10 -07:00
Krish Dholakia
435d39bb24
Merge branch 'main' into litellm_azure_ai_openai_support 2024-08-14 17:53:27 -07:00
Krish Dholakia
aa2267bddb
Merge branch 'main' into litellm_fix_pass_through_endpoints 2024-08-14 14:59:38 -07:00
Krrish Dholakia
179dd7b893 docs(model_management.md): add section on adding additional model information to proxy config 2024-08-14 14:39:48 -07:00
Krrish Dholakia
583a3b330d fix(utils.py): support calling openai models via azure_ai/ 2024-08-14 13:41:04 -07:00
Krish Dholakia
22243c6571
Merge pull request #5176 from BerriAI/litellm_key_logging
Allow specifying langfuse project for logging in key metadata
2024-08-14 12:55:07 -07:00
Ishaan Jaff
bb877f6ead fix test for gcs bucket 2024-08-14 08:40:02 -07:00
Ishaan Jaff
63af2942ab feat log fail events on gcs 2024-08-14 08:39:16 -07:00
Krrish Dholakia
5fdbfcee44 fix(user_api_key_auth.py): more precisely expand scope to handle 'basic' tokens 2024-08-13 22:00:33 -07:00
Ishaan Jaff
e0978378c1 return detailed error message on check_valid_ip 2024-08-13 21:29:21 -07:00
Ishaan Jaff
09535b25f4 fix use s3 get_credentials to get boto3 creds 2024-08-13 21:18:06 -07:00
Ishaan Jaff
86818ddffc comment on using boto3 2024-08-13 20:28:40 -07:00
Krrish Dholakia
742e3cbccf feat(user_api_key_auth.py): support calling langfuse with litellm user_api_key_auth 2024-08-13 20:26:24 -07:00
Ishaan Jaff
4e7b0ce76e feat read config from s3 2024-08-13 20:18:59 -07:00