Commit graph

2899 commits

Author SHA1 Message Date
Krrish Dholakia
af2055c2b7 feat(aporio_ai.py): support aporio ai prompt injection for chat completion requests
Closes https://github.com/BerriAI/litellm/issues/2950
2024-07-17 16:38:47 -07:00
Krrish Dholakia
830c21b1fd fix(utils.py): return optional params from groq 2024-07-17 12:09:08 -07:00
Krrish Dholakia
04a646c321 fix(test_key_generate_prisma.py): pass user_api_key_dict to test call 2024-07-17 08:29:21 -07:00
Krrish Dholakia
ae78f00ba6 fix(team_endpoints.py): fix check 2024-07-16 22:05:48 -07:00
Krrish Dholakia
65e65e99a5 fix(internal_user_endpoints.py): delete associated invitation links before deleting user in /user/delete
Fixes https://github.com/BerriAI/litellm/issues/4740
2024-07-16 21:43:17 -07:00
Ishaan Jaff
a6e3a82e58 fix check if user passed custom header 2024-07-16 21:43:17 -07:00
Ishaan Jaff
ecc0163409 add example on how to use litellm_key_header_name 2024-07-16 21:43:17 -07:00
Ishaan Jaff
af151450cc feat - use custom api key name 2024-07-16 21:43:17 -07:00
Krrish Dholakia
6ffa77cd8f fix(team_endpoints.py): check if key belongs to team before returning /team/info 2024-07-16 21:43:17 -07:00
Ishaan Jaff
3203d825e0 fix calculate correct alerting threshold 2024-07-16 21:43:17 -07:00
Ishaan Jaff
f1a3fd99b7 fix tracking hanging requests 2024-07-16 21:43:16 -07:00
Ishaan Jaff
4eef814a35 fix storing request status in mem 2024-07-16 21:43:16 -07:00
Ishaan Jaff
c6d898ff44 fix set default value for max_file_size_mb 2024-07-16 21:43:16 -07:00
Ishaan Jaff
ed8e368ba3 ui new build 2024-07-16 20:04:36 -07:00
Krrish Dholakia
b022099712 fix(proxy/utils.py): fix failure logging for rejected requests. + unit tests 2024-07-16 17:15:20 -07:00
Ishaan Jaff
7944450074 Merge pull request #4724 from BerriAI/litellm_Set_max_file_size_transc
[Feat] - set max file size on /audio/transcriptions
2024-07-15 20:42:24 -07:00
Ishaan Jaff
b982436ece ui new build 2024-07-15 20:09:17 -07:00
Ishaan Jaff
c755522724 Merge pull request #4719 from BerriAI/litellm_fix_audio_transcript
[Fix] /audio/transcription - don't write to the local file system
2024-07-15 20:05:42 -07:00
Ishaan Jaff
470642c92c fix linting 2024-07-15 20:02:41 -07:00
Ishaan Jaff
e2096badbc fix error from max file size 2024-07-15 19:57:33 -07:00
Ishaan Jaff
20fe3a64b2 fix set max_file_size 2024-07-15 19:41:38 -07:00
Ishaan Jaff
57f0633b99 use helper to check check_file_size_under_limit 2024-07-15 19:40:05 -07:00
Ishaan Jaff
df6995f067 check_file_size_under_limit 2024-07-15 19:38:08 -07:00
Krrish Dholakia
4687b12732 fix(litellm_logging.py): log response_cost=0 for failed calls
Fixes https://github.com/BerriAI/litellm/issues/4604
2024-07-15 19:25:56 -07:00
Krrish Dholakia
cdfe1e6847 fix(utils.py): update get_model_info docstring
Fixes https://github.com/BerriAI/litellm/issues/4711
2024-07-15 18:18:50 -07:00
Ishaan Jaff
0b43d37f76 fix - don't write file.filename 2024-07-15 14:56:01 -07:00
Krrish Dholakia
7947cb5c2c docs(enterprise.md): cleanup docs 2024-07-15 14:52:08 -07:00
Ishaan Jaff
a9f77c2af3 fix show debugging utils on in mem usage 2024-07-15 10:05:57 -07:00
Krish Dholakia
808653ced6 Merge pull request #4696 from BerriAI/litellm_guardrail_logging_only
Allow setting `logging_only` in guardrails config
2024-07-13 21:50:43 -07:00
Krish Dholakia
38f61b50bd Merge branch 'main' into litellm_add_azure_ai_pricing 2024-07-13 21:50:26 -07:00
Krish Dholakia
833b5c4b34 Merge pull request #4706 from BerriAI/litellm_retry_after
Return `retry-after` header for rate limited requests
2024-07-13 21:37:41 -07:00
Krrish Dholakia
e0c2940f22 fix(proxy_server.py): fix returning response headers on exception 2024-07-13 19:11:30 -07:00
Ishaan Jaff
e3e649ea55 Merge pull request #4704 from BerriAI/litellm_debug_mem
[Debug-Utils] Add some useful memory usage debugging utils
2024-07-13 18:44:40 -07:00
Ishaan Jaff
186d8fafc5 Merge pull request #4703 from BerriAI/litellm_only_use_internal_use_cache
[Fix Memory Usage] - only use per request tracking if slack alerting is being used
2024-07-13 18:40:22 -07:00
Ishaan Jaff
668070a2bb feat - return size of in memory cache 2024-07-13 18:22:44 -07:00
Ishaan Jaff
4c7cc2208f debug mem issues show growth 2024-07-13 18:05:19 -07:00
Ishaan Jaff
36b24209eb fix only use per request tracking if slack alerting is being used 2024-07-13 18:01:53 -07:00
Krrish Dholakia
17635450cd feat(proxy_server.py): return 'retry-after' param for rate limited requests
Closes https://github.com/BerriAI/litellm/issues/4695
2024-07-13 17:15:20 -07:00
Krrish Dholakia
9b51b93980 fix: cleanup 2024-07-13 16:36:04 -07:00
Krrish Dholakia
219dbf4f36 build(model_prices_and_context_window.json): add azure ai jamba instruct pricing + token details
Adds jamba instruct, mistral, llama3 pricing + token info for azure_ai
2024-07-13 16:34:31 -07:00
Krish Dholakia
5439eb9ba3 Merge pull request #4701 from BerriAI/litellm_rpm_support_passthrough
Support key-rpm limits on pass-through endpoints
2024-07-13 15:22:29 -07:00
Krrish Dholakia
7a95132a03 fix(pass_through_endpoints.py): fix client init 2024-07-13 14:46:56 -07:00
Ishaan Jaff
8bdd4594fe Merge pull request #4697 from BerriAI/litellm_fix_sso_bug
[Fix] Bug - Clear user_id from cache when /user/update is called
2024-07-13 14:39:47 -07:00
Krrish Dholakia
4ca677638f fix: fix linting errors 2024-07-13 14:39:42 -07:00
Ishaan Jaff
02fe7cfb3a ui new build 2024-07-13 14:38:13 -07:00
Krrish Dholakia
e82616342b test(test_pass_through_endpoints.py): add test for rpm limit support 2024-07-13 13:49:20 -07:00
Krrish Dholakia
1d6643df22 feat(pass_through_endpoint.py): support enforcing key rpm limits on pass through endpoints
Closes https://github.com/BerriAI/litellm/issues/4698
2024-07-13 13:29:44 -07:00
Ishaan Jaff
b7387156c8 delete updated / deleted values from cache 2024-07-13 13:16:57 -07:00
Ishaan Jaff
f2b0929284 correctly clear cache when updating a user 2024-07-13 12:33:43 -07:00
Ishaan Jaff
8cce7d2df1 use wrapper on /user endpoints 2024-07-13 12:29:15 -07:00