Commit graph

3089 commits

Author SHA1 Message Date
Ishaan Jaff
36be9967d1 fix storing request status in mem 2024-07-16 21:43:16 -07:00
Ishaan Jaff
86b311eeca fix set default value for max_file_size_mb 2024-07-16 21:43:16 -07:00
Ishaan Jaff
ac7849ee47 ui new build 2024-07-16 20:04:36 -07:00
Krrish Dholakia
ec03e675c9 fix(proxy/utils.py): fix failure logging for rejected requests. + unit tests 2024-07-16 17:15:20 -07:00
Vinnie Giarrusso
6ff863ee00 Add enabled_roles to Guardrails configuration, Update Lakera guardrail moderation hook 2024-07-16 01:52:08 -07:00
Ishaan Jaff
254ac37f65
Merge pull request #4724 from BerriAI/litellm_Set_max_file_size_transc
[Feat] - set max file size on /audio/transcriptions
2024-07-15 20:42:24 -07:00
Ishaan Jaff
af19a2aff3 ui new build 2024-07-15 20:09:17 -07:00
Ishaan Jaff
979b5d8eea
Merge pull request #4719 from BerriAI/litellm_fix_audio_transcript
[Fix] /audio/transcription - don't write to the local file system
2024-07-15 20:05:42 -07:00
Ishaan Jaff
bac6685bfc fix linting 2024-07-15 20:02:41 -07:00
Ishaan Jaff
38cef1c58d fix error from max file size 2024-07-15 19:57:33 -07:00
Ishaan Jaff
48d28e37a4 fix set max_file_size 2024-07-15 19:41:38 -07:00
Ishaan Jaff
b5a2090720 use helper to check check_file_size_under_limit 2024-07-15 19:40:05 -07:00
Ishaan Jaff
6c060b1fdc check_file_size_under_limit 2024-07-15 19:38:08 -07:00
Krrish Dholakia
959c627dd3 fix(litellm_logging.py): log response_cost=0 for failed calls
Fixes https://github.com/BerriAI/litellm/issues/4604
2024-07-15 19:25:56 -07:00
Krrish Dholakia
9cc2daeec9 fix(utils.py): update get_model_info docstring
Fixes https://github.com/BerriAI/litellm/issues/4711
2024-07-15 18:18:50 -07:00
Ishaan Jaff
a900f352b5 fix - don't write file.filename 2024-07-15 14:56:01 -07:00
Krrish Dholakia
e8e31c4029 docs(enterprise.md): cleanup docs 2024-07-15 14:52:08 -07:00
Ishaan Jaff
3dc2ec8119 fix show debugging utils on in mem usage 2024-07-15 10:05:57 -07:00
Krish Dholakia
6bf60d773e
Merge pull request #4696 from BerriAI/litellm_guardrail_logging_only
Allow setting `logging_only` in guardrails config
2024-07-13 21:50:43 -07:00
Krish Dholakia
7bc9a189e7
Merge branch 'main' into litellm_add_azure_ai_pricing 2024-07-13 21:50:26 -07:00
Krish Dholakia
d0fb685c56
Merge pull request #4706 from BerriAI/litellm_retry_after
Return `retry-after` header for rate limited requests
2024-07-13 21:37:41 -07:00
Krrish Dholakia
de8230ed41 fix(proxy_server.py): fix returning response headers on exception 2024-07-13 19:11:30 -07:00
Ishaan Jaff
4d7d6504b6
Merge pull request #4704 from BerriAI/litellm_debug_mem
[Debug-Utils] Add some useful memory usage debugging utils
2024-07-13 18:44:40 -07:00
Ishaan Jaff
ed5114c680
Merge pull request #4703 from BerriAI/litellm_only_use_internal_use_cache
[Fix Memory Usage] - only use per request tracking if slack alerting is being used
2024-07-13 18:40:22 -07:00
Ishaan Jaff
31783196c0 feat - return size of in memory cache 2024-07-13 18:22:44 -07:00
Ishaan Jaff
759e02bdaa debug mem issues show growth 2024-07-13 18:05:19 -07:00
Ishaan Jaff
69f74c1e6c fix only use per request tracking if slack alerting is being used 2024-07-13 18:01:53 -07:00
Krrish Dholakia
fde434be66 feat(proxy_server.py): return 'retry-after' param for rate limited requests
Closes https://github.com/BerriAI/litellm/issues/4695
2024-07-13 17:15:20 -07:00
Krrish Dholakia
bc9fe23ebf fix: cleanup 2024-07-13 16:36:04 -07:00
Krrish Dholakia
b1be355d42 build(model_prices_and_context_window.json): add azure ai jamba instruct pricing + token details
Adds jamba instruct, mistral, llama3 pricing + token info for azure_ai
2024-07-13 16:34:31 -07:00
Krish Dholakia
bc58e44d8f
Merge pull request #4701 from BerriAI/litellm_rpm_support_passthrough
Support key-rpm limits on pass-through endpoints
2024-07-13 15:22:29 -07:00
Krrish Dholakia
77325358b4 fix(pass_through_endpoints.py): fix client init 2024-07-13 14:46:56 -07:00
Ishaan Jaff
c1a9881d5c
Merge pull request #4697 from BerriAI/litellm_fix_sso_bug
[Fix] Bug - Clear user_id from cache when /user/update is called
2024-07-13 14:39:47 -07:00
Krrish Dholakia
7e769f3b89 fix: fix linting errors 2024-07-13 14:39:42 -07:00
Ishaan Jaff
fad37a969b ui new build 2024-07-13 14:38:13 -07:00
Krrish Dholakia
55e153556a test(test_pass_through_endpoints.py): add test for rpm limit support 2024-07-13 13:49:20 -07:00
Krrish Dholakia
0cc273d77b feat(pass_through_endpoint.py): support enforcing key rpm limits on pass through endpoints
Closes https://github.com/BerriAI/litellm/issues/4698
2024-07-13 13:29:44 -07:00
Ishaan Jaff
a447e4dd1a delete updated / deleted values from cache 2024-07-13 13:16:57 -07:00
Ishaan Jaff
893ed4e5f1 correctly clear cache when updating a user 2024-07-13 12:33:43 -07:00
Ishaan Jaff
bc91025307 use wrapper on /user endpoints 2024-07-13 12:29:15 -07:00
Krrish Dholakia
6b78e39600 feat(guardrails.py): allow setting logging_only in guardrails_config for presidio pii masking integration 2024-07-13 12:22:17 -07:00
Ishaan Jaff
670bf1b98d correctly flush cache when updating user 2024-07-13 12:05:09 -07:00
Krish Dholakia
66cedccd6b
Merge pull request #4686 from BerriAI/litellm_custom_chat_endpoints
docs(pass_through.md): Creating custom chat endpoints on proxy
2024-07-13 09:45:17 -07:00
Ishaan Jaff
70b96d12e9
Merge pull request #4685 from BerriAI/litellm_return_type_expired_key
[Fix] Proxy Return type=expire_key on expired Key errors
2024-07-12 18:52:51 -07:00
Krrish Dholakia
667fd2b376 docs(pass_through.md): add doc on creating custom chat endpoints on proxy
Allows developers to call proxy with anthropic sdk/boto3/etc.
2024-07-12 18:48:40 -07:00
Ishaan Jaff
57ced1d25e raise roxyErrorTypes.expired_key on expired key 2024-07-12 18:41:39 -07:00
Ishaan Jaff
34ff0a7e57 raise expired_key error 2024-07-12 18:39:00 -07:00
Ishaan Jaff
92bf98b30f
Merge pull request #4684 from BerriAI/litellm_safe_memory_mode
[Feat] Allow safe memory mode
2024-07-12 18:32:16 -07:00
Ishaan Jaff
24918c5041
Merge pull request #4682 from BerriAI/litellm_mem_leak_debug
show stack trace of 10 files taking up memory
2024-07-12 18:31:41 -07:00
Ishaan Jaff
cf5f11cc84
Merge pull request #4681 from BerriAI/litellm_mem_usage
[Fix] Reduce Mem Usage - only set ttl for requests to 2 mins
2024-07-12 18:31:19 -07:00