Commit graph

9188 commits

Author SHA1 Message Date
Ishaan Jaff
dc5c72d04e router return get_deployment_by_model_group_name 2024-07-15 19:27:12 -07:00
Ishaan Jaff
11ed40be80 allow setting max_file_size_mb 2024-07-15 19:25:24 -07:00
Ishaan Jaff
0b43d37f76 fix - don't write file.filename 2024-07-15 14:56:01 -07:00
Ishaan Jaff
a9f77c2af3 fix show debugging utils on in mem usage 2024-07-15 10:05:57 -07:00
Krrish Dholakia
015d1a4fd9 fix(vertex_httpx.py): google search grounding fix 2024-07-14 08:06:17 -07:00
Krish Dholakia
808653ced6 Merge pull request #4696 from BerriAI/litellm_guardrail_logging_only
Allow setting `logging_only` in guardrails config
2024-07-13 21:50:43 -07:00
Krish Dholakia
38f61b50bd Merge branch 'main' into litellm_add_azure_ai_pricing 2024-07-13 21:50:26 -07:00
Krrish Dholakia
fdaa0443d3 test(test_presidio_pii_masking.py): fix presidio test 2024-07-13 21:44:22 -07:00
Krish Dholakia
833b5c4b34 Merge pull request #4706 from BerriAI/litellm_retry_after
Return `retry-after` header for rate limited requests
2024-07-13 21:37:41 -07:00
Krrish Dholakia
e0c2940f22 fix(proxy_server.py): fix returning response headers on exception 2024-07-13 19:11:30 -07:00
Ishaan Jaff
e3e649ea55 Merge pull request #4704 from BerriAI/litellm_debug_mem
[Debug-Utils] Add some useful memory usage debugging utils
2024-07-13 18:44:40 -07:00
Ishaan Jaff
186d8fafc5 Merge pull request #4703 from BerriAI/litellm_only_use_internal_use_cache
[Fix Memory Usage] - only use per request tracking if slack alerting is being used
2024-07-13 18:40:22 -07:00
Ishaan Jaff
668070a2bb feat - return size of in memory cache 2024-07-13 18:22:44 -07:00
Ishaan Jaff
4c7cc2208f debug mem issues show growth 2024-07-13 18:05:19 -07:00
Ishaan Jaff
36b24209eb fix only use per request tracking if slack alerting is being used 2024-07-13 18:01:53 -07:00
Krrish Dholakia
17635450cd feat(proxy_server.py): return 'retry-after' param for rate limited requests
Closes https://github.com/BerriAI/litellm/issues/4695
2024-07-13 17:15:20 -07:00
Krrish Dholakia
9b51b93980 fix: cleanup 2024-07-13 16:36:04 -07:00
Krrish Dholakia
219dbf4f36 build(model_prices_and_context_window.json): add azure ai jamba instruct pricing + token details
Adds jamba instruct, mistral, llama3 pricing + token info for azure_ai
2024-07-13 16:34:31 -07:00
Krish Dholakia
5439eb9ba3 Merge pull request #4701 from BerriAI/litellm_rpm_support_passthrough
Support key-rpm limits on pass-through endpoints
2024-07-13 15:22:29 -07:00
Ishaan Jaff
0cb2136c5c Merge pull request #4693 from BerriAI/litellm_bad_req_error_mapping
fix -  Raise `BadRequestError` when passing the wrong role
2024-07-13 15:05:54 -07:00
Krrish Dholakia
8240cf8997 test: test fixes 2024-07-13 15:04:13 -07:00
Krrish Dholakia
7a95132a03 fix(pass_through_endpoints.py): fix client init 2024-07-13 14:46:56 -07:00
Ishaan Jaff
8bdd4594fe Merge pull request #4697 from BerriAI/litellm_fix_sso_bug
[Fix] Bug - Clear user_id from cache when /user/update is called
2024-07-13 14:39:47 -07:00
Krrish Dholakia
4ca677638f fix: fix linting errors 2024-07-13 14:39:42 -07:00
Ishaan Jaff
02fe7cfb3a ui new build 2024-07-13 14:38:13 -07:00
Krrish Dholakia
e82616342b test(test_pass_through_endpoints.py): add test for rpm limit support 2024-07-13 13:49:20 -07:00
Krrish Dholakia
1d6643df22 feat(pass_through_endpoint.py): support enforcing key rpm limits on pass through endpoints
Closes https://github.com/BerriAI/litellm/issues/4698
2024-07-13 13:29:44 -07:00
Ishaan Jaff
e384bf254d fix test rules 2024-07-13 13:23:23 -07:00
Ishaan Jaff
b7387156c8 delete updated / deleted values from cache 2024-07-13 13:16:57 -07:00
Ishaan Jaff
411ebf8f85 test updating user role 2024-07-13 13:13:40 -07:00
Ishaan Jaff
f2b0929284 correctly clear cache when updating a user 2024-07-13 12:33:43 -07:00
Ishaan Jaff
8cce7d2df1 use wrapper on /user endpoints 2024-07-13 12:29:15 -07:00
Ishaan Jaff
7c489856e3 add doc string to explain what delete cache does 2024-07-13 12:25:31 -07:00
Krrish Dholakia
6641683d66 feat(guardrails.py): allow setting logging_only in guardrails_config for presidio pii masking integration 2024-07-13 12:22:17 -07:00
Ishaan Jaff
2505dcf530 correctly flush cache when updating user 2024-07-13 12:05:09 -07:00
Krrish Dholakia
d5f5415add fix(types/guardrails.py): add 'logging_only' param support 2024-07-13 11:44:37 -07:00
Krrish Dholakia
5147b8df13 build: re-run ci/cd 2024-07-13 11:41:35 -07:00
Ishaan Jaff
52370fee12 fix testing exception mapping 2024-07-13 11:10:13 -07:00
Ishaan Jaff
54c84d7139 fix exception raised in factory.py 2024-07-13 09:55:04 -07:00
Ishaan Jaff
31c61951f2 test - test_completion_bedrock_invalid_role_exception 2024-07-13 09:54:32 -07:00
Ishaan Jaff
1d99eeb188 fix str from BadRequestError 2024-07-13 09:54:32 -07:00
Ishaan Jaff
954410de64 fix exception raised in factory.py 2024-07-13 09:54:32 -07:00
Krish Dholakia
e628171d82 Merge pull request #4686 from BerriAI/litellm_custom_chat_endpoints
docs(pass_through.md): Creating custom chat endpoints on proxy
2024-07-13 09:45:17 -07:00
Ishaan Jaff
f2bdad3569 ci/cd run again 2024-07-12 19:08:59 -07:00
Ishaan Jaff
35824ab66a test_async_response_azure 2024-07-12 19:04:05 -07:00
Krrish Dholakia
fd16974996 fix(factory.py): handle message content being a list instead of string
Fixes https://github.com/BerriAI/litellm/issues/4679
2024-07-12 19:00:39 -07:00
Ishaan Jaff
f373ac90f6 Merge pull request #4685 from BerriAI/litellm_return_type_expired_key
[Fix] Proxy Return type=expire_key on expired Key errors
2024-07-12 18:52:51 -07:00
Krrish Dholakia
c8a2782df8 docs(pass_through.md): add doc on creating custom chat endpoints on proxy
Allows developers to call proxy with anthropic sdk/boto3/etc.
2024-07-12 18:48:40 -07:00
Ishaan Jaff
59bf9275b2 test expired key raises correct exception 2024-07-12 18:45:01 -07:00
Ishaan Jaff
d76a09681f raise roxyErrorTypes.expired_key on expired key 2024-07-12 18:41:39 -07:00