Commit graph

2873 commits

Author SHA1 Message Date
Ishaan Jaff
0b43d37f76 fix - don't write file.filename 2024-07-15 14:56:01 -07:00
Ishaan Jaff
a9f77c2af3 fix show debugging utils on in mem usage 2024-07-15 10:05:57 -07:00
Krish Dholakia
808653ced6 Merge pull request #4696 from BerriAI/litellm_guardrail_logging_only
Allow setting `logging_only` in guardrails config
2024-07-13 21:50:43 -07:00
Krish Dholakia
38f61b50bd Merge branch 'main' into litellm_add_azure_ai_pricing 2024-07-13 21:50:26 -07:00
Krish Dholakia
833b5c4b34 Merge pull request #4706 from BerriAI/litellm_retry_after
Return `retry-after` header for rate limited requests
2024-07-13 21:37:41 -07:00
Krrish Dholakia
e0c2940f22 fix(proxy_server.py): fix returning response headers on exception 2024-07-13 19:11:30 -07:00
Ishaan Jaff
e3e649ea55 Merge pull request #4704 from BerriAI/litellm_debug_mem
[Debug-Utils] Add some useful memory usage debugging utils
2024-07-13 18:44:40 -07:00
Ishaan Jaff
186d8fafc5 Merge pull request #4703 from BerriAI/litellm_only_use_internal_use_cache
[Fix Memory Usage] - only use per request tracking if slack alerting is being used
2024-07-13 18:40:22 -07:00
Ishaan Jaff
668070a2bb feat - return size of in memory cache 2024-07-13 18:22:44 -07:00
Ishaan Jaff
4c7cc2208f debug mem issues show growth 2024-07-13 18:05:19 -07:00
Ishaan Jaff
36b24209eb fix only use per request tracking if slack alerting is being used 2024-07-13 18:01:53 -07:00
Krrish Dholakia
17635450cd feat(proxy_server.py): return 'retry-after' param for rate limited requests
Closes https://github.com/BerriAI/litellm/issues/4695
2024-07-13 17:15:20 -07:00
Krrish Dholakia
9b51b93980 fix: cleanup 2024-07-13 16:36:04 -07:00
Krrish Dholakia
219dbf4f36 build(model_prices_and_context_window.json): add azure ai jamba instruct pricing + token details
Adds jamba instruct, mistral, llama3 pricing + token info for azure_ai
2024-07-13 16:34:31 -07:00
Krish Dholakia
5439eb9ba3 Merge pull request #4701 from BerriAI/litellm_rpm_support_passthrough
Support key-rpm limits on pass-through endpoints
2024-07-13 15:22:29 -07:00
Krrish Dholakia
7a95132a03 fix(pass_through_endpoints.py): fix client init 2024-07-13 14:46:56 -07:00
Ishaan Jaff
8bdd4594fe Merge pull request #4697 from BerriAI/litellm_fix_sso_bug
[Fix] Bug - Clear user_id from cache when /user/update is called
2024-07-13 14:39:47 -07:00
Krrish Dholakia
4ca677638f fix: fix linting errors 2024-07-13 14:39:42 -07:00
Ishaan Jaff
02fe7cfb3a ui new build 2024-07-13 14:38:13 -07:00
Krrish Dholakia
e82616342b test(test_pass_through_endpoints.py): add test for rpm limit support 2024-07-13 13:49:20 -07:00
Krrish Dholakia
1d6643df22 feat(pass_through_endpoint.py): support enforcing key rpm limits on pass through endpoints
Closes https://github.com/BerriAI/litellm/issues/4698
2024-07-13 13:29:44 -07:00
Ishaan Jaff
b7387156c8 delete updated / deleted values from cache 2024-07-13 13:16:57 -07:00
Ishaan Jaff
f2b0929284 correctly clear cache when updating a user 2024-07-13 12:33:43 -07:00
Ishaan Jaff
8cce7d2df1 use wrapper on /user endpoints 2024-07-13 12:29:15 -07:00
Krrish Dholakia
6641683d66 feat(guardrails.py): allow setting logging_only in guardrails_config for presidio pii masking integration 2024-07-13 12:22:17 -07:00
Ishaan Jaff
2505dcf530 correctly flush cache when updating user 2024-07-13 12:05:09 -07:00
Krish Dholakia
e628171d82 Merge pull request #4686 from BerriAI/litellm_custom_chat_endpoints
docs(pass_through.md): Creating custom chat endpoints on proxy
2024-07-13 09:45:17 -07:00
Ishaan Jaff
f373ac90f6 Merge pull request #4685 from BerriAI/litellm_return_type_expired_key
[Fix] Proxy Return type=expire_key on expired Key errors
2024-07-12 18:52:51 -07:00
Krrish Dholakia
c8a2782df8 docs(pass_through.md): add doc on creating custom chat endpoints on proxy
Allows developers to call proxy with anthropic sdk/boto3/etc.
2024-07-12 18:48:40 -07:00
Ishaan Jaff
d76a09681f raise roxyErrorTypes.expired_key on expired key 2024-07-12 18:41:39 -07:00
Ishaan Jaff
f32ecd2238 raise expired_key error 2024-07-12 18:39:00 -07:00
Ishaan Jaff
1adff9cbd6 Merge pull request #4684 from BerriAI/litellm_safe_memory_mode
[Feat] Allow safe memory mode
2024-07-12 18:32:16 -07:00
Ishaan Jaff
fc7b1d78e2 Merge pull request #4682 from BerriAI/litellm_mem_leak_debug
show stack trace of 10 files taking up memory
2024-07-12 18:31:41 -07:00
Ishaan Jaff
ad93e940fc Merge pull request #4681 from BerriAI/litellm_mem_usage
[Fix] Reduce Mem Usage - only set ttl for requests to 2 mins
2024-07-12 18:31:19 -07:00
Ishaan Jaff
c43948545f feat add safe_memory_mode 2024-07-12 18:18:39 -07:00
Krrish Dholakia
d2a0977af7 feat(opentelemetry.py): support logging call metadata to otel 2024-07-12 15:41:34 -07:00
Ishaan Jaff
946d48d286 show stack trace of 10 files tking up memory 2024-07-12 15:33:03 -07:00
Ishaan Jaff
bc7b3f28b9 reduce ttil for update_request_status 2024-07-12 15:14:54 -07:00
Krrish Dholakia
d4ba87840e fix(proxy_server.py): fix linting errors 2024-07-11 22:12:33 -07:00
Krish Dholakia
35a17b7d99 Merge pull request #4669 from BerriAI/litellm_logging_only_masking
Flag for PII masking on Logging only
2024-07-11 22:03:37 -07:00
Krish Dholakia
8a4c428a7c Merge branch 'main' into litellm_call_id_in_response 2024-07-11 21:54:49 -07:00
Krish Dholakia
12e0f50812 Merge pull request #4651 from msabramo/docs-logging-cleanup
Docs: Miscellaneous cleanup of `docs/my-website/docs/proxy/logging.md`
2024-07-11 21:52:20 -07:00
Ishaan Jaff
2eef673ca8 ui new build 2024-07-11 19:13:08 -07:00
Krrish Dholakia
1a57e49e46 fix(presidio_pii_masking.py): support logging_only pii masking 2024-07-11 18:04:12 -07:00
Krrish Dholakia
abd682323c feat(guardrails): Flag for PII Masking on Logging
Fixes https://github.com/BerriAI/litellm/issues/4580
2024-07-11 16:09:34 -07:00
Ishaan Jaff
92228d9104 Merge pull request #4647 from msabramo/msabramo/remove-unnecessary-imports
Remove unnecessary imports
2024-07-11 15:07:30 -07:00
Ishaan Jaff
bf50c8e087 Merge pull request #4661 from BerriAI/litellm_fix_mh
[Fix] Model Hub - Show supports vision correctly
2024-07-11 15:03:37 -07:00
Ishaan Jaff
a16cd02cd9 fix supports vision 2024-07-11 12:59:42 -07:00
Krrish Dholakia
3f965df68b fix(llm_cost_calc/google.py): fix google embedding cost calculation
Fixes https://github.com/BerriAI/litellm/issues/4630
2024-07-11 11:55:48 -07:00
Ishaan Jaff
db7d417727 Merge pull request #4658 from BerriAI/litellm_check_otel_spans
[Test-Proxy] Otel Traces
2024-07-11 10:41:51 -07:00