Commit graph

27 commits

Author SHA1 Message Date
Krrish Dholakia
ea96eebe85 refactor: move all testing to top-level of repo
Closes https://github.com/BerriAI/litellm/issues/486
2024-09-28 21:08:14 -07:00
Ishaan Jaff
cad944d031 [Fix proxy perf] Use correct cache key when reading from redis cache (#5928)
* fix parallel request limiter use correct user id

* async def get_user_object(
fix

* use safe get_internal_user_object

* fix store internal users in redis correctly
2024-09-26 18:13:35 -07:00
Ishaan Jaff
4d253e473a [Feat] Improve OTEL Tracking - Require all Redis Cache reads to be logged on OTEL (#5881)
* fix use previous internal usage caching logic

* fix test_dual_cache_uses_redis

* redis track event_metadata in service logging

* show otel error on _get_parent_otel_span_from_kwargs

* track parent otel span on internal usage cache

* update_request_status

* fix internal usage cache

* fix linting

* fix test internal usage cache

* fix linting error

* show event metadata in redis set

* fix test_get_team_redis

* fix test_get_team_redis

* test_proxy_logging_setup
2024-09-25 10:57:08 -07:00
Ishaan Jaff
64398fe123 fix flaky tests 2024-08-27 22:47:11 -07:00
Ishaan Jaff
94e74b9ede inly write model tpm/rpm tracking when user set it 2024-08-18 09:58:09 -07:00
Ishaan Jaff
45cc7291f5 fix parallel request limiter tests 2024-08-17 14:21:59 -07:00
Ishaan Jaff
cf807d1541 fix parallel request limiter 2024-08-17 14:14:12 -07:00
Ishaan Jaff
8ae626b31f feat add settings for rpm/tpm limits for a model 2024-08-17 09:16:01 -07:00
Ishaan Jaff
824ea32452 track rpm/tpm usage per key+model 2024-08-16 18:28:58 -07:00
Krrish Dholakia
17635450cd feat(proxy_server.py): return 'retry-after' param for rate limited requests
Closes https://github.com/BerriAI/litellm/issues/4695
2024-07-13 17:15:20 -07:00
Krrish Dholakia
b8c6f3dd61 test(test_parallel_request_limiter.py): fix test 2024-06-13 17:13:44 -07:00
Krrish Dholakia
56fd0c60d1 fix(parallel_request_limiter.py): fix user+team tpm/rpm limit check
Closes https://github.com/BerriAI/litellm/issues/3788
2024-05-27 08:48:23 -07:00
Krrish Dholakia
3f339cb694 fix(parallel_request_limiter.py): fix max parallel request limiter on retries 2024-05-15 20:16:11 -07:00
Krrish Dholakia
29368b5d71 test: refactor testing to handle the hash token fix 2024-04-17 17:31:39 -07:00
Krrish Dholakia
6346839574 test(test_parallel_request_limiter.py): add more verbose logging 2024-03-06 19:21:57 -08:00
Krrish Dholakia
7fff5119de fix(parallel_request_limiter.py): fix team rate limit enforcement 2024-02-26 18:06:13 -08:00
ishaan-jaff
74d66d5ac5 (feat) tpm/rpm limit by User 2024-02-22 18:44:03 -08:00
Krrish Dholakia
07aa05bf17 fix(test_parallel_request_limiter.py): use mock responses for streaming 2024-02-08 21:45:38 -08:00
ishaan-jaff
f33815aaf8 (fix) test_normal_router_tpm_limit 2024-02-06 12:44:30 -08:00
ishaan-jaff
de4d36cf96 (fix) test_normal_router_tpm_limit 2024-02-06 06:46:49 -08:00
ishaan-jaff
f9916cc6e4 (fix) test_parallel limiter fix 2024-02-03 21:03:15 -08:00
Krrish Dholakia
6bdb081e2c test(test_parallel_request_limiter.py): fix test to handle minute changes 2024-02-03 20:39:31 -08:00
Krrish Dholakia
0302e4b27d test(test_parallel_request_limiter.py): fix test import 2024-02-02 10:15:33 -08:00
Krrish Dholakia
c91ab81fde fix(test_parallel_request_limiter): increase time limit for waiting for success logging event to happen 2024-01-30 13:26:17 -08:00
Krrish Dholakia
3957a8303a test(test_parallel_request_limiter.py): unit testing for tpm/rpm rate limits 2024-01-18 15:25:29 -08:00
Krrish Dholakia
13b013b28d feat(parallel_request_limiter.py): add support for tpm/rpm limits 2024-01-18 13:52:15 -08:00
Krrish Dholakia
44553bcc3a fix(parallel_request_limiter.py): decrement count for failed llm calls
https://github.com/BerriAI/litellm/issues/1477
2024-01-18 12:42:14 -08:00