Commit graph

3040 commits

Author SHA1 Message Date
Ishaan Jaff
d208dedb35 (ci/cd) run again 2024-05-15 17:39:21 -07:00
Ishaan Jaff
240b183d7a ci/cd run again 2024-05-15 17:31:14 -07:00
Ishaan Jaff
ed0a815c2b test - exceptions predibase 2024-05-15 16:53:41 -07:00
Ishaan Jaff
f138c15859 (ci/cd) fix test_vertex_ai_stream 2024-05-15 16:32:40 -07:00
Ishaan Jaff
f2e8b2500f fix function calling mistral large latest 2024-05-15 16:05:17 -07:00
Ishaan Jaff
e518b1e6c1 fix - vertex exception test 2024-05-15 15:37:59 -07:00
Ishaan Jaff
6d8ea641ec (ci/cd) fix test_content_policy_exception_azure 2024-05-15 14:47:39 -07:00
Ishaan Jaff
371043d683 fix - test mistral/large _parallel_function_call 2024-05-15 14:31:00 -07:00
Ishaan Jaff
3e831b4e1a fix debug logs on router test 2024-05-15 14:28:17 -07:00
Ishaan Jaff
fdf7a4d8c8 fix - test_lowest_latency_routing_first_pick 2024-05-15 14:24:13 -07:00
Ishaan Jaff
5177e4408e
Merge pull request #3651 from BerriAI/litellm_improve_load_balancing
[Feat] Proxy + router - don't cooldown on 4XX error that are not 429, 408, 401
2024-05-15 10:24:34 -07:00
Ishaan Jaff
ae80148c12 test - router cooldowns 2024-05-15 09:43:30 -07:00
Krrish Dholakia
f43da3597d test: fix test 2024-05-15 08:51:40 -07:00
Krrish Dholakia
1840919ebd fix(main.py): testing fix 2024-05-15 08:23:00 -07:00
Krrish Dholakia
8117af664c fix(huggingface_restapi.py): fix task extraction from model name 2024-05-15 07:28:19 -07:00
Krrish Dholakia
900bb9aba8 test(test_token_counter.py): fix load test 2024-05-15 07:12:43 -07:00
Ishaan Jaff
0bac40b0f2 ci/cd run again 2024-05-14 21:53:14 -07:00
Ishaan Jaff
6290de36df (ci/cd) run again 2024-05-14 21:39:09 -07:00
Krrish Dholakia
73b6b5e804 test(test_token_counter.py): fix token counting test 2024-05-14 21:35:28 -07:00
Ishaan Jaff
faa58c7938 (ci/cd) run again 2024-05-14 20:45:07 -07:00
Ishaan Jaff
6d1ae5b9c4 (ci/cd) run again 2024-05-14 20:18:12 -07:00
Krrish Dholakia
298fd9b25c fix(main.py): ignore model_config param 2024-05-14 19:03:17 -07:00
Krrish Dholakia
a1dd341ca1 fix(utils.py): default claude-3 to tiktoken (0.8s faster than hf tokenizer) 2024-05-14 18:37:14 -07:00
Krish Dholakia
b04a8d878a
Revert "Logfire Integration" 2024-05-14 17:38:47 -07:00
Krrish Dholakia
c0d701a51e test(test_config.py): fix linting error 2024-05-14 17:32:31 -07:00
Krrish Dholakia
4e30d7cf5e build(model_prices_and_context_window.json): add gemini 1.5 flash model info 2024-05-14 17:30:31 -07:00
Ishaan Jaff
aa1615c757
Merge pull request #3626 from BerriAI/litellm_reset_spend_per_team_api_key
feat - reset spend per team, api_key [Only Master Key]
2024-05-14 11:49:07 -07:00
Krish Dholakia
adaafd72be
Merge pull request #3599 from taralika/patch-1
Ignore 0 failures and 0s latency in daily slack reports
2024-05-14 11:47:46 -07:00
alisalim17
765c382b2a Merge remote-tracking branch 'upstream/main' 2024-05-14 22:32:57 +04:00
Ishaan Jaff
ca41e6590e test - auth on /reset/spend 2024-05-14 11:28:35 -07:00
Krrish Dholakia
7557b3e2ff fix(init.py): set 'default_fallbacks' as a litellm_setting 2024-05-14 11:15:53 -07:00
Ishaan Jaff
0c8f5e5649
Merge pull request #3266 from antonioloison/litellm_add_disk_cache
[Feature] Add cache to disk
2024-05-14 09:24:01 -07:00
alisalim17
18bf68298f Merge remote-tracking branch 'upstream/main' 2024-05-14 18:42:20 +04:00
Anand Taralika
bd2e4cdfe0 Fixed the test alert sequence
Also fixed the issue that MagicMock does not create asynchronous mocks by default.
2024-05-13 22:43:12 -07:00
Krish Dholakia
2c867ea9a5
Merge pull request #3600 from msabramo/msabramo/fix-pydantic-warnings
Update pydantic code to fix warnings
2024-05-13 22:00:39 -07:00
sumanth
71e0294485 addressed comments 2024-05-14 10:05:19 +05:30
Anand Taralika
3b0c7d2321
Merge branch 'BerriAI:main' into patch-1 2024-05-13 21:31:00 -07:00
Anand Taralika
30332a6d68
Added tests for ignoring 0 metrics when alerting 2024-05-13 21:29:52 -07:00
Ishaan Jaff
ffc637969b (ci/cd) run again 2024-05-13 21:07:12 -07:00
Ishaan Jaff
5de31e9318 (ci/cd) run again 2024-05-13 20:54:50 -07:00
Ishaan Jaff
9bde3ccd1d (ci/cd) fixes 2024-05-13 20:49:02 -07:00
SUMANTH
978672a56d
Merge branch 'BerriAI:main' into usage-based-routing-ttl-on-cache 2024-05-14 09:08:01 +05:30
Ishaan Jaff
da0e10954a (ci/cd) run again 2024-05-13 20:31:59 -07:00
Krrish Dholakia
071a70c5fc test: fix watsonx api error 2024-05-13 19:01:19 -07:00
Krrish Dholakia
724d880a45 test(test_completion.py): handle async watsonx call fail 2024-05-13 18:40:51 -07:00
Krrish Dholakia
d4123951d9 test: handle watsonx rate limit error 2024-05-13 18:27:39 -07:00
Krrish Dholakia
29449aa5c1 fix(utils.py): fix watsonx exception mapping 2024-05-13 18:13:13 -07:00
Krrish Dholakia
38988f030a fix(router.py): fix typing 2024-05-13 18:06:10 -07:00
Krrish Dholakia
5488bf4921 feat(router.py): enable default fallbacks
allow user to define a generic list of fallbacks, in case a new deployment is bad

Closes https://github.com/BerriAI/litellm/issues/3623
2024-05-13 17:49:56 -07:00
Marc Abramowitz
bdb995f436 Revert = None changes
because @krrishdholakia said [here](https://github.com/BerriAI/litellm/pull/3600#discussion_r1599108453)
that they were problematic in the past.
2024-05-13 16:49:36 -07:00