Commit graph

1648 commits

Author SHA1 Message Date
Krrish Dholakia
ecbb3c54c3 fix(utils.py): support get_max_tokens() call with same model_name as completion
Closes https://github.com/BerriAI/litellm/issues/3921
2024-05-31 21:37:51 -07:00
Krrish Dholakia
3896e3e88f fix: fix streaming with httpx client
prevent overwriting streams in parallel streaming calls
2024-05-31 10:55:18 -07:00
lj
f1fe41db74 Merge branch 'main' into fix-pydantic-warnings-again 2024-05-31 11:35:42 +08:00
Krish Dholakia
73e3dba2f6 Merge pull request #3928 from BerriAI/litellm_audio_speech_endpoint
feat(main.py): support openai tts endpoint
2024-05-30 17:30:42 -07:00
Krrish Dholakia
6b4153ff03 fix(main.py): add logging to audio_transcription calls 2024-05-30 16:57:11 -07:00
KX
ddb998fac1 fix: add missing seed parameter to ollama input
Current ollama interfacing does not allow for seed, which is supported in https://github.com/ollama/ollama/blob/main/docs/api.md#parameters and https://github.com/ollama/ollama/blob/main/docs/modelfile.md#valid-parameters-and-values

This resolves that by adding in handling of seed parameter.
2024-05-31 01:47:56 +08:00
Nir Gazit
8aebad9d25 Revert "Revert "fix: Log errors in Traceloop Integration (reverts previous revert)"" 2024-05-30 04:06:45 +03:00
Krish Dholakia
06ae6cad8d Revert "fix: Log errors in Traceloop Integration (reverts previous revert)" 2024-05-29 16:30:09 -07:00
Krish Dholakia
5063f0eab8 Merge pull request #3846 from nirga/revert-3831-revert-3780-traceloop-failures
fix: Log errors in Traceloop Integration (reverts previous revert)
2024-05-29 08:54:01 -07:00
Ishaan Jaff
000f23d005 Merge branch 'main' into litellm_show_openai_params_model_hub 2024-05-27 09:27:56 -07:00
Krrish Dholakia
23542fc1d2 fix(utils.py): support deepinfra optional params
Fixes https://github.com/BerriAI/litellm/issues/3855
2024-05-27 09:16:56 -07:00
Ishaan Jaff
50f1cbb1dd fix - return supported_openai_params from get_model_info 2024-05-27 09:00:12 -07:00
Krrish Dholakia
8e9a3fef81 feat(proxy_server.py): expose new /model_group/info endpoint
returns model-group level info on supported params, max tokens, pricing, etc.
2024-05-26 14:07:35 -07:00
Nir Gazit
5509e9f531 Revert "Revert "Log errors in Traceloop Integration"" 2024-05-26 12:01:10 +03:00
Ishaan Jaff
af82336cad Merge pull request #3824 from BerriAI/litellm_include_litellm_exception-in-error
[Feature]: Attach litellm exception in error string
2024-05-25 17:09:22 -07:00
Krrish Dholakia
b0afacf7e3 fix(proxy_server.py): fix model check for /v1/models endpoint when team has restricted access 2024-05-25 13:02:03 -07:00
Ishaan Jaff
b16c58d521 Revert "Log errors in Traceloop Integration" 2024-05-24 21:25:17 -07:00
Krish Dholakia
40791ee1f8 Merge pull request #3828 from BerriAI/litellm_outage_alerting
fix(slack_alerting.py): support region based outage alerting
2024-05-24 19:13:17 -07:00
Krrish Dholakia
4536ed6f6e feat(slack_alerting.py): refactor region outage alerting to do model based alerting instead
Unable to extract azure region from api base, makes sense to start with model alerting and then move to region
2024-05-24 19:10:33 -07:00
Krrish Dholakia
7368406c24 fix(slack_alerting.py): support region based outage alerting 2024-05-24 16:59:16 -07:00
Krish Dholakia
0f195d6b94 Merge pull request #3780 from nirga/traceloop-failures
Log errors in Traceloop Integration
2024-05-24 14:23:26 -07:00
Nir Gazit
43c30a4489 fix(traceloop): log errors 2024-05-24 22:05:31 +03:00
Ishaan Jaff
2b85d0faf9 feat - include litellm exception type when raising exception 2024-05-24 10:45:37 -07:00
Krish Dholakia
baa53d94f0 Merge pull request #3812 from afbarbaro/main
Fix issue with delta being None when Deferred / Async Content Filter is enabled on Azure OpenAI
2024-05-24 10:05:08 -07:00
Andres Barbaro
8dd4838d96 Fix issue with delta being None when Deferred / Async Content Filter is enabled on Azure OpenAI 2024-05-23 22:42:42 -05:00
Krrish Dholakia
c50074a0b7 feat(ui/model_dashboard.tsx): add databricks models via admin ui 2024-05-23 20:28:54 -07:00
Krish Dholakia
edb349a9ab Merge pull request #3808 from BerriAI/litellm_databricks_api
feat(databricks.py): adds databricks support - completion, async, streaming
2024-05-23 19:23:19 -07:00
Krrish Dholakia
e3c5e004c5 feat(databricks.py): add embedding model support 2024-05-23 18:22:03 -07:00
Krrish Dholakia
143a44823a feat(databricks.py): adds databricks support - completion, async, streaming
Closes https://github.com/BerriAI/litellm/issues/2160
2024-05-23 16:29:46 -07:00
Ishaan Jaff
bff4227f6a feat - add prixing for vertex_ai image gen 2024-05-23 16:27:08 -07:00
Krish Dholakia
10e1b43751 Merge branch 'main' into litellm_filter_invalid_params 2024-05-21 20:42:21 -07:00
Krrish Dholakia
c989b92801 feat(router.py): Fixes https://github.com/BerriAI/litellm/issues/3769 2024-05-21 17:24:51 -07:00
Krrish Dholakia
413be6d805 fix(utils.py): filter out hf eos token
Closes https://github.com/BerriAI/litellm/issues/3757
2024-05-21 14:31:54 -07:00
alisalim17
fe0e600062 Revert "Revert "Logfire Integration""
This reverts commit b04a8d878a.
2024-05-21 11:07:40 +04:00
Krish Dholakia
db77e41833 Merge pull request #3740 from BerriAI/litellm_return_rejected_response
feat(proxy_server.py): allow admin to return rejected response as string to user
2024-05-20 17:48:21 -07:00
Ishaan Jaff
7e6c9274fc Merge branch 'main' into litellm_standardize_slack_exception_msg_format 2024-05-20 16:39:41 -07:00
Ishaan Jaff
dc55a57d8a Merge pull request #3716 from BerriAI/litellm_set_cooldown_time_based_on_exception_header
[Feat] Router/ Proxy - set cooldown_time based on Azure exception headers
2024-05-20 16:34:12 -07:00
Ishaan Jaff
233828e16f fix - standardize slack alerting format 2024-05-20 16:26:11 -07:00
Ishaan Jaff
28d1bde250 Merge pull request #3739 from BerriAI/litellm_add_imagen_support
[FEAT] Async VertexAI Image Generation
2024-05-20 14:14:43 -07:00
Ishaan Jaff
883a9eb69a add parameter mapping with vertex ai 2024-05-20 13:28:20 -07:00
Krrish Dholakia
d4d4550bb6 fix(proxy_server.py): fixes for making rejected responses work with streaming 2024-05-20 12:32:19 -07:00
Ishaan Jaff
8ea3482676 fix python3.8 error 2024-05-20 10:55:10 -07:00
Ishaan Jaff
76a1444621 add ImageObject 2024-05-20 10:45:37 -07:00
Krrish Dholakia
45fedb83c6 feat(proxy_server.py): allow admin to return rejected response as string to user
Closes https://github.com/BerriAI/litellm/issues/3671
2024-05-20 10:30:23 -07:00
Krrish Dholakia
7da925cada fix(utils.py): drop response_format if 'drop_params=True' for gpt-4 2024-05-18 13:02:48 -07:00
Krrish Dholakia
ab747c8fe9 fix(utils.py): fix replicate completion cost calculation 2024-05-17 22:18:57 -07:00
Ishaan Jaff
cdfa9c9232 fix - cooldown based on exception header 2024-05-17 18:52:45 -07:00
Krish Dholakia
60615f46c4 Merge branch 'main' into litellm_bedrock_anthropic_fix 2024-05-17 17:47:32 -07:00
Krrish Dholakia
b20f4f65b4 fix(bedrock_httpx.py): raise better timeout exception 2024-05-17 17:16:36 -07:00
Krrish Dholakia
54b4e24427 fix(utils.py): exception map bedrock error 2024-05-17 16:18:25 -07:00