Commit graph

1508 commits

Author SHA1 Message Date
Ishaan Jaff
3b823c7587 fix - by default log raw curl command on langfuse 2024-06-04 16:30:25 -07:00
Krrish Dholakia
52a2f5150c fix(utils.py): fix cost calculation for openai-compatible streaming object 2024-06-04 10:36:25 -07:00
Krrish Dholakia
7b474ec267 fix(utils.py): add coverage for azure img gen content policy violation error 2024-06-04 08:29:30 -07:00
Krrish Dholakia
1de5235ba0 fix(router.py): use litellm.request_timeout as default for router clients 2024-06-03 14:19:53 -07:00
Ishaan Jaff
dd7d0a2895
Merge pull request #3983 from BerriAI/litellm_log_request_boddy_langfuse
[Feat] Log Raw Request from LiteLLM on Langfuse - when `"log_raw_request": true`
2024-06-03 13:42:06 -07:00
Ishaan Jaff
857ceb40bc feat - log raw_request to langfuse / other logging providers 2024-06-03 07:53:52 -07:00
Krrish Dholakia
ea30359b38 fix(utils.py): handle else block for get optional params 2024-06-03 07:45:44 -07:00
Krrish Dholakia
9ef83126d7 fix(utils.py): correctly instrument passing through api version in optional param check 2024-06-01 19:31:52 -07:00
Krrish Dholakia
7efac4d36c fix(azure.py): support dropping 'tool_choice=required' for older azure API versions
Closes https://github.com/BerriAI/litellm/issues/3876
2024-06-01 18:44:50 -07:00
Krish Dholakia
e7ff3adc26
Merge pull request #3944 from BerriAI/litellm_fix_parallel_streaming
fix: fix streaming with httpx client
2024-05-31 21:42:37 -07:00
Krrish Dholakia
7523f803d2 fix(utils.py): support get_max_tokens() call with same model_name as completion
Closes https://github.com/BerriAI/litellm/issues/3921
2024-05-31 21:37:51 -07:00
Krrish Dholakia
93c3635b64 fix: fix streaming with httpx client
prevent overwriting streams in parallel streaming calls
2024-05-31 10:55:18 -07:00
lj
27ed72405b
Merge branch 'main' into fix-pydantic-warnings-again 2024-05-31 11:35:42 +08:00
Krish Dholakia
d3a247bf20
Merge pull request #3928 from BerriAI/litellm_audio_speech_endpoint
feat(main.py): support openai tts endpoint
2024-05-30 17:30:42 -07:00
Krrish Dholakia
d65b7fe01b fix(main.py): add logging to audio_transcription calls 2024-05-30 16:57:11 -07:00
KX
d3921a3d28 fix: add missing seed parameter to ollama input
Current ollama interfacing does not allow for seed, which is supported in https://github.com/ollama/ollama/blob/main/docs/api.md#parameters and https://github.com/ollama/ollama/blob/main/docs/modelfile.md#valid-parameters-and-values

This resolves that by adding in handling of seed parameter.
2024-05-31 01:47:56 +08:00
Nir Gazit
b8d97c688c Revert "Revert "fix: Log errors in Traceloop Integration (reverts previous revert)"" 2024-05-30 04:06:45 +03:00
Krish Dholakia
77cc9cded9
Revert "fix: Log errors in Traceloop Integration (reverts previous revert)" 2024-05-29 16:30:09 -07:00
Krish Dholakia
c76deb8f76
Merge pull request #3846 from nirga/revert-3831-revert-3780-traceloop-failures
fix: Log errors in Traceloop Integration (reverts previous revert)
2024-05-29 08:54:01 -07:00
Ishaan Jaff
75222d7d4b
Merge branch 'main' into litellm_show_openai_params_model_hub 2024-05-27 09:27:56 -07:00
Krrish Dholakia
f0f853b941 fix(utils.py): support deepinfra optional params
Fixes https://github.com/BerriAI/litellm/issues/3855
2024-05-27 09:16:56 -07:00
Ishaan Jaff
245990597e fix - return supported_openai_params from get_model_info 2024-05-27 09:00:12 -07:00
Krrish Dholakia
22b6b99b34 feat(proxy_server.py): expose new /model_group/info endpoint
returns model-group level info on supported params, max tokens, pricing, etc.
2024-05-26 14:07:35 -07:00
Nir Gazit
7602c6f436 Revert "Revert "Log errors in Traceloop Integration"" 2024-05-26 12:01:10 +03:00
Ishaan Jaff
0ae6b337a3
Merge pull request #3824 from BerriAI/litellm_include_litellm_exception-in-error
[Feature]: Attach litellm exception in error string
2024-05-25 17:09:22 -07:00
Krrish Dholakia
25a2f00db6 fix(proxy_server.py): fix model check for /v1/models endpoint when team has restricted access 2024-05-25 13:02:03 -07:00
Ishaan Jaff
0083776a14
Revert "Log errors in Traceloop Integration" 2024-05-24 21:25:17 -07:00
Krish Dholakia
d25ed9c4d3
Merge pull request #3828 from BerriAI/litellm_outage_alerting
fix(slack_alerting.py): support region based outage alerting
2024-05-24 19:13:17 -07:00
Krrish Dholakia
8dec87425e feat(slack_alerting.py): refactor region outage alerting to do model based alerting instead
Unable to extract azure region from api base, makes sense to start with model alerting and then move to region
2024-05-24 19:10:33 -07:00
Krrish Dholakia
f8350b9461 fix(slack_alerting.py): support region based outage alerting 2024-05-24 16:59:16 -07:00
Krish Dholakia
bcd7534153
Merge pull request #3780 from nirga/traceloop-failures
Log errors in Traceloop Integration
2024-05-24 14:23:26 -07:00
Nir Gazit
7c279e05a7 fix(traceloop): log errors 2024-05-24 22:05:31 +03:00
Ishaan Jaff
33632df4c5 feat - include litellm exception type when raising exception 2024-05-24 10:45:37 -07:00
Krish Dholakia
391a31c0ce
Merge pull request #3812 from afbarbaro/main
Fix issue with delta being None when Deferred / Async Content Filter is enabled on Azure OpenAI
2024-05-24 10:05:08 -07:00
Andres Barbaro
b002328c8d Fix issue with delta being None when Deferred / Async Content Filter is enabled on Azure OpenAI 2024-05-23 22:42:42 -05:00
Krrish Dholakia
f04e4b921b feat(ui/model_dashboard.tsx): add databricks models via admin ui 2024-05-23 20:28:54 -07:00
Krish Dholakia
c14584722e
Merge pull request #3808 from BerriAI/litellm_databricks_api
feat(databricks.py): adds databricks support - completion, async, streaming
2024-05-23 19:23:19 -07:00
Krrish Dholakia
43353c28b3 feat(databricks.py): add embedding model support 2024-05-23 18:22:03 -07:00
Krrish Dholakia
d2229dcd21 feat(databricks.py): adds databricks support - completion, async, streaming
Closes https://github.com/BerriAI/litellm/issues/2160
2024-05-23 16:29:46 -07:00
Ishaan Jaff
e0af091b5e feat - add prixing for vertex_ai image gen 2024-05-23 16:27:08 -07:00
Krish Dholakia
beb61702c3
Merge branch 'main' into litellm_filter_invalid_params 2024-05-21 20:42:21 -07:00
Krrish Dholakia
988970f4c2 feat(router.py): Fixes https://github.com/BerriAI/litellm/issues/3769 2024-05-21 17:24:51 -07:00
Krrish Dholakia
62e5559e0a fix(utils.py): filter out hf eos token
Closes https://github.com/BerriAI/litellm/issues/3757
2024-05-21 14:31:54 -07:00
alisalim17
01bb26bbba Revert "Revert "Logfire Integration""
This reverts commit b04a8d878a.
2024-05-21 11:07:40 +04:00
Krish Dholakia
c6bb6e325b
Merge pull request #3740 from BerriAI/litellm_return_rejected_response
feat(proxy_server.py): allow admin to return rejected response as string to user
2024-05-20 17:48:21 -07:00
Ishaan Jaff
8413fdf4c7
Merge branch 'main' into litellm_standardize_slack_exception_msg_format 2024-05-20 16:39:41 -07:00
Ishaan Jaff
25fae37efc
Merge pull request #3716 from BerriAI/litellm_set_cooldown_time_based_on_exception_header
[Feat] Router/ Proxy - set cooldown_time based on Azure exception headers
2024-05-20 16:34:12 -07:00
Ishaan Jaff
8263d15ca3 fix - standardize slack alerting format 2024-05-20 16:26:11 -07:00
Ishaan Jaff
91a89eb4ed
Merge pull request #3739 from BerriAI/litellm_add_imagen_support
[FEAT] Async VertexAI Image Generation
2024-05-20 14:14:43 -07:00
Ishaan Jaff
518db13982 add parameter mapping with vertex ai 2024-05-20 13:28:20 -07:00