Commit graph

2362 commits

Author SHA1 Message Date
Krish Dholakia
1c6f6592ea Merge pull request #3153 from BerriAI/litellm_usage_based_routing_v2_improvements
usage based routing v2 improvements - unit testing + *NEW* async + sync 'pre_call_checks'
2024-04-18 22:16:16 -07:00
Krrish Dholakia
5bb73dc9c0 fix(router.py): instrument pre-call-checks for all openai endpoints 2024-04-18 21:54:25 -07:00
Krrish Dholakia
376ee4e9d7 fix(test_lowest_tpm_rpm_routing_v2.py): unit testing for usage-based-routing-v2 2024-04-18 21:38:00 -07:00
Ishaan Jaff
94ea8c44a7 (ci/cd) vertexai - skip 429 errors 2024-04-18 21:27:22 -07:00
Ishaan Jaff
1a780b8bc0 (ci/cd) run again 2024-04-18 21:04:07 -07:00
Ishaan Jaff
29b1745a43 (ci/cd) run again 2024-04-18 20:35:21 -07:00
Ishaan Jaff
44207fd404 fix - view spend per user 2024-04-18 20:23:04 -07:00
Krrish Dholakia
72691e05f4 fix(tpm_rpm_routing_v2.py): fix tpm rpm routing 2024-04-18 20:01:22 -07:00
Krish Dholakia
741a18a040 Merge pull request #3144 from BerriAI/litellm_prometheus_latency_tracking
feat(prometheus_services.py): emit proxy latency for successful llm api requests
2024-04-18 19:10:58 -07:00
Krrish Dholakia
852455e7fc fix(caching.py): fix error logging 2024-04-18 17:21:10 -07:00
Ishaan Jaff
3cbbbd5861 Merge pull request #3142 from BerriAI/litellm_slack_alerting_show_model_passed
[Fix] Show `model` passed on `"400: {'error': 'Invalid model name passed in mode` errors 👻
2024-04-18 16:18:17 -07:00
Ishaan Jaff
8a3937a7ae test - invalid model name passed test 2024-04-18 15:46:49 -07:00
Krish Dholakia
fe5c63e80b Merge pull request #3105 from BerriAI/litellm_fix_hashing
fix(_types.py): hash api key in UserAPIKeyAuth
2024-04-18 08:16:24 -07:00
Krrish Dholakia
280d9b4405 fix(utils.py): function_setup empty message fix
fixes https://github.com/BerriAI/litellm/issues/2858
2024-04-18 07:32:29 -07:00
Krrish Dholakia
64fe5b146c fix(utils.py): fix azure streaming logic 2024-04-18 07:08:36 -07:00
Krrish Dholakia
a21630dac7 test(test_completion.py): skip local test 2024-04-17 19:14:41 -07:00
Krrish Dholakia
fdd73a4e26 fix(utils.py): support azure mistral function calling 2024-04-17 19:10:26 -07:00
Krrish Dholakia
10d65f0676 test(test_function_calling.py): handle for when model returns a text response 2024-04-17 18:32:34 -07:00
Krrish Dholakia
ed70680e4a test(test_function_calling.py): fix test to check if tool call in response 2024-04-17 18:28:59 -07:00
Krrish Dholakia
da05d86dce test(test_max_tpm_rpm_limiter.py): fix test to use hashed api key in cache check 2024-04-17 18:10:18 -07:00
Krrish Dholakia
caa46ca905 fix(utils.py): fix streaming special character flushing logic 2024-04-17 18:03:40 -07:00
Krrish Dholakia
1b4462ee70 fix(utils.py): ensure streaming output parsing only applied for hf / sagemaker models
selectively applies the <s>
</s> checking
2024-04-17 17:43:41 -07:00
Krrish Dholakia
29368b5d71 test: refactor testing to handle the hash token fix 2024-04-17 17:31:39 -07:00
Krrish Dholakia
013611a676 fix(_types.py): hash api key, if given
fixes https://github.com/BerriAI/litellm/issues/2114
2024-04-17 16:43:35 -07:00
Krrish Dholakia
2a2b97f093 fix(utils.py): accept {custom_llm_provider}/{model_name} in get_model_info
fixes https://github.com/BerriAI/litellm/issues/3100
2024-04-17 16:38:53 -07:00
Ishaan Jaff
72e7571e51 ci/cd run again 2024-04-17 08:01:39 -07:00
Ishaan Jaff
12afabca08 fix - azure test - don't pop api key 2024-04-17 07:54:16 -07:00
Krish Dholakia
d55aada92a Merge pull request #3062 from cwang/cwang/trim-messages-fix
Use `max_input_token` for `trim_messages`
2024-04-16 22:29:45 -07:00
Krrish Dholakia
db4ea71f0f test(test_amazing_vertex_completion.py): handle api errors 2024-04-16 21:47:02 -07:00
Ishaan Jaff
9d601d2549 (ci/cd) run again 2024-04-16 21:44:11 -07:00
Ishaan Jaff
539b5dd562 fix test_key_name_null 2024-04-16 21:37:31 -07:00
Ishaan Jaff
b00ad976af fix function calling prompt - ask llm to respond in fahrenheit 2024-04-16 21:09:53 -07:00
Ishaan Jaff
f2c1c20cc1 fix - gemini function calling 2024-04-16 20:54:22 -07:00
Ishaan Jaff
50770ef3b0 test vertex ai exception mapping 2024-04-16 20:00:32 -07:00
Ishaan Jaff
99065cb6b4 Merge pull request #3079 from BerriAI/litellm_router_save_settings_ui
UI - Save / Edit router settings UI
2024-04-16 16:57:42 -07:00
Krrish Dholakia
c6ad02b167 fix(proxy_server.py): ensure id used in delete deployment matches id used in litellm Router 2024-04-16 15:17:18 -07:00
Ishaan Jaff
fb3edc6d92 router - get settings 2024-04-16 14:22:54 -07:00
Ishaan Jaff
3ad0651b59 return correct response from user/new 2024-04-16 08:17:45 -07:00
Krish Dholakia
b46e9058df Merge pull request #3019 from ligaz/fix-anthropic-system-msg-handling
Fix Anthropic system message handling
2024-04-16 07:18:47 -07:00
Chen Wang
2567f9a3a6 Use max_input_token for trim_messages 2024-04-16 13:36:25 +01:00
Krrish Dholakia
89017448d4 test(test_amazing_vertex_completion.py): fix vertex anthropic streaming text 2024-04-15 14:37:22 -07:00
Krrish Dholakia
8c3c45fbb5 fix(vertex_ai_anthropic.py): set vertex_credentials for vertex ai anthropic calls
allows setting vertex credentials as a json string for vertex ai anthropic calls
2024-04-15 14:16:28 -07:00
Krrish Dholakia
3d645f95a5 fix(main.py): accept vertex service account credentials as json string
allows us to dynamically set vertex ai credentials
2024-04-15 13:28:59 -07:00
Krish Dholakia
cfd2bc030f Merge pull request #3028 from BerriAI/litellm_anthropic_text_completion_fix
fix(anthropic_text.py): add support for async text completion calls
2024-04-15 09:26:28 -07:00
Ishaan Jaff
c356173d73 test groq tool calling 2024-04-15 08:41:53 -07:00
Ishaan Jaff
2a54d9a6ec fix test groq function call 2024-04-15 08:40:39 -07:00
Krrish Dholakia
1cd0551a1e fix(anthropic_text.py): add support for async text completion calls 2024-04-15 08:15:00 -07:00
Ishaan Jaff
82b050dcc8 test - groq tool calling 2024-04-15 08:13:05 -07:00
Stefan Dobrev
b3b52dbe32 Fix Anthropic system message handling 2024-04-14 16:23:59 +03:00
Krrish Dholakia
b0fc2b342d fix(caching.py): don't decode a string 2024-04-13 18:48:03 -07:00