Krish Dholakia
|
1c6f6592ea
|
Merge pull request #3153 from BerriAI/litellm_usage_based_routing_v2_improvements
usage based routing v2 improvements - unit testing + *NEW* async + sync 'pre_call_checks'
|
2024-04-18 22:16:16 -07:00 |
|
Krrish Dholakia
|
5bb73dc9c0
|
fix(router.py): instrument pre-call-checks for all openai endpoints
|
2024-04-18 21:54:25 -07:00 |
|
Krrish Dholakia
|
376ee4e9d7
|
fix(test_lowest_tpm_rpm_routing_v2.py): unit testing for usage-based-routing-v2
|
2024-04-18 21:38:00 -07:00 |
|
Ishaan Jaff
|
94ea8c44a7
|
(ci/cd) vertexai - skip 429 errors
|
2024-04-18 21:27:22 -07:00 |
|
Ishaan Jaff
|
1a780b8bc0
|
(ci/cd) run again
|
2024-04-18 21:04:07 -07:00 |
|
Ishaan Jaff
|
29b1745a43
|
(ci/cd) run again
|
2024-04-18 20:35:21 -07:00 |
|
Ishaan Jaff
|
44207fd404
|
fix - view spend per user
|
2024-04-18 20:23:04 -07:00 |
|
Krrish Dholakia
|
72691e05f4
|
fix(tpm_rpm_routing_v2.py): fix tpm rpm routing
|
2024-04-18 20:01:22 -07:00 |
|
Krish Dholakia
|
741a18a040
|
Merge pull request #3144 from BerriAI/litellm_prometheus_latency_tracking
feat(prometheus_services.py): emit proxy latency for successful llm api requests
|
2024-04-18 19:10:58 -07:00 |
|
Krrish Dholakia
|
852455e7fc
|
fix(caching.py): fix error logging
|
2024-04-18 17:21:10 -07:00 |
|
Ishaan Jaff
|
3cbbbd5861
|
Merge pull request #3142 from BerriAI/litellm_slack_alerting_show_model_passed
[Fix] Show `model` passed on `"400: {'error': 'Invalid model name passed in mode` errors 👻
|
2024-04-18 16:18:17 -07:00 |
|
Ishaan Jaff
|
8a3937a7ae
|
test - invalid model name passed test
|
2024-04-18 15:46:49 -07:00 |
|
Krish Dholakia
|
fe5c63e80b
|
Merge pull request #3105 from BerriAI/litellm_fix_hashing
fix(_types.py): hash api key in UserAPIKeyAuth
|
2024-04-18 08:16:24 -07:00 |
|
Krrish Dholakia
|
280d9b4405
|
fix(utils.py): function_setup empty message fix
fixes https://github.com/BerriAI/litellm/issues/2858
|
2024-04-18 07:32:29 -07:00 |
|
Krrish Dholakia
|
64fe5b146c
|
fix(utils.py): fix azure streaming logic
|
2024-04-18 07:08:36 -07:00 |
|
Krrish Dholakia
|
a21630dac7
|
test(test_completion.py): skip local test
|
2024-04-17 19:14:41 -07:00 |
|
Krrish Dholakia
|
fdd73a4e26
|
fix(utils.py): support azure mistral function calling
|
2024-04-17 19:10:26 -07:00 |
|
Krrish Dholakia
|
10d65f0676
|
test(test_function_calling.py): handle for when model returns a text response
|
2024-04-17 18:32:34 -07:00 |
|
Krrish Dholakia
|
ed70680e4a
|
test(test_function_calling.py): fix test to check if tool call in response
|
2024-04-17 18:28:59 -07:00 |
|
Krrish Dholakia
|
da05d86dce
|
test(test_max_tpm_rpm_limiter.py): fix test to use hashed api key in cache check
|
2024-04-17 18:10:18 -07:00 |
|
Krrish Dholakia
|
caa46ca905
|
fix(utils.py): fix streaming special character flushing logic
|
2024-04-17 18:03:40 -07:00 |
|
Krrish Dholakia
|
1b4462ee70
|
fix(utils.py): ensure streaming output parsing only applied for hf / sagemaker models
selectively applies the <s>
</s> checking
|
2024-04-17 17:43:41 -07:00 |
|
Krrish Dholakia
|
29368b5d71
|
test: refactor testing to handle the hash token fix
|
2024-04-17 17:31:39 -07:00 |
|
Krrish Dholakia
|
013611a676
|
fix(_types.py): hash api key, if given
fixes https://github.com/BerriAI/litellm/issues/2114
|
2024-04-17 16:43:35 -07:00 |
|
Krrish Dholakia
|
2a2b97f093
|
fix(utils.py): accept {custom_llm_provider}/{model_name} in get_model_info
fixes https://github.com/BerriAI/litellm/issues/3100
|
2024-04-17 16:38:53 -07:00 |
|
Ishaan Jaff
|
72e7571e51
|
ci/cd run again
|
2024-04-17 08:01:39 -07:00 |
|
Ishaan Jaff
|
12afabca08
|
fix - azure test - don't pop api key
|
2024-04-17 07:54:16 -07:00 |
|
Krish Dholakia
|
d55aada92a
|
Merge pull request #3062 from cwang/cwang/trim-messages-fix
Use `max_input_token` for `trim_messages`
|
2024-04-16 22:29:45 -07:00 |
|
Krrish Dholakia
|
db4ea71f0f
|
test(test_amazing_vertex_completion.py): handle api errors
|
2024-04-16 21:47:02 -07:00 |
|
Ishaan Jaff
|
9d601d2549
|
(ci/cd) run again
|
2024-04-16 21:44:11 -07:00 |
|
Ishaan Jaff
|
539b5dd562
|
fix test_key_name_null
|
2024-04-16 21:37:31 -07:00 |
|
Ishaan Jaff
|
b00ad976af
|
fix function calling prompt - ask llm to respond in fahrenheit
|
2024-04-16 21:09:53 -07:00 |
|
Ishaan Jaff
|
f2c1c20cc1
|
fix - gemini function calling
|
2024-04-16 20:54:22 -07:00 |
|
Ishaan Jaff
|
50770ef3b0
|
test vertex ai exception mapping
|
2024-04-16 20:00:32 -07:00 |
|
Ishaan Jaff
|
99065cb6b4
|
Merge pull request #3079 from BerriAI/litellm_router_save_settings_ui
UI - Save / Edit router settings UI
|
2024-04-16 16:57:42 -07:00 |
|
Krrish Dholakia
|
c6ad02b167
|
fix(proxy_server.py): ensure id used in delete deployment matches id used in litellm Router
|
2024-04-16 15:17:18 -07:00 |
|
Ishaan Jaff
|
fb3edc6d92
|
router - get settings
|
2024-04-16 14:22:54 -07:00 |
|
Ishaan Jaff
|
3ad0651b59
|
return correct response from user/new
|
2024-04-16 08:17:45 -07:00 |
|
Krish Dholakia
|
b46e9058df
|
Merge pull request #3019 from ligaz/fix-anthropic-system-msg-handling
Fix Anthropic system message handling
|
2024-04-16 07:18:47 -07:00 |
|
Chen Wang
|
2567f9a3a6
|
Use max_input_token for trim_messages
|
2024-04-16 13:36:25 +01:00 |
|
Krrish Dholakia
|
89017448d4
|
test(test_amazing_vertex_completion.py): fix vertex anthropic streaming text
|
2024-04-15 14:37:22 -07:00 |
|
Krrish Dholakia
|
8c3c45fbb5
|
fix(vertex_ai_anthropic.py): set vertex_credentials for vertex ai anthropic calls
allows setting vertex credentials as a json string for vertex ai anthropic calls
|
2024-04-15 14:16:28 -07:00 |
|
Krrish Dholakia
|
3d645f95a5
|
fix(main.py): accept vertex service account credentials as json string
allows us to dynamically set vertex ai credentials
|
2024-04-15 13:28:59 -07:00 |
|
Krish Dholakia
|
cfd2bc030f
|
Merge pull request #3028 from BerriAI/litellm_anthropic_text_completion_fix
fix(anthropic_text.py): add support for async text completion calls
|
2024-04-15 09:26:28 -07:00 |
|
Ishaan Jaff
|
c356173d73
|
test groq tool calling
|
2024-04-15 08:41:53 -07:00 |
|
Ishaan Jaff
|
2a54d9a6ec
|
fix test groq function call
|
2024-04-15 08:40:39 -07:00 |
|
Krrish Dholakia
|
1cd0551a1e
|
fix(anthropic_text.py): add support for async text completion calls
|
2024-04-15 08:15:00 -07:00 |
|
Ishaan Jaff
|
82b050dcc8
|
test - groq tool calling
|
2024-04-15 08:13:05 -07:00 |
|
Stefan Dobrev
|
b3b52dbe32
|
Fix Anthropic system message handling
|
2024-04-14 16:23:59 +03:00 |
|
Krrish Dholakia
|
b0fc2b342d
|
fix(caching.py): don't decode a string
|
2024-04-13 18:48:03 -07:00 |
|