Ishaan Jaff
41ca6fd52a
feat - check max response size
2024-07-27 16:53:00 -07:00
Ishaan Jaff
4ab8d2229d
security - check max request size
2024-07-27 16:08:41 -07:00
Ishaan Jaff
2e9fb5ca1f
Merge pull request #4924 from BerriAI/litellm_log_writing_spend_to_db_otel
...
[Feat] - log writing BatchSpendUpdate events on OTEL
2024-07-27 16:07:56 -07:00
Ishaan Jaff
19fb5cc11c
use common helpers for writing to otel
2024-07-27 11:40:39 -07:00
Ishaan Jaff
d5d9ed73af
use _get_parent_otel_span_from_kwargs
2024-07-27 11:14:06 -07:00
Ishaan Jaff
61c10e60a4
feat - use log_to_opentelemetry for _PROXY_track_cost_callback
2024-07-27 11:08:22 -07:00
Krrish Dholakia
2719860c46
build(model_prices_and_context_window.json): add mistral-large on vertex ai pricing
2024-07-27 10:37:18 -07:00
Ishaan Jaff
1adf71b9b7
feat - clearly show version litellm enterprise
2024-07-27 09:50:03 -07:00
Ishaan Jaff
6f428a16fa
fix update public key
2024-07-27 09:45:58 -07:00
Krish Dholakia
9bdcef238b
Merge pull request #4907 from BerriAI/litellm_proxy_get_secret
...
fix(proxy_server.py): fix get secret for environment_variables
2024-07-26 22:17:11 -07:00
Krish Dholakia
f9c2fec1a6
Merge pull request #4918 from BerriAI/litellm_ollama_tool_calling
...
feat(ollama_chat.py): support ollama tool calling
2024-07-26 22:16:58 -07:00
Krrish Dholakia
77fe8f57cf
docs(ollama.md): add ollama tool calling to docs
2024-07-26 22:12:52 -07:00
Krrish Dholakia
b25d4a8cb3
feat(ollama_chat.py): support ollama tool calling
...
Closes https://github.com/BerriAI/litellm/issues/4812
2024-07-26 21:51:54 -07:00
Ishaan Jaff
2501b4eccd
feat link to model cost map on swagger
2024-07-26 21:34:42 -07:00
Ishaan Jaff
548adea8cf
add litellm_header_name endpoint
2024-07-26 21:04:31 -07:00
Ishaan Jaff
a7f964b869
Merge pull request #4913 from BerriAI/litellm_fix_error_limit
...
[Proxy-Fix] - raise more descriptive errors when crossing tpm / rpm limits on keys, user, global limits
2024-07-26 20:25:28 -07:00
Ishaan Jaff
3c463ccbe6
Merge pull request #4914 from BerriAI/litellm_fix_batches
...
[Proxy-Fix + Test] - /batches endpoint
2024-07-26 20:12:03 -07:00
Krrish Dholakia
fe0b55f2ca
fix(utils.py): fix cache hits for streaming
...
Fixes https://github.com/BerriAI/litellm/issues/4109
2024-07-26 19:04:08 -07:00
Ishaan Jaff
f627fa9b40
fix for GET /v1/batches{batch_id:path}
2024-07-26 18:23:15 -07:00
Ishaan Jaff
56ce7e892d
fix batches inserting metadata
2024-07-26 18:08:54 -07:00
Ishaan Jaff
159a880dcc
fix /v1/batches POST
2024-07-26 18:06:00 -07:00
Ishaan Jaff
c4e4b4675c
fix raise better error when crossing tpm / rpm limits
2024-07-26 17:35:08 -07:00
Krrish Dholakia
9943c6d607
fix(proxy_server.py): fix get secret for environment_variables
2024-07-26 13:33:02 -07:00
Krrish Dholakia
84482703b8
docs(config.md): update wildcard docs
2024-07-26 08:59:53 -07:00
Krrish Dholakia
1d6c39a607
feat(proxy_server.py): handle pydantic mockselvar error
...
Fixes https://github.com/BerriAI/litellm/issues/4898#issuecomment-2252105485
2024-07-26 08:38:51 -07:00
Krrish Dholakia
2f773d9cb6
fix(litellm_cost_calc/google.py): support meta llama vertex ai cost tracking
2024-07-25 22:12:07 -07:00
Ishaan Jaff
079a41fbe1
Merge branch 'main' into litellm_proxy_support_all_providers
2024-07-25 20:15:37 -07:00
Ishaan Jaff
68e94f0976
example mistral sdk
2024-07-25 19:48:54 -07:00
Krrish Dholakia
f5fd9bd3ce
feat(proxy_server.py): support custom llm handler on proxy
2024-07-25 19:35:52 -07:00
Ishaan Jaff
1a89f9e86d
feat - support health check audio_speech
2024-07-25 19:35:48 -07:00
Krrish Dholakia
a5ecf6b1c1
fix(proxy_server.py): check if input list > 0 before indexing into it
...
resolves 'list index out of range' error
2024-07-25 19:35:48 -07:00
Krrish Dholakia
86973d6aca
fix(main.py): fix calling openai gpt-3.5-turbo-instruct via /completions
...
Fixes https://github.com/BerriAI/litellm/issues/749
2024-07-25 19:35:40 -07:00
Krrish Dholakia
9955b1b486
fix(internal_user_endpoints.py): support updating budgets for /user/update
2024-07-25 19:35:29 -07:00
Krrish Dholakia
2c71f6dd04
feat(auth_check.py): support using redis cache for team objects
...
Allows team update / check logic to work across instances instantly
2024-07-25 19:35:29 -07:00
Ishaan Jaff
693bcfac39
fix using pass_through_all_models
2024-07-25 19:32:49 -07:00
Krish Dholakia
c2086300b7
Merge branch 'main' into litellm_redis_team_object
2024-07-25 19:31:52 -07:00
Krish Dholakia
a306b83b2d
Merge pull request #4887 from BerriAI/litellm_custom_llm
...
feat(custom_llm.py): Support Custom LLM Handlers
2024-07-25 19:05:29 -07:00
Ishaan Jaff
9863520376
support using */*
2024-07-25 18:48:56 -07:00
Ishaan Jaff
8f4c5437b8
router support setting pass_through_all_models
2024-07-25 18:34:12 -07:00
Krrish Dholakia
bd7af04a72
feat(proxy_server.py): support custom llm handler on proxy
2024-07-25 17:56:34 -07:00
Ishaan Jaff
2432c90515
feat - support health check audio_speech
2024-07-25 17:26:14 -07:00
Krrish Dholakia
bfdda089c8
fix(proxy_server.py): check if input list > 0 before indexing into it
...
resolves 'list index out of range' error
2024-07-25 14:23:07 -07:00
Krrish Dholakia
4e51f712f3
fix(main.py): fix calling openai gpt-3.5-turbo-instruct via /completions
...
Fixes https://github.com/BerriAI/litellm/issues/749
2024-07-25 09:57:19 -07:00
Krrish Dholakia
b376ee71b0
fix(internal_user_endpoints.py): support updating budgets for /user/update
2024-07-24 21:51:46 -07:00
Krish Dholakia
e6963217ba
Merge branch 'main' into litellm_parallel_requests
2024-07-24 19:25:56 -07:00
Krrish Dholakia
6ab2527fdc
feat(auth_check.py): support using redis cache for team objects
...
Allows team update / check logic to work across instances instantly
2024-07-24 18:14:49 -07:00
Marc Abramowitz
6faaa8aa50
Allow not displaying feedback box
...
by setting an env var called `LITELLM_DONT_SHOW_FEEDBACK_BOX` to `"true"`.
I liked the feedback box when I first started using LiteLLM, because it showed
me that the authors care about customers. But now that I've seen it a bunch of
times, I don't need to see it every time I start the server and I'd rather have
less output on startup.
2024-07-24 16:50:10 -07:00
Krrish Dholakia
b5c5ed2209
fix(key_management_endpoints.py): if budget duration set, set budget_reset_at
2024-07-24 15:02:22 -07:00
Krrish Dholakia
99d8b0ad68
fix(bedrock_httpx.py): fix async client check
2024-07-24 12:43:52 -07:00
Marc Abramowitz
642f1a7bcc
Check existence of multiple views in 1 query
...
instead of multiple queries. This is more efficient because it lets us check for
all views in one query instead of multiple queries.
2024-07-23 21:55:02 -07:00