Commit graph

1831 commits

Author SHA1 Message Date
Ishaan Jaff
20a5bbe6a6 fix allow general guardrails on free tier 2024-09-04 19:59:32 -07:00
Krish Dholakia
be3c7b401e
LiteLLM Minor fixes + improvements (08/03/2024) (#5488)
* fix(internal_user_endpoints.py): set budget_reset_at for /user/update

* fix(vertex_and_google_ai_studio_gemini.py): handle accumulated json

Fixes https://github.com/BerriAI/litellm/issues/5479

* fix(vertex_ai_and_gemini.py): fix assistant message function call when content is not None

Fixes https://github.com/BerriAI/litellm/issues/5490

* fix(proxy_server.py): generic state uuid for okta sso

* fix(lago.py): improve debug logs

Debugging for https://github.com/BerriAI/litellm/issues/5477

* docs(bedrock.md): add bedrock cross-region inferencing to docs

* fix(azure.py): return azure response headers on aembedding call

* feat(azure.py): return azure response headers for `/audio/transcription`

* fix(types/utils.py): standardize deepseek / anthropic prompt caching usage information

Closes https://github.com/BerriAI/litellm/issues/5285

* docs(usage.md): add docs on litellm usage object

* test(test_completion.py): mark flaky test
2024-09-03 21:21:34 -07:00
Ishaan Jaff
3c898e23ea refactor secret managers 2024-09-03 10:58:02 -07:00
Ishaan Jaff
b0178a85cf refactor get_secret 2024-09-03 10:42:12 -07:00
Krish Dholakia
9f3fa29624
feat(router.py): Support Loadbalancing batch azure api endpoints (#5469)
* feat(router.py): initial commit for loadbalancing azure batch api endpoints

Closes https://github.com/BerriAI/litellm/issues/5396

* fix(router.py): working `router.acreate_file()`

* feat(router.py): working router.acreate_batch endpoint

* feat(router.py): expose router.aretrieve_batch function

Make it easy for user to retrieve the batch information

* feat(router.py): support 'router.alist_batches' endpoint

Adds support for getting all batches across all endpoints

* feat(router.py): working loadbalancing on `/v1/files`

* feat(proxy_server.py): working loadbalancing on `/v1/batches`

* feat(proxy_server.py): working loadbalancing on Retrieve + List batch
2024-09-02 21:32:55 -07:00
Ishaan Jaff
aa13977136 refactor vtx image gen 2024-09-02 17:35:51 -07:00
Ishaan Jaff
56f10224df
Merge pull request #5457 from BerriAI/litellm_track_spend_logs_for_vertex_pass_through_endpoints
[Feat-Proxy] track spend logs for vertex pass through endpoints
2024-08-31 16:30:15 -07:00
Ishaan Jaff
b35bfb0302 fix cost tracking for vertex ai native 2024-08-31 08:22:27 -07:00
Ishaan Jaff
7d746064ab add gcs bucket base 2024-08-30 10:41:39 -07:00
Ishaan Jaff
ad88c7d0a8 show all error types on swagger 2024-08-29 18:50:41 -07:00
Ishaan Jaff
fb5be57bb8 v0 add rerank on litellm proxy 2024-08-27 17:28:39 -07:00
Ishaan Jaff
74f0e60962 fix set Caching Default Off 2024-08-24 09:43:39 -07:00
Krrish Dholakia
ac9a1e65ab fix(proxy_server.py): fix post /v1/batches endpoint
Fixes https://github.com/BerriAI/litellm/issues/5279#issuecomment-2307919820
2024-08-23 20:38:00 -07:00
Krrish Dholakia
ab28e55b76 fix(proxy_server.py): support env vars for controlling global max parallel request retry/timeouts
fixes issue where litellm module level settings weren't working for global retries, due to time of init
2024-08-23 16:06:08 -07:00
Ishaan Jaff
1b1e0f2d77 init custom guardrail class 2024-08-23 10:54:42 -07:00
Krish Dholakia
76b3db334b
Merge branch 'main' into litellm_azure_batch_apis 2024-08-22 19:07:54 -07:00
Krrish Dholakia
735fc804ed fix(proxy_server.py): expose flag to disable retries when max parallel request limit is hit 2024-08-22 16:49:52 -07:00
Krrish Dholakia
63cd94c32a fix: fix linting errors 2024-08-22 15:51:59 -07:00
Krrish Dholakia
8625663458 feat(proxy_server.py): support azure batch api endpoints 2024-08-22 15:21:43 -07:00
Krish Dholakia
68cb5cae58
Merge branch 'main' into litellm_redis_cluster 2024-08-22 11:06:14 -07:00
Ishaan Jaff
a120135dd1 fix allow setting LiteLLM license as .env 2024-08-22 10:05:00 -07:00
Ishaan Jaff
cc8e6f1d44 fix allow setting license in config.yaml 2024-08-22 09:45:15 -07:00
Ishaan Jaff
2be984ebee add docstring for /embeddings and /completions 2024-08-22 09:30:47 -07:00
Ishaan Jaff
f6e80b0031 add doc string for /chat/completions swagger 2024-08-22 09:27:40 -07:00
Ishaan Jaff
a174cbdd72
Merge branch 'main' into litellm_pass_through_vtx_multi_modal 2024-08-21 17:23:22 -07:00
Ishaan Jaff
e9537c6560 proxy - print embedding request when recieved 2024-08-21 17:00:18 -07:00
Krish Dholakia
72169fd5c4
Merge branch 'main' into litellm_disable_storing_master_key_hash_in_db 2024-08-21 15:37:25 -07:00
Krrish Dholakia
e2d7539690 feat(caching.py): redis cluster support
Closes https://github.com/BerriAI/litellm/issues/4358
2024-08-21 15:01:52 -07:00
Ishaan Jaff
d6493b0e7f docs semantic caching qdrant 2024-08-21 13:03:41 -07:00
Krrish Dholakia
89014dfc07 feat(proxy_server.py): support disabling storing master key hash in db, for spend tracking 2024-08-21 12:35:37 -07:00
Krrish Dholakia
6f8840daa1 fix(proxy_server.py): fix invalid login message to not show passed in pwd
Closes https://github.com/BerriAI/litellm/issues/5290
2024-08-20 08:56:57 -07:00
Ishaan Jaff
9ef6ae2f7c
Merge pull request #4868 from msabramo/allow-not-displaying-feedback-box
Allow not displaying feedback box
2024-08-20 08:53:45 -07:00
Ishaan Jaff
c7b3978655
Merge pull request #5288 from BerriAI/litellm_aporia_refactor
[Feat] V2 aporia guardrails litellm
2024-08-19 20:41:45 -07:00
Ishaan Jaff
8cd1963c11 feat - guardrails v2 2024-08-19 18:24:20 -07:00
Krrish Dholakia
1701c48ad5 feat(langfuse_endpoints.py): support langfuse pass through endpoints by default 2024-08-19 17:28:34 -07:00
Ishaan Jaff
613bd1babd feat - return applied guardrails in response headers 2024-08-19 11:56:20 -07:00
Ishaan Jaff
4685b9909a feat - allow accessing data post success call 2024-08-19 11:35:33 -07:00
Krish Dholakia
ff6ff133ee
Merge pull request #5260 from BerriAI/google_ai_studio_pass_through
Pass-through endpoints for Gemini - Google AI Studio
2024-08-17 13:51:51 -07:00
Ishaan Jaff
feb8c3c5b4
Merge pull request #5259 from BerriAI/litellm_return_remaining_tokens_in_header
[Feat] return `x-litellm-key-remaining-requests-{model}`: 1, `x-litellm-key-remaining-tokens-{model}: None` in response headers
2024-08-17 12:41:16 -07:00
Ishaan Jaff
ee0f772b5c feat return rmng tokens for model for api key 2024-08-17 12:35:10 -07:00
Krrish Dholakia
bc0023a409 feat(google_ai_studio_endpoints.py): support pass-through endpoint for all google ai studio requests
New Feature
2024-08-17 10:46:59 -07:00
Ishaan Jaff
5985c7e933 feat - use commong helper for getting model group 2024-08-17 10:46:04 -07:00
Krrish Dholakia
61f4b71ef7 refactor: replace .error() with .exception() logging for better debugging on sentry 2024-08-16 09:22:47 -07:00
Krish Dholakia
b3d15ace89
Merge branch 'main' into litellm_pass_through_endpoints_api 2024-08-15 22:39:19 -07:00
Krrish Dholakia
6fc6df134f feat(ui): for adding pass-through endpoints 2024-08-15 21:58:11 -07:00
Krrish Dholakia
589da45c24 feat(pass_through_endpoints.py): initial working CRUD endpoints for /pass_through_endoints 2024-08-15 21:23:26 -07:00
Ishaan Jaff
58828403ea refactor use 1 util for llm routing 2024-08-15 08:42:20 -07:00
Ishaan Jaff
fdd6664420 use route_request for making llm call 2024-08-15 08:16:44 -07:00
Krrish Dholakia
28faafadb1 feat(pass_through_endpoints.py): initial commit of crud endpoints for pass through endpoints 2024-08-14 21:36:07 -07:00
Krrish Dholakia
a020563149 feat(proxy_server.py): support returning available fields for pass_through_endpoints via `/config/field/list 2024-08-14 19:07:10 -07:00