Commit graph

3109 commits

Author SHA1 Message Date
Ishaan Jaff
265ec00d0f fix test routes on litellm proxy 2024-07-10 16:51:47 -07:00
Ishaan Jaff
a313174ecb
Merge pull request #4648 from BerriAI/litellm_add_remaining_file_endpoints
[Feat] Add LIST, DELETE, GET `/files`
2024-07-10 16:42:05 -07:00
Ishaan Jaff
a5f12aba81 add file content 2024-07-10 16:15:43 -07:00
Ishaan Jaff
cb300d30a9 add file delete path 2024-07-10 16:08:58 -07:00
Marc Abramowitz
3a2cb151aa Proxy: Add x-litellm-call-id response header
This gives the value of `logging_obj.litellm_call_id` and one particular use of
this is to correlate the HTTP response from a request with a trace in an LLM
logging tool like Langfuse, Langsmith, etc.

For example, if a user in my environment (w/ Langfuse) gets back this in the
response headers:

```
x-litellm-call-id: ffcb49e7-bd6e-4e56-9c08-a7243802b26e
```

then they know that they can see the trace for this request in Langfuse by
visiting https://langfuse.domain.com/trace/ffcb49e7-bd6e-4e56-9c08-a7243802b26e

They can also use this ID to submit scores for this request to the Langfuse
scoring API.
2024-07-10 16:05:37 -07:00
Ishaan Jaff
0b39ec6a8e add /v1/files LIST 2024-07-10 16:04:07 -07:00
Marc Abramowitz
2db9c23bce Remove unnecessary imports
from `litellm/proxy/proxy_server.py`
2024-07-10 15:06:47 -07:00
Ishaan Jaff
efca9daf5d add "/v1/files/{file_id}" as openai route 2024-07-10 14:56:53 -07:00
Ishaan Jaff
393ce7df14 add /files endpoints 2024-07-10 14:55:10 -07:00
Krrish Dholakia
aace0b22a3 fix(proxy_server.py): fix proxy_server.py premium user check for encrypted license key 2024-07-10 12:25:31 -07:00
Ishaan Jaff
09fe40791e add "/v1/assistants/{assistant_id}", as openai route 2024-07-10 11:42:02 -07:00
Ishaan Jaff
62f475919b feat - add DELETE assistants endpoint 2024-07-10 11:37:37 -07:00
Krrish Dholakia
5d6e172d5c feat(anthropic_adapter.py): support for translating anthropic params to openai format 2024-07-10 00:32:28 -07:00
Krrish Dholakia
d077148135 style(litellm_license.py): add debug statement for litellm license 2024-07-09 22:43:33 -07:00
Ishaan Jaff
3a06e2e425 fix show exact prisma exception when starting proxy 2024-07-09 18:20:09 -07:00
Ishaan Jaff
a250e99934 ui new build 2024-07-09 16:33:00 -07:00
Ishaan Jaff
22df67edb7 feat - add mgtm endpoint routes 2024-07-09 15:29:41 -07:00
Ishaan Jaff
362c01c21f ui - add Create, get, delete endpoints for IP Addresses 2024-07-09 15:12:08 -07:00
Krrish Dholakia
a1986fab60 fix(vertex_httpx.py): add sync vertex image gen support
Fixes https://github.com/BerriAI/litellm/issues/4623
2024-07-09 13:33:54 -07:00
Ishaan Jaff
6bce7e73a3
Merge pull request #4627 from BerriAI/litellm_fix_thread_auth
[Fix] Authentication on /thread endpoints on Proxy
2024-07-09 12:19:19 -07:00
Ishaan Jaff
e9b14a56d2 test /threads endpoint 2024-07-09 12:17:42 -07:00
Ishaan Jaff
b30fa64e2a fix - use helper to check if a route is openai route 2024-07-09 12:00:07 -07:00
Ishaan Jaff
e380954de1 add helper to check is_openai_route 2024-07-09 11:50:12 -07:00
Ishaan Jaff
9423ab95ba fix add assistant settings on config 2024-07-09 10:05:32 -07:00
Ishaan Jaff
12f207b499 feat - support /create assistants endpoint 2024-07-09 10:03:47 -07:00
Ishaan Jaff
6479442d5b feat - support acreate_assistants endpoint 2024-07-09 09:49:38 -07:00
Krrish Dholakia
e4105a6959 docs(configs.md): add ip address filtering to docs 2024-07-08 21:59:26 -07:00
Krish Dholakia
a986413df3
Merge pull request #4615 from BerriAI/litellm_user_api_key_auth
Enable `allowed_ip's` for proxy
2024-07-08 17:35:11 -07:00
Krrish Dholakia
fe62e4e1c4 fix(proxy_cli.py): bump default azure api version 2024-07-08 16:28:22 -07:00
Ishaan Jaff
4af1d032a8 ui new build 2024-07-08 16:19:25 -07:00
Krrish Dholakia
3045a2d9b3 fix(proxy_server.py): add license protection for 'allowed_ip' address feature 2024-07-08 16:04:44 -07:00
Krrish Dholakia
f3cc57bc6f feat(user_api_key_auth.py): allow restricting calls by IP address
Allows admin to restrict which IP addresses can make calls to the proxy
2024-07-08 15:58:15 -07:00
Ishaan Jaff
6930881891
Merge pull request #4611 from BerriAI/litellm_fix_assistants_routes
[Proxy-Fix]: Add /assistants, /threads as OpenAI routes
2024-07-08 15:16:11 -07:00
Ishaan Jaff
9d5853943a fix routes on assistants endpoints 2024-07-08 15:02:12 -07:00
Andres Guzman
a650160390 fix(utils.py): change update to upsert 2024-07-08 15:49:29 -06:00
Ishaan Jaff
5c0d4a4162 use ProxyErrorTypes, 2024-07-08 12:47:53 -07:00
Ishaan Jaff
25ddbbb41d raise budget_exceeded in user_api_key_auth 2024-07-08 12:45:39 -07:00
Ishaan Jaff
c6256284da use types for ProxyErrorTypes 2024-07-08 12:42:27 -07:00
Ishaan Jaff
fad595fd47
Merge pull request #4603 from BerriAI/litellm_track_user_ip
[Enterprise-Feature: Proxy] Track user-ip address in requests & in LiteLLM_SpendLogs
2024-07-08 12:18:52 -07:00
Ishaan Jaff
b99ffcf8ae SpendLogsPayload- track user ip 2024-07-08 10:16:58 -07:00
Krrish Dholakia
298505c47c fix(whisper---handle-openai/azure-vtt-response-format): Fixes https://github.com/BerriAI/litellm/issues/4595 2024-07-08 09:10:40 -07:00
Ishaan Jaff
626c630eaf track user_ip address per request 2024-07-08 09:00:08 -07:00
Ishaan Jaff
bbf3de74dd fix - setting rpm/tpm 2024-07-08 07:43:00 -07:00
Krrish Dholakia
1193ee8803 fix(presidio_pii_masking.py): fix presidio unset url check + add same check for langfuse 2024-07-06 17:50:55 -07:00
Krrish Dholakia
d57d3df1d6 fix(presidio_pii_masking.py): add support for setting 'http://' if unset by render env for presidio base url 2024-07-06 17:42:10 -07:00
Krrish Dholakia
20e39d6acc fix(utils.py): cleanup 'additionalProperties=False' for tool calling with zod
Fixes issue with zod passing in additionalProperties=False, causing vertex ai / gemini calls to fail
2024-07-06 17:27:37 -07:00
Ishaan Jaff
196884a779 fix trace hierarchy on otel 2024-07-06 15:37:23 -07:00
Ishaan Jaff
547f6ffdd5 ui new build 2024-07-06 15:26:26 -07:00
Ishaan Jaff
e815563638
Merge pull request #4578 from BerriAI/litellm_allow_querying_spend_report_by_key
[Feat-Enterprise] /spend/report view spend for a specific key
2024-07-06 15:11:46 -07:00
Ishaan Jaff
d61cc598b0
Merge pull request #4576 from BerriAI/litellm_encrypt_decrypt_using_salt
[Refactor] Use helper function to encrypt/decrypt model credentials
2024-07-06 15:11:09 -07:00