Ishaan Jaff
a40ecc3fe4
fix import loc
2025-04-23 17:33:29 -07:00
Ishaan Jaff
ce9d084dd4
fix _assemble_key_object
2025-04-23 17:18:21 -07:00
Ishaan Jaff
6c0e1bd38b
fix imports
2025-04-23 17:03:46 -07:00
Ishaan Jaff
5df4dcd33f
fix imports
2025-04-23 16:47:47 -07:00
Ishaan Jaff
cc7a2bfa50
fix imports
2025-04-23 16:25:37 -07:00
Ishaan Jaff
8777e38880
fix imports
2025-04-23 16:22:01 -07:00
Ishaan Jaff
fec670c341
fix import loc2
2025-04-23 16:17:42 -07:00
Ishaan Jaff
1b6b05e8d4
fix import loc2
2025-04-23 16:15:58 -07:00
Ishaan Jaff
daa8c23675
fix import loc2
2025-04-23 16:11:28 -07:00
Ishaan Jaff
5d85af5cee
refactor location of proxy fixes
2025-04-23 15:55:47 -07:00
Ishaan Jaff
ce58c53ff1
refactor location of proxy
2025-04-23 14:38:44 -07:00
Ishaan Jaff
baa5564f95
cleanup remove stale dir
2025-04-23 14:07:43 -07:00
Ishaan Jaff
36ee132514
[Feat] Add gpt-image-1 cost tracking ( #10241 )
...
* add gpt-image-1
* add gpt-image-1 example to docs
2025-04-23 12:20:55 -07:00
Christian Owusu
47420d8d68
Require auth for all dashboard pages ( #10229 )
...
* Require authentication for all Dashboard pages
* Add test
* Add test
2025-04-23 07:08:25 -07:00
Krish Dholakia
217681eb5e
Litellm dev 04 22 2025 p1 ( #10206 )
...
* fix(openai.py): initial commit adding generic event type for openai responses api streaming
Ensures handling for undocumented event types - e.g. "response.reasoning_summary_part.added"
* fix(transformation.py): handle unknown openai response type
* fix(datadog_llm_observability.py): handle dict[str, any] -> dict[str, str] conversion
Fixes https://github.com/BerriAI/litellm/issues/9494
* test: add more unit testing
* test: add unit test
* fix(common_utils.py): fix message with content list
* test: update testing
2025-04-22 23:58:43 -07:00
Krrish Dholakia
31f704a370
fix(internal_user_endpoints.py): add check on sortby value
2025-04-22 21:41:13 -07:00
Ishaan Jaff
96e31d205c
feat: Added Missing Attributes For Arize & Phoenix Integration ( #10043 ) ( #10215 )
...
* feat: Added Missing Attributes For Arize & Phoenix Integration
* chore: Added noqa for PLR0915 to suppress warning
* chore: Moved Contributor Test to Correct Location
* chore: Removed Redundant Fallback
Co-authored-by: Ali Saleh <saleh.a@turing.com>
2025-04-22 21:34:51 -07:00
Krish Dholakia
5f98d4d7de
UI - Users page - Enable global sorting (allows finding users with highest spend) ( #10211 )
...
* fix(view_users.tsx): add time tracking logic to debounce search - prevent new queries from being overwritten by previous ones
* fix(internal_user_endpoints.py): add sort functionality to user list endpoint
* feat(internal_user_endpoints.py): support sort by on `/user/list`
* fix(view_users.tsx): enable global sorting
allows finding user with highest spend
* feat(view_users.tsx): support filtering by sso user id
* test(search_users.spec.ts): add tests to ensure filtering works
* test: add more unit testing
2025-04-22 19:59:53 -07:00
Ishaan Jaff
b96d2ea422
Bug Fix - Address deprecation of open_text ( #10208 )
...
* Update utils.py (#10201 )
* fixes importlib
---------
Co-authored-by: Nathan Brake <33383515+njbrake@users.noreply.github.com>
2025-04-22 18:29:56 -07:00
Ishaan Jaff
868cdd0226
[Feat] Add Support for DELETE /v1/responses/{response_id} on OpenAI, Azure OpenAI ( #10205 )
...
* add transform_delete_response_api_request to base responses config
* add transform_delete_response_api_request
* add delete_response_api_handler
* fixes for deleting responses, response API
* add adelete_responses
* add async test_basic_openai_responses_delete_endpoint
* test_basic_openai_responses_delete_endpoint
* working delete for streaming on responses API
* fixes azure transformation
* TestAnthropicResponsesAPITest
* fix code check
* fix linting
* fixes for get_complete_url
* test_basic_openai_responses_streaming_delete_endpoint
* streaming fixes
2025-04-22 18:27:03 -07:00
Ishaan Jaff
2bb51866b1
fix azure/computer-use-preview native streaming
2025-04-22 18:21:06 -07:00
Krish Dholakia
66680c421d
Add global filtering to Users tab ( #10195 )
...
* style(internal_user_endpoints.py): add response model to `/user/list` endpoint
make sure we maintain consistent response spec
* fix(key_management_endpoints.py): return 'created_at' and 'updated_at' on `/key/generate`
Show 'created_at' on UI when key created
* test(test_keys.py): add e2e test to ensure created at is always returned
* fix(view_users.tsx): support global search by user email
allows easier search
* test(search_users.spec.ts): add e2e test ensure user search works on admin ui
* fix(view_users.tsx): support filtering user by role and user id
More powerful filtering on internal users table
* fix(view_users.tsx): allow filtering users by team
* style(view_users.tsx): cleanup ui to show filters in consistent style
* refactor(view_users.tsx): cleanup to just use 1 variable for the data
* fix(view_users.tsx): cleanup use effect hooks
* fix(internal_user_endpoints.py): fix check to pass testing
* test: update tests
* test: update tests
* Revert "test: update tests"
This reverts commit 6553eeb232
.
* fix(view_userts.tsx): add back in 'previous' and 'next' tabs for pagination
2025-04-22 13:59:43 -07:00
Dwij
b2955a2bdd
Add AgentOps Integration to LiteLLM ( #9685 )
...
* feat(sidebars): add new item for agentops integration in Logging & Observability category
* Update agentops_integration.md to enhance title formatting and remove redundant section
* Enhance AgentOps integration in documentation and codebase by removing LiteLLMCallbackHandler references, adding environment variable configurations, and updating logging initialization for AgentOps support.
* Update AgentOps integration documentation to include instructions for obtaining API keys and clarify environment variable setup.
* Add unit tests for AgentOps integration and improve error handling in token fetching
* Add unit tests for AgentOps configuration and token fetching functionality
* Corrected agentops test directory
* Linting fix
* chore: add OpenTelemetry dependencies to pyproject.toml
* chore: update OpenTelemetry dependencies and add new packages in pyproject.toml and poetry.lock
2025-04-22 10:29:01 -07:00
Krish Dholakia
a7db0df043
Gemini-2.5-flash improvements ( #10198 )
...
* fix(vertex_and_google_ai_studio_gemini.py): allow thinking budget = 0
Fixes https://github.com/BerriAI/litellm/issues/10121
* fix(vertex_and_google_ai_studio_gemini.py): handle nuance in counting exclusive vs. inclusive tokens
Addresses https://github.com/BerriAI/litellm/pull/10141#discussion_r2052272035
2025-04-21 22:48:00 -07:00
Ishaan Jaff
7cb95bcc96
[Bug Fix] caching does not account for thinking or reasoning_effort config ( #10140 )
...
* _get_litellm_supported_chat_completion_kwargs
* test caching with thinking
2025-04-21 22:39:40 -07:00
Ishaan Jaff
104e4cb1bc
[Feat] Add infinity embedding support (contributor pr) ( #10196 )
...
* Feature - infinity support for #8764 (#10009 )
* Added support for infinity embeddings
* Added test cases
* Fixed tests and api base
* Updated docs and tests
* Removed unused import
* Updated signature
* Added support for infinity embeddings
* Added test cases
* Fixed tests and api base
* Updated docs and tests
* Removed unused import
* Updated signature
* Updated validate params
---------
Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
* fix InfinityEmbeddingConfig
---------
Co-authored-by: Prathamesh Saraf <pratamesh1867@gmail.com>
2025-04-21 20:01:29 -07:00
Ishaan Jaff
0c2f705417
[Feat] Add Responses API - Routing Affinity logic for sessions ( #10193 )
...
* test for test_responses_api_routing_with_previous_response_id
* test_responses_api_routing_with_previous_response_id
* add ResponsesApiDeploymentCheck
* ResponsesApiDeploymentCheck
* ResponsesApiDeploymentCheck
* fix ResponsesApiDeploymentCheck
* test_responses_api_routing_with_previous_response_id
* ResponsesApiDeploymentCheck
* test_responses_api_deployment_check.py
* docs routing affinity
* simplify ResponsesApiDeploymentCheck
* test response id
* fix code quality check
2025-04-21 20:00:27 -07:00
Ishaan Jaff
4eac0f64f3
[Feat] Pass through endpoints - ensure PassthroughStandardLoggingPayload
is logged and contains method, url, request/response body ( #10194 )
...
* ensure passthrough_logging_payload is filled in kwargs
* test_assistants_passthrough_logging
* test_assistants_passthrough_logging
* test_assistants_passthrough_logging
* test_threads_passthrough_logging
* test _init_kwargs_for_pass_through_endpoint
* _init_kwargs_for_pass_through_endpoint
2025-04-21 19:46:22 -07:00
Krrish Dholakia
4a50cf10fb
build: update ui build
Read Version from pyproject.toml / read-version (push) Successful in 23s
Helm unit test / unit-test (push) Successful in 25s
2025-04-21 16:26:36 -07:00
Krish Dholakia
89131d8ed3
Remove user_id from url ( #10192 )
...
* fix(user_dashboard.tsx): initial commit using user id from jwt instead of url
* fix(proxy_server.py): remove user id from url
fixes security issue around sharing url's
* fix(user_dashboard.tsx): handle user id being null
2025-04-21 16:22:57 -07:00
Krrish Dholakia
a34778dda6
build(ui/): update ui build
...
supports new non-user id in url flow
2025-04-21 16:22:28 -07:00
Krish Dholakia
0c3b7bb37d
fix(router.py): handle edge case where user sets 'model_group' inside… ( #10191 )
...
* fix(router.py): handle edge case where user sets 'model_group' inside 'model_info'
* fix(key_management_endpoints.py): security fix - return hashed token in 'token' field
Ensures when creating a key on UI - only hashed token shown
* test(test_key_management_endpoints.py): add unit test
* test: update test
2025-04-21 16:17:45 -07:00
Nilanjan De
03245c732a
Fix: Potential SQLi in spend_management_endpoints.py ( #9878 )
...
* fix: Potential SQLi in spend_management_endpoints.py
* fix tests
* test: add tests for global spend keys endpoint
* chore: update error message
* chore: lint
* chore: rename test
2025-04-21 14:29:38 -07:00
Li Yang
10257426a2
fix(bedrock): wrong system prompt transformation ( #10120 )
...
Read Version from pyproject.toml / read-version (push) Successful in 16s
Helm unit test / unit-test (push) Successful in 25s
* fix(bedrock): wrong system transformation
* chore: add one more test case
---------
Co-authored-by: Krish Dholakia <krrishdholakia@gmail.com>
2025-04-21 08:48:14 -07:00
Marty Sullivan
0b63c7a2eb
Model pricing updates for Azure & VertexAI ( #10178 )
Read Version from pyproject.toml / read-version (push) Successful in 23s
Helm unit test / unit-test (push) Successful in 27s
2025-04-20 11:33:45 -07:00
Krish Dholakia
ce828408da
fix(proxy_server.py): pass llm router to get complete model list ( #10176 )
...
Read Version from pyproject.toml / read-version (push) Successful in 21s
Helm unit test / unit-test (push) Successful in 27s
allows model auth to work
2025-04-19 22:27:49 -07:00
Krish Dholakia
e0a613f88a
fix(common_daily_activity.py): support empty entity id field ( #10175 )
...
* fix(common_daily_activity.py): support empty entity id field
allows returning empty response when user is not admin and does not belong to any team
* test(test_common_daily_activity.py): add unit testing
2025-04-19 22:20:28 -07:00
Ishaan Jaff
72f6bd3972
fix azure foundry phi error
2025-04-19 22:10:18 -07:00
Ishaan Jaff
36bcb3de4e
fix models appearing under test key page
2025-04-19 21:37:08 -07:00
Ishaan Jaff
1be36be72e
Litellm docs SCIM ( #10174 )
...
* docs scim
* docs SCIM stash
* docs litellm SCIM
* docs fix
* docs scim with LiteLLM
2025-04-19 18:29:09 -07:00
Krish Dholakia
55a17730fb
fix(transformation.py): pass back in gemini thinking content to api ( #10173 )
...
Ensures thinking content always returned
2025-04-19 18:03:05 -07:00
Krish Dholakia
bbfcb1ac7e
Litellm release notes 04 19 2025 ( #10169 )
...
* docs(index.md): initial draft release notes
* docs: note all pending docs
* build(model_prices_and_context_window.json): add o3, gpt-4.1, o4-mini pricing
* docs(vllm.md): update vllm doc to show file message type support
* docs(mistral.md): add mistral passthrough route doc
* docs(gemini.md): add gemini thinking to docs
* docs(vertex.md): add thinking/reasoning content for gemini models to docs
* docs(index.md): more links
* docs(index.md): add more links, images
* docs(index.md): cleanup highlights
2025-04-19 17:26:30 -07:00
Ishaan Jaff
f39d917886
[Docs] Responses API ( #10172 )
...
* docs litellm responses api
* doc fix
* docs responses API
* add get_supported_openai_params for LiteLLMCompletionResponsesConfig
* add Supported Responses API Parameters
2025-04-19 17:10:45 -07:00
Ishaan Jaff
7c3df984da
can_user_call_model ( #10170 )
Read Version from pyproject.toml / read-version (push) Successful in 51s
Helm unit test / unit-test (push) Successful in 51s
2025-04-19 16:46:00 -07:00
Ishaan Jaff
653570824a
Bug Fix - Responses API, Loosen restrictions on allowed environments for computer use tool ( #10168 )
...
* loosen allowed types on ComputerToolParam
* test_basic_computer_use_preview_tool_call
2025-04-19 14:40:32 -07:00
Ishaan Jaff
c80e984d7e
ui new build
2025-04-19 14:19:33 -07:00
Ishaan Jaff
b0024bb229
[Bug Fix] Spend Tracking Bug Fix, don't modify in memory default litellm params ( #10167 )
...
* _update_kwargs_with_default_litellm_params
* test_update_kwargs_does_not_mutate_defaults_and_merges_metadata
2025-04-19 14:13:59 -07:00
Ishaan Jaff
0717369ae6
[Feat] Expose Responses API on LiteLLM UI Test Key Page ( #10166 )
...
* add /responses API on UI
* add makeOpenAIResponsesRequest
* add makeOpenAIResponsesRequest
* fix add responses API on UI
* fix endpoint selector
* responses API render chunks on litellm chat ui
* fixes to streaming iterator
* fix render responses completed events
* fixes for MockResponsesAPIStreamingIterator
* transform_responses_api_request_to_chat_completion_request
* fix for responses API
* test_basic_openai_responses_api_streaming
* fix base responses api tests
2025-04-19 13:18:54 -07:00
Krish Dholakia
03b5399f86
test(utils.py): handle scenario where text tokens + reasoning tokens … ( #10165 )
...
* test(utils.py): handle scenario where text tokens + reasoning tokens set, but reasoning tokens not charged separately
Addresses https://github.com/BerriAI/litellm/pull/10141#discussion_r2051555332
* fix(vertex_and_google_ai_studio.py): only set content if non-empty str
2025-04-19 12:32:38 -07:00
Ong Khai Wei
99db1b7690
to get API key from environment viarble of WATSONX_APIKEY ( #10131 )
2025-04-19 11:25:14 -07:00