litellm-mirror/litellm
Krish Dholakia 0a2a51a5a5 UI - Allow admin to control default model access for internal users (#8912)
* fix(create_user_button.tsx): allow admin to set models user has access to, on invite

Enables controlling model access on invite

* feat(auth_checks.py): enforce 'no-model-access' special model name on backend

prevent user from calling models if default key has no model access

* fix(chat_ui.tsx): allow user to input custom model

* fix(chat_ui.tsx): pull available models based on models key has access to

* style(create_user_button.tsx): move default model inside 'personal key creation' accordion

* fix(chat_ui.tsx): fix linting error

* test(test_auth_checks.py): add unit-test for special model name

* docs(internal_user_endpoints.py): update docstring

* fix test_moderations_bad_model

* Litellm dev 02 27 2025 p6 (#8891)

* fix(http_parsing_utils.py): orjson can throw errors on some emoji's in text, default to json.loads

* fix(sagemaker/handler.py): support passing model id on async streaming

* fix(litellm_pre_call_utils.py): Fixes https://github.com/BerriAI/litellm/issues/7237

* Fix calling claude via invoke route + response_format support for claude on invoke route (#8908)

* fix(anthropic_claude3_transformation.py): fix amazon anthropic claude 3 tool calling transformation on invoke route

move to using anthropic config as base

* fix(utils.py): expose anthropic config via providerconfigmanager

* fix(llm_http_handler.py): support json mode on async completion calls

* fix(invoke_handler/make_call): support json mode for anthropic called via bedrock invoke

* fix(anthropic/): handle 'response_format: {"type": "text"}` + migrate amazon claude 3 invoke config to inherit from anthropic config

Prevents error when passing in 'response_format: {"type": "text"}

* test: fix test

* fix(utils.py): fix base invoke provider check

* fix(anthropic_claude3_transformation.py): don't pass 'stream' param

* fix: fix linting errors

* fix(converse_transformation.py): handle response_format type=text for converse

* converse_transformation: pass 'description' if set in response_format (#8907)

* test(test_bedrock_completion.py): e2e test ensuring tool description is passed in

* fix(converse_transformation.py): pass description, if set

* fix(transformation.py): Fixes https://github.com/BerriAI/litellm/issues/8767#issuecomment-2689887663

* Fix bedrock passing `response_format: {"type": "text"}` (#8900)

* fix(converse_transformation.py): ignore type: text, value in response_format

no-op for bedrock

* fix(converse_transformation.py): handle adding response format value to tools

* fix(base_invoke_transformation.py): fix 'get_bedrock_invoke_provider' to handle cross-region-inferencing models

* test(test_bedrock_completion.py): add unit testing for bedrock invoke provider logic

* test: update test

* fix(exception_mapping_utils.py): add context window exceeded error handling for databricks provider route

* fix(fireworks_ai/): support passing tools + response_format together

* fix: cleanup

* fix(base_invoke_transformation.py): fix imports

* (Feat) - Show Error Logs on LiteLLM UI  (#8904)

* fix test_moderations_bad_model

* use async_post_call_failure_hook

* basic logging errors in DB

* show status on ui

* show status on ui

* ui show request / response side by side

* stash fixes

* working, track raw request

* track error info in metadata

* fix showing error / request / response logs

* show traceback on error viewer

* ui with traceback of error

* fix async_post_call_failure_hook

* fix(http_parsing_utils.py): orjson can throw errors on some emoji's in text, default to json.loads

* test_get_error_information

* fix code quality

* rename proxy track cost callback test

* _should_store_errors_in_spend_logs

* feature flag error logs

* Revert "_should_store_errors_in_spend_logs"

This reverts commit 7f345df477.

* Revert "feature flag error logs"

This reverts commit 0e90c022bb.

* test_spend_logs_payload

* fix OTEL log_db_metrics

* fix import json

* fix ui linting error

* test_async_post_call_failure_hook

* test_chat_completion_bad_model_with_spend_logs

---------

Co-authored-by: Krrish Dholakia <krrishdholakia@gmail.com>

* ui new build

* test_chat_completion_bad_model_with_spend_logs

* docs(release_cycle.md): document release cycle

* bump: version 1.62.0 → 1.62.1

---------

Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
2025-02-28 23:23:03 -08:00
..
adapters (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
assistants Revert "fix: add missing parameters order, limit, before, and after in get_as…" (#7542) 2025-01-03 16:32:12 -08:00
batch_completion (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
batches (Feat - Batches API) add support for retrieving vertex api batch jobs (#7661) 2025-01-09 18:35:03 -08:00
caching (Redis fix) - use mget_non_atomic (#8682) 2025-02-20 17:51:31 -08:00
files (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
fine_tuning fix linting 2025-02-14 21:42:51 -08:00
integrations Litellm contributor prs 02 24 2025 (#8775) 2025-02-24 18:55:48 -08:00
litellm_core_utils (Feat) - Show Error Logs on LiteLLM UI (#8904) 2025-02-28 20:10:09 -08:00
llms Fix bedrock passing response_format: {"type": "text"} (#8900) 2025-02-28 20:09:59 -08:00
proxy UI - Allow admin to control default model access for internal users (#8912) 2025-02-28 23:23:03 -08:00
realtime_api (Refactor) - Re use litellm.completion/litellm.embedding etc for health checks (#7455) 2024-12-28 18:38:54 -08:00
rerank_api Add new gpt-4.5-preview model + other updates (#8879) 2025-02-27 15:27:14 -08:00
router_strategy fix code quality 2025-02-18 21:29:23 -08:00
router_utils (Router) - If allowed_fails or allowed_fail_policy set, use that for single deployment cooldown logic (#8668) 2025-02-25 15:15:01 -08:00
secret_managers fix: add default credential for azure (#7095) (#7891) 2025-01-21 09:01:49 -08:00
types (Feat) - Show Error Logs on LiteLLM UI (#8904) 2025-02-28 20:10:09 -08:00
__init__.py Fix calling claude via invoke route + response_format support for claude on invoke route (#8908) 2025-02-28 17:56:26 -08:00
_logging.py (sdk perf fix) - only print args passed to litellm when debugging mode is on (#7708) 2025-01-11 22:56:20 -08:00
_redis.py (Redis Cluster) - Fixes for using redis cluster + pipeline (#8442) 2025-02-12 18:01:32 -08:00
_service_logger.py fix svc logger (#7727) 2025-01-12 22:00:25 -08:00
_version.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
budget_manager.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
constants.py Fix calling claude via invoke route + response_format support for claude on invoke route (#8908) 2025-02-28 17:56:26 -08:00
cost.json
cost_calculator.py Add cohere v2/rerank support (#8421) (#8605) 2025-02-22 22:25:29 -08:00
exceptions.py fix(main.py): fix key leak error when unknown provider given (#8556) 2025-02-15 14:02:55 -08:00
main.py fix(main.py): pass 'thinking' param on async completion call 2025-02-26 23:16:39 -08:00
model_prices_and_context_window_backup.json Install Node.js 2025-02-27 21:09:04 -08:00
py.typed feature - Types for mypy - #360 2024-05-30 14:14:41 -04:00
router.py (Observability) - Add more detailed dd tracing on Proxy Auth, Bedrock Auth (#8693) 2025-02-20 18:00:41 -08:00
scheduler.py (refactor) caching use LLMCachingHandler for async_get_cache and set_cache (#6208) 2024-10-14 16:34:01 +05:30
timeout.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
utils.py Fix calling claude via invoke route + response_format support for claude on invoke route (#8908) 2025-02-28 17:56:26 -08:00