litellm-mirror/litellm
Krish Dholakia f30260343b Litellm dev 12 26 2024 p3 (#7434)
* build(model_prices_and_context_window.json): update groq models to specify 'supports_vision' parameter

Closes https://github.com/BerriAI/litellm/issues/7433

* docs(groq.md): add groq vision example to docs

Closes https://github.com/BerriAI/litellm/issues/7433

* fix(prometheus.py): refactor self.litellm_proxy_failed_requests_metric to use label factory

* feat(prometheus.py): new 'litellm_proxy_failed_requests_by_tag_metric'

allows tracking failed requests by tag on proxy

* fix(prometheus.py): fix exception logging

* feat(prometheus.py): add new 'litellm_request_total_latency_by_tag_metric'

enables tracking latency by use-case

* feat(prometheus.py): add new llm api latency by tag metric

* feat(prometheus.py): new litellm_deployment_latency_per_output_token_by_tag metric

allows tracking deployment latency by tag

* fix(prometheus.py): refactor 'litellm_requests_metric' to use enum values + label factory

* feat(prometheus.py): new litellm_proxy_total_requests_by_tag metric

allows tracking total requests by tag

* feat(prometheus.py): new metric litellm_deployment_successful_fallbacks_by_tag

allows tracking deployment fallbacks by tag

* fix(prometheus.py): new 'litellm_deployment_failed_fallbacks_by_tag' metric

allows tracking failed fallbacks on deployment by custom tag

* test: fix test

* test: rename test to run earlier

* test: skip flaky test
2024-12-26 21:21:16 -08:00
..
adapters (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
assistants (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
batch_completion (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
batches (Feat) add `"/v1/batches/{batch_id:path}/cancel" endpoint (#7406) 2024-12-24 20:23:50 -08:00
caching (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
deprecated_litellm_server (refactor) caching use LLMCachingHandler for async_get_cache and set_cache (#6208) 2024-10-14 16:34:01 +05:30
files (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
fine_tuning (Feat) Add logging for POST v1/fine_tuning/jobs (#7426) 2024-12-26 08:58:47 -08:00
integrations Litellm dev 12 26 2024 p3 (#7434) 2024-12-26 21:21:16 -08:00
litellm_core_utils (fix) initializing OTEL Logging on LiteLLM Proxy - ensure OTEL logger is initialized only once (#7435) 2024-12-26 21:17:19 -08:00
llms Litellm dev 12 25 2025 p2 (#7420) 2024-12-25 18:35:34 -08:00
proxy (fix) initializing OTEL Logging on LiteLLM Proxy - ensure OTEL logger is initialized only once (#7435) 2024-12-26 21:17:19 -08:00
realtime_api (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
rerank_api (feat) /batches - track user_api_key_alias, user_api_key_team_alias etc for /batch requests (#7401) 2024-12-24 17:44:28 -08:00
router_strategy Support budget/rate limit tiers for keys (#7429) 2024-12-26 19:05:27 -08:00
router_utils Controll fallback prompts client-side (#7334) 2024-12-20 19:09:53 -08:00
secret_managers (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
types Litellm dev 12 26 2024 p3 (#7434) 2024-12-26 21:21:16 -08:00
__init__.py (fix) initializing OTEL Logging on LiteLLM Proxy - ensure OTEL logger is initialized only once (#7435) 2024-12-26 21:17:19 -08:00
_logging.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
_redis.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
_service_logger.py LiteLLM Minor Fixes & Improvements (12/05/2024) (#7037) 2024-12-05 00:02:31 -08:00
_version.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
budget_manager.py (code quality) run ruff rule to ban unused imports (#7313) 2024-12-19 12:33:42 -08:00
constants.py (feat) /batches - track user_api_key_alias, user_api_key_team_alias etc for /batch requests (#7401) 2024-12-24 17:44:28 -08:00
cost.json
cost_calculator.py LiteLLM Minor Fixes & Improvements (12/23/2024) - p3 (#7394) 2024-12-23 22:02:52 -08:00
exceptions.py Litellm 12 02 2024 (#6994) 2024-12-02 22:00:01 -08:00
main.py Litellm dev 12 25 2025 p2 (#7420) 2024-12-25 18:35:34 -08:00
model_prices_and_context_window_backup.json Litellm dev 12 26 2024 p3 (#7434) 2024-12-26 21:21:16 -08:00
py.typed
router.py Support budget/rate limit tiers for keys (#7429) 2024-12-26 19:05:27 -08:00
scheduler.py (refactor) caching use LLMCachingHandler for async_get_cache and set_cache (#6208) 2024-10-14 16:34:01 +05:30
timeout.py Litellm ruff linting enforcement (#5992) 2024-10-01 19:44:20 -04:00
utils.py (Feat) Add logging for POST v1/fine_tuning/jobs (#7426) 2024-12-26 08:58:47 -08:00