litellm/docs/my-website/docs
Krish Dholakia 1e403a8447
Litellm dev 10 29 2024 (#6502)
* fix(core_helpers.py): return None, instead of raising kwargs is None error

Closes https://github.com/BerriAI/litellm/issues/6500

* docs(cost_tracking.md): cleanup doc

* fix(vertex_and_google_ai_studio.py): handle function call with no params passed in

Closes https://github.com/BerriAI/litellm/issues/6495

* test(test_router_timeout.py): add test for router timeout + retry logic

* test: update test to use module level values

* (fix) Prometheus - Log Postgres DB latency, status on prometheus  (#6484)

* fix logging DB fails on prometheus

* unit testing log to otel wrapper

* unit testing for service logger + prometheus

* use LATENCY buckets for service logging

* fix service logging

* docs clarify vertex vs gemini

* (router_strategy/) ensure all async functions use async cache methods (#6489)

* fix router strat

* use async set / get cache in router_strategy

* add coverage for router strategy

* fix imports

* fix batch_get_cache

* use async methods for least busy

* fix least busy use async methods

* fix test_dual_cache_increment

* test async_get_available_deployment when routing_strategy="least-busy"

* (fix) proxy - fix when `STORE_MODEL_IN_DB` should be set (#6492)

* set store_model_in_db at the top

* correctly use store_model_in_db global

* (fix) `PrometheusServicesLogger` `_get_metric` should return metric in Registry  (#6486)

* fix logging DB fails on prometheus

* unit testing log to otel wrapper

* unit testing for service logger + prometheus

* use LATENCY buckets for service logging

* fix service logging

* fix _get_metric in prom services logger

* add clear doc string

* unit testing for prom service logger

* bump: version 1.51.0 → 1.51.1

* Add `azure/gpt-4o-mini-2024-07-18` to model_prices_and_context_window.json (#6477)

* Update utils.py (#6468)

Fixed missing keys

* (perf) Litellm redis router fix - ~100ms improvement (#6483)

* docs(exception_mapping.md): add missing exception types

Fixes https://github.com/Aider-AI/aider/issues/2120#issuecomment-2438971183

* fix(main.py): register custom model pricing with specific key

Ensure custom model pricing is registered to the specific model+provider key combination

* test: make testing more robust for custom pricing

* fix(redis_cache.py): instrument otel logging for sync redis calls

ensures complete coverage for all redis cache calls

* refactor: pass parent_otel_span for redis caching calls in router

allows for more observability into what calls are causing latency issues

* test: update tests with new params

* refactor: ensure e2e otel tracing for router

* refactor(router.py): add more otel tracing acrosss router

catch all latency issues for router requests

* fix: fix linting error

* fix(router.py): fix linting error

* fix: fix test

* test: fix tests

* fix(dual_cache.py): pass ttl to redis cache

* fix: fix param

* perf(cooldown_cache.py): improve cooldown cache, to store cache results in memory for 5s, prevents redis call from being made on each request

reduces 100ms latency per call with caching enabled on router

* fix: fix test

* fix(cooldown_cache.py): handle if a result is None

* fix(cooldown_cache.py): add debug statements

* refactor(dual_cache.py): move to using an in-memory check for batch get cache, to prevent redis from being hit for every call

* fix(cooldown_cache.py): fix linting erropr

* refactor(prometheus.py): move to using standard logging payload for reading the remaining request / tokens

Ensures prometheus token tracking works for anthropic as well

* fix: fix linting error

* fix(redis_cache.py): make sure ttl is always int (handle float values)

Fixes issue where redis_client.ex was not working correctly due to float ttl

* fix: fix linting error

* test: update test

* fix: fix linting error

---------

Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
Co-authored-by: Xingyao Wang <xingyao@all-hands.dev>
Co-authored-by: vibhanshu-ob <115142120+vibhanshu-ob@users.noreply.github.com>
2024-10-29 22:04:16 -07:00
..
caching (refactor) caching use LLMCachingHandler for async_get_cache and set_cache (#6208) 2024-10-14 16:34:01 +05:30
completion doc - using gpt-4o-audio-preview (#6326) 2024-10-19 09:34:56 +05:30
debugging docs - fix doc build time errors 2024-06-15 14:58:02 -07:00
embedding LiteLLM Minor Fixes & Improvements (10/24/2024) (#6421) 2024-10-25 15:55:56 -07:00
extras docs(code_quality.md): add doc on litellm code qa 2024-10-02 11:20:15 -04:00
langchain update langchain documentation to reflect refactor 2024-05-07 17:01:55 +05:30
observability (feat) Arize - Allow using Arize HTTP endpoint (#6364) 2024-10-23 09:38:35 +05:30
pass_through docs(pass_through/bedrock.md): add bedrock agents support 2024-09-04 22:30:14 -07:00
projects add dbally project 2024-08-22 23:21:40 +02:00
providers LiteLLM Minor Fixes & Improvements (10/28/2024) (#6475) 2024-10-29 17:20:24 -07:00
proxy Litellm dev 10 29 2024 (#6502) 2024-10-29 22:04:16 -07:00
tutorials docs: fix dead links 2024-08-26 13:28:25 -07:00
anthropic_completion.md docs(anthropic_completion.md): add doc on anthropic /v1/messages endpoint support 2024-07-10 22:56:33 -07:00
assistants.md [Feat-Proxy] Add Azure Assistants API - Create Assistant, Delete Assistant Support (#5777) 2024-09-18 16:27:33 -07:00
audio_transcription.md docs(image_generation.md): add proxy image gen example to docs 2024-07-09 11:56:21 -07:00
batches.md [Feat-Proxy] Add Azure Assistants API - Create Assistant, Delete Assistant Support (#5777) 2024-09-18 16:27:33 -07:00
budget_manager.md docs - use consistent name for LiteLLM proxy server 2024-08-03 12:54:07 -07:00
contact.md
contributing.md docs(contributing.md): add docs for contributing to litellm ui 2024-02-23 13:01:34 -08:00
data_security.md docs(data_security.md): cleanup docs 2024-10-01 15:33:10 -04:00
default_code_snippet.md
enterprise.md Merge branch 'main' into litellm_allow_setting_route_controls 2024-09-03 17:00:22 -07:00
exception_mapping.md Litellm dev 10 26 2024 (#6472) 2024-10-28 15:05:43 -07:00
fine_tuning.md docs native vertex ft endpoint 2024-08-03 11:56:43 -07:00
getting_started.md Add Helicone 2024-07-08 15:41:35 -07:00
hosted.md docs gdpr regions cloud 2024-07-10 18:00:16 -07:00
image_generation.md docs(image_generation.md): add proxy image gen example to docs 2024-07-09 11:56:21 -07:00
index.md docs docker quick start 2024-09-19 15:10:59 -07:00
load_test.md (docs) add 1k rps load test doc (#6059) 2024-10-04 16:56:34 +05:30
load_test_advanced.md update load test doc 2024-10-04 18:47:26 +05:30
load_test_rpm.md (docs) add 1k rps load test doc (#6059) 2024-10-04 16:56:34 +05:30
load_test_sdk.md (docs) add 1k rps load test doc (#6059) 2024-10-04 16:56:34 +05:30
migration.md
migration_policy.md docs migration policy 2024-08-09 18:06:37 -07:00
oidc.md (oidc): Improve docs for unofficial provider. 2024-08-16 20:30:41 +00:00
old_guardrails.md docs move lakera to free 2024-08-20 16:38:37 -07:00
projects.md
prompt_injection.md docs(sidebars.js): refactor docs 2024-08-22 18:22:50 -07:00
proxy_api.md
proxy_server.md docs - use consistent name for LiteLLM proxy server 2024-08-03 12:54:07 -07:00
realtime.md Litellm dev 10 22 2024 (#6384) 2024-10-22 21:18:54 -07:00
rerank.md docs(azure_ai.md): add rerank api endpoint to docs 2024-09-17 23:06:19 -07:00
routing.md docs - show correct rpm - > tpm conversion for Azure 2024-09-27 17:18:55 -07:00
rules.md
scheduler.md docs(scheduler.md): cleanup docs to use /chat/completion endpoint 2024-08-07 21:49:06 -07:00
sdk_custom_pricing.md organize docs 2024-08-03 12:54:07 -07:00
secret.md docs secret manager link 2024-09-03 11:41:43 -07:00
set_keys.md feat(utils.py): unify common auth params across azure/vertex_ai/bedrock/watsonx 2024-04-27 11:06:18 -07:00
simple_proxy_old_doc.md (docs + testing) Correctly document the timeout value used by litellm proxy is 6000 seconds + add to best practices for prod (#6339) 2024-10-23 14:09:35 +05:30
text_to_speech.md docs tts 2024-08-23 18:29:36 -07:00
troubleshoot.md docs(troubleshoot.md): cleanup docs 2024-05-27 09:51:35 -07:00