Commit graph

17999 commits

Author SHA1 Message Date
Krrish Dholakia
cc960da4b6 docs(azure.md): add o1 model support to config 2024-10-07 22:37:49 -07:00
Krrish Dholakia
9ee1a3ff8c bump: version 1.48.18 → 1.48.19 2024-10-07 22:22:02 -07:00
Krish Dholakia
6729c9ca7f
LiteLLM Minor Fixes & Improvements (10/07/2024) (#6101)
* fix(utils.py): support dropping temperature param for azure o1 models

* fix(main.py): handle azure o1 streaming requests

o1 doesn't support streaming, fake it to ensure code works as expected

* feat(utils.py): expose `hosted_vllm/` endpoint, with tool handling for vllm

Fixes https://github.com/BerriAI/litellm/issues/6088

* refactor(internal_user_endpoints.py): cleanup unused params + update docstring

Closes https://github.com/BerriAI/litellm/issues/6100

* fix(main.py): expose custom image generation api support

Fixes https://github.com/BerriAI/litellm/issues/6097

* fix: fix linting errors

* docs(custom_llm_server.md): add docs on custom api for image gen calls

* fix(types/utils.py): handle dict type

* fix(types/utils.py): fix linting errors
2024-10-07 22:17:22 -07:00
Ishaan Jaff
5de69cb1b2 fix using Dockerfile 2024-10-08 08:45:40 +05:30
Ishaan Jaff
59b247ab23 fix config.yml 2024-10-08 08:36:03 +05:30
Ishaan Jaff
d742e8cb43
(clean up) move docker files from root to docker folder (#6109)
* fix move docker files to docker folders

* move check file length

* fix docker hub deploy
2024-10-08 08:23:52 +05:30
Ishaan Jaff
ef815f3a84
(docs) add remaining litellm settings on configs.md doc (#6108)
* docs add litellm settings configs

* docs langfuse tags on config
2024-10-08 07:57:04 +05:30
Ishaan Jaff
2b370f8e9e
(docs) key based callbacks (#6107) 2024-10-08 07:12:01 +05:30
Pradyumna Singh Rathore
b7ba558b74
fix links due to broken list (#6103) 2024-10-07 15:47:29 -04:00
Ishaan Jaff
5afc45d411 bump: version 1.48.17 → 1.48.18 2024-10-07 18:22:21 +05:30
Ishaan Jaff
b1e9d344b2
Update readme.md 2024-10-07 18:15:15 +05:30
Ishaan Jaff
a0cbf31fcf
Update readme.md 2024-10-07 18:12:43 +05:30
Ishaan Jaff
1bafbf8382
(feat proxy) add v2 maintained LiteLLM grafana dashboard (#6098)
* add new grafana dashboard litellm

* add v2 grafana dashboard
2024-10-07 18:11:20 +05:30
Ishaan Jaff
2c8bba293f
(bug fix) TTL not being set for embedding caching requests (#6095)
* fix ttl for cache pipeline settings

* add test for caching

* add test for setting ttls on redis caching
2024-10-07 15:53:18 +05:30
Ishaan Jaff
285b589095 ui new build 2024-10-07 13:01:19 +05:30
Ishaan Jaff
51af0d5d94
(proxy ui sso flow) - fix invite user sso flow (#6093)
* return if sso setup on ui_settings

* use helper to get invite link
2024-10-07 12:32:08 +05:30
Ishaan Jaff
a7628317cd
(proxy ui) - fix view user pagination (#6094)
* ui - fix view user pagination

* add new internal user test
2024-10-07 12:31:55 +05:30
Ishaan Jaff
abe8059713 ui - fix view user pagination 2024-10-07 12:15:29 +05:30
kvadros
e007bb65b5
Proxy: include customer budget in responses (#5977) 2024-10-07 10:05:28 +05:30
Ishaan Jaff
b2fbee3923 docs key logging 2024-10-06 13:49:27 +05:30
Ishaan Jaff
fd7014a326 correct use of healthy / unhealthy 2024-10-06 13:48:30 +05:30
Krish Dholakia
49d8b2be46
fix(utils.py): fix fix pydantic obj to schema creation for vertex en… (#6071)
* fix(utils.py): fix  fix pydantic obj to schema creation for vertex endpoints

Fixes https://github.com/BerriAI/litellm/issues/6027

* test(test_completion.pyu): skip test - avoid hitting gemini rate limits

* fix(common_utils.py): fix ruff linting error
2024-10-06 00:25:55 -04:00
Krrish Dholakia
29da2d49d6 bump: version 1.48.16 → 1.48.17 2024-10-05 21:27:22 -04:00
Krish Dholakia
04e5963b65
Litellm expose disable schema update flag (#6085)
* fix: enable new 'disable_prisma_schema_update' flag

* build(config.yml): remove setup remote docker step

* ci(config.yml): give container time to start up

* ci(config.yml): update test

* build(config.yml): actually start docker

* build(config.yml): simplify grep check

* fix(prisma_client.py): support reading disable_schema_update via env vars

* ci(config.yml): add test to check if all general settings are documented

* build(test_General_settings.py): check available dir

* ci: check ../ repo path

* build: check ./

* build: fix test
2024-10-05 21:26:51 -04:00
Krish Dholakia
f2c0a31e3c
LiteLLM Minor Fixes & Improvements (10/05/2024) (#6083)
* docs(prompt_caching.md): add prompt caching cost calc example to docs

* docs(prompt_caching.md): add proxy examples to docs

* feat(utils.py): expose new helper `supports_prompt_caching()` to check if a model supports prompt caching

* docs(prompt_caching.md): add docs on checking model support for prompt caching

* build: fix invalid json
2024-10-05 18:59:11 -04:00
Krish Dholakia
fac3b2ee42
Add pyright to ci/cd + Fix remaining type-checking errors (#6082)
* fix: fix type-checking errors

* fix: fix additional type-checking errors

* fix: additional type-checking error fixes

* fix: fix additional type-checking errors

* fix: additional type-check fixes

* fix: fix all type-checking errors + add pyright to ci/cd

* fix: fix incorrect import

* ci(config.yml): use mypy on ci/cd

* fix: fix type-checking errors in utils.py

* fix: fix all type-checking errors on main.py

* fix: fix mypy linting errors

* fix(anthropic/cost_calculator.py): fix linting errors

* fix: fix mypy linting errors

* fix: fix linting errors
2024-10-05 17:04:00 -04:00
Ishaan Jaff
f7ce1173f3 bump: version 1.48.15 → 1.48.16 2024-10-05 16:59:16 +05:30
Ishaan Jaff
3cb04480fb
(code clean up) use a folder for gcs bucket logging + add readme in folder (#6080)
* refactor gcs bucket

* add readme
2024-10-05 16:58:10 +05:30
Ishaan Jaff
6e6d38841f docs fix 2024-10-05 15:25:25 +05:30
GTonehour
d533acd24a
openrouter/openai's litellm_provider should be openrouter, not openai (#6079)
In model_prices_and_context_window.json, openrouter/* models all have litellm_provider set as "openrouter", except for four openrouter/openai/* models, which were set to "openai".
I suppose they must be set to "openrouter", so one can know it should use the openrouter API for this model.
2024-10-05 15:20:44 +05:30
Ishaan Jaff
ab0b536143
(feat) add azure openai cost tracking for prompt caching (#6077)
* add azure o1 models to model cost map

* add azure o1 cost tracking

* fix azure cost calc

* add get llm provider test
2024-10-05 15:04:18 +05:30
Ishaan Jaff
7267852511 linting error fix 2024-10-05 15:03:39 +05:30
Ishaan Jaff
5ee1342d37
(docs) reference router settings general settings etc (#6078) 2024-10-05 15:01:28 +05:30
Ishaan Jaff
d2f17cf97c docs routing config table 2024-10-05 14:40:07 +05:30
Ishaan Jaff
530915da51 add o-1 to Azure docs 2024-10-05 14:23:54 +05:30
Ishaan Jaff
3682f661d8
(feat) add cost tracking for OpenAI prompt caching (#6055)
* add cache_read_input_token_cost for prompt caching models

* add prompt caching for latest models

* add openai cost calculator

* add openai prompt caching test

* fix lint check

* add not on how usage._cache_read_input_tokens is used

* fix cost calc whisper openai

* use output_cost_per_second

* add input_cost_per_second
2024-10-05 14:20:15 +05:30
Ishaan Jaff
930606ad63
add azure o1 models to model cost map (#6075) 2024-10-05 13:22:06 +05:30
Ishaan Jaff
c84cfe977e
(feat) add /key/health endpoint to test key based logging (#6073)
* add /key/health endpoint

* add /key/health endpoint

* fix return from /key/health

* update doc string

* fix doc string for /key/health

* add test for /key/health

* fix linting

* docs /key/health
2024-10-05 11:56:55 +05:30
Krish Dholakia
4e921bee2b
fix(gcs_bucket.py): show error response text in exception (#6072) 2024-10-05 11:56:43 +05:30
Krrish Dholakia
4c9dea9f36 bump: version 1.48.14 → 1.48.15 2024-10-04 21:32:45 -04:00
Krish Dholakia
2e5c46ef6d
LiteLLM Minor Fixes & Improvements (10/04/2024) (#6064)
* fix(litellm_logging.py): ensure cache hits are scrubbed if 'turn_off_message_logging' is enabled

* fix(sagemaker.py): fix streaming to raise error immediately

Fixes https://github.com/BerriAI/litellm/issues/6054

* (fixes)  gcs bucket key based logging  (#6044)

* fixes for gcs bucket logging

* fix StandardCallbackDynamicParams

* fix - gcs logging when payload is not serializable

* add test_add_callback_via_key_litellm_pre_call_utils_gcs_bucket

* working success callbacks

* linting fixes

* fix linting error

* add type hints to functions

* fixes for dynamic success and failure logging

* fix for test_async_chat_openai_stream

* fix handle case when key based logging vars are set as os.environ/ vars

* fix prometheus track cooldown events on custom logger (#6060)

* (docs) add 1k rps load test doc  (#6059)

* docs 1k rps load test

* docs load testing

* docs load testing litellm

* docs load testing

* clean up load test doc

* docs prom metrics for load testing

* docs using prometheus on load testing

* doc load testing with prometheus

* (fixes) docs + qa - gcs key based logging  (#6061)

* fixes for required values for gcs bucket

* docs gcs bucket logging

* bump: version 1.48.12 → 1.48.13

* ci/cd run again

* bump: version 1.48.13 → 1.48.14

* update load test doc

* (docs) router settings - on litellm config  (#6037)

* add yaml with all router settings

* add docs for router settings

* docs router settings litellm settings

* (feat)  OpenAI prompt caching models to model cost map (#6063)

* add prompt caching for latest models

* add cache_read_input_token_cost for prompt caching models

* fix(litellm_logging.py): check if param is iterable

Fixes https://github.com/BerriAI/litellm/issues/6025#issuecomment-2393929946

* fix(factory.py): support passing an 'assistant_continue_message' to prevent bedrock error

Fixes https://github.com/BerriAI/litellm/issues/6053

* fix(databricks/chat): handle streaming responses

* fix(factory.py): fix linting error

* fix(utils.py): unify anthropic + deepseek prompt caching information to openai format

Fixes https://github.com/BerriAI/litellm/issues/6069

* test: fix test

* fix(types/utils.py): support all openai roles

Fixes https://github.com/BerriAI/litellm/issues/6052

* test: fix test

---------

Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
2024-10-04 21:28:53 -04:00
Ishaan Jaff
fc6e0dd6cb
(feat) OpenAI prompt caching models to model cost map (#6063)
* add prompt caching for latest models

* add cache_read_input_token_cost for prompt caching models
2024-10-04 19:12:13 +05:30
Ishaan Jaff
6d1de8e1ee
(docs) router settings - on litellm config (#6037)
* add yaml with all router settings

* add docs for router settings

* docs router settings litellm settings
2024-10-04 18:59:01 +05:30
Ishaan Jaff
0c9c42915f update load test doc 2024-10-04 18:47:26 +05:30
Ishaan Jaff
45a981a37e bump: version 1.48.13 → 1.48.14 2024-10-04 17:19:33 +05:30
Ishaan Jaff
3c59d188ef ci/cd run again 2024-10-04 17:19:26 +05:30
Ishaan Jaff
69c96d9ba4 bump: version 1.48.12 → 1.48.13 2024-10-04 17:18:49 +05:30
Ishaan Jaff
e394ed1e5b
(fixes) docs + qa - gcs key based logging (#6061)
* fixes for required values for gcs bucket

* docs gcs bucket logging
2024-10-04 16:58:04 +05:30
Ishaan Jaff
2449d258cf
(docs) add 1k rps load test doc (#6059)
* docs 1k rps load test

* docs load testing

* docs load testing litellm

* docs load testing

* clean up load test doc

* docs prom metrics for load testing

* docs using prometheus on load testing

* doc load testing with prometheus
2024-10-04 16:56:34 +05:30
Ishaan Jaff
224460d4c9
fix prometheus track cooldown events on custom logger (#6060) 2024-10-04 16:56:22 +05:30