Commit graph

870 commits

Author SHA1 Message Date
Ishaan Jaff
a22e473636 set timeout on predibase test 2024-09-18 17:13:13 -07:00
Ishaan Jaff
8762b64b1d ci/cd run again 2024-09-16 18:26:53 -07:00
Ishaan Jaff
85acdb9193
[Feat] Add max_completion_tokens param (#5691)
* add max_completion_tokens

* add max_completion_tokens

* add max_completion_tokens support for OpenAI models

* add max_completion_tokens param

* add max_completion_tokens for bedrock converse models

* add test for converse maxTokens

* fix openai o1 param mapping test

* move test optional params

* add max_completion_tokens for anthropic api

* fix conftest

* add max_completion tokens for vertex ai partner models

* add max_completion_tokens for fireworks ai

* add max_completion_tokens for hf rest api

* add test for param mapping

* add param mapping for vertex, gemini + testing

* predibase is the most unstable and unusable llm api in prod, can't handle our ci/cd

* add max_completion_tokens to openai supported params

* fix fireworks ai param mapping
2024-09-14 14:57:01 -07:00
Krish Dholakia
60709a0753
LiteLLM Minor Fixes and Improvements (09/13/2024) (#5689)
* refactor: cleanup unused variables + fix pyright errors

* feat(health_check.py): Closes https://github.com/BerriAI/litellm/issues/5686

* fix(o1_reasoning.py): add stricter check for o-1 reasoning model

* refactor(mistral/): make it easier to see mistral transformation logic

* fix(openai.py): fix openai o-1 model param mapping

Fixes https://github.com/BerriAI/litellm/issues/5685

* feat(main.py): infer finetuned gemini model from base model

Fixes https://github.com/BerriAI/litellm/issues/5678

* docs(vertex.md): update docs to call finetuned gemini models

* feat(proxy_server.py): allow admin to hide proxy model aliases

Closes https://github.com/BerriAI/litellm/issues/5692

* docs(load_balancing.md): add docs on hiding alias models from proxy config

* fix(base.py): don't raise notimplemented error

* fix(user_api_key_auth.py): fix model max budget check

* fix(router.py): fix elif

* fix(user_api_key_auth.py): don't set team_id to empty str

* fix(team_endpoints.py): fix response type

* test(test_completion.py): handle predibase error

* test(test_proxy_server.py): fix test

* fix(o1_transformation.py): fix max_completion_token mapping

* test(test_image_generation.py): mark flaky test
2024-09-14 10:02:55 -07:00
Ishaan Jaff
cd8d7ca915
[Fix] Performance - use in memory cache when downloading images from a url (#5657)
* fix use in memory cache when getting images

* fix linting

* fix load testing

* fix load test size

* fix load test size

* trigger ci/cd again
2024-09-13 07:23:42 -07:00
Ishaan Jaff
bea34c9231 fix config.yml 2024-09-12 09:28:45 -07:00
Ishaan Jaff
90d096b639 ci/cd run again 2024-09-12 08:42:34 -07:00
Ishaan Jaff
d038568be4 ci/cd run again 2024-09-12 08:31:17 -07:00
Ishaan Jaff
ce5182710f ci/cd run again 2024-09-10 19:21:13 -07:00
Krish Dholakia
2d2282101b
LiteLLM Minor Fixes and Improvements (09/09/2024) (#5602)
* fix(main.py): pass default azure api version as alternative in completion call

Fixes api error caused due to api version

Closes https://github.com/BerriAI/litellm/issues/5584

* Fixed gemini-1.5-flash pricing (#5590)

* add /key/list endpoint

* bump: version 1.44.21 → 1.44.22

* docs architecture

* Fixed gemini-1.5-flash pricing

---------

Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>

* fix(bedrock/chat.py): fix converse api stop sequence param mapping

Fixes https://github.com/BerriAI/litellm/issues/5592

* fix(databricks/cost_calculator.py): handle databricks model name changes

Fixes https://github.com/BerriAI/litellm/issues/5597

* fix(azure.py): support azure api version 2024-08-01-preview

Closes https://github.com/BerriAI/litellm/issues/5377

* fix(proxy/_types.py): allow dev keys to call cohere /rerank endpoint

Fixes issue where only admin could call rerank endpoint

* fix(azure.py): check if model is gpt-4o

* fix(proxy/_types.py): support /v1/rerank on non-admin routes as well

* fix(cost_calculator.py): fix split on `/` logic in cost calculator

---------

Co-authored-by: F1bos <44951186+F1bos@users.noreply.github.com>
Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
2024-09-09 21:56:12 -07:00
Krish Dholakia
4ac66bd843
LiteLLM Minor Fixes and Improvements (09/07/2024) (#5580)
* fix(litellm_logging.py): set completion_start_time_float to end_time_float if none

Fixes https://github.com/BerriAI/litellm/issues/5500

* feat(_init_.py): add new 'openai_text_completion_compatible_providers' list

Fixes https://github.com/BerriAI/litellm/issues/5558

Handles correctly routing fireworks ai calls when done via text completions

* fix: fix linting errors

* fix: fix linting errors

* fix(openai.py): fix exception raised

* fix(openai.py): fix error handling

* fix(_redis.py): allow all supported arguments for redis cluster (#5554)

* Revert "fix(_redis.py): allow all supported arguments for redis cluster (#5554)" (#5583)

This reverts commit f2191ef4cb.

* fix(router.py): return model alias w/ underlying deployment on router.get_model_list()

Fixes https://github.com/BerriAI/litellm/issues/5524#issuecomment-2336410666

* test: handle flaky tests

---------

Co-authored-by: Jonas Dittrich <58814480+Kakadus@users.noreply.github.com>
2024-09-09 18:54:17 -07:00
Ishaan Jaff
b573847dad run ci cd again 2024-09-05 19:07:16 -07:00
Ishaan Jaff
f42a0528db
Merge branch 'main' into litellm_allow_internal_user_view_usage 2024-09-05 16:46:06 -07:00
Ishaan Jaff
30137b0b72 run ci - cd again 2024-09-05 15:42:40 -07:00
Ishaan Jaff
9eb35e47ee run ci/cd on main 2024-09-05 15:39:30 -07:00
Ishaan Jaff
05e02fbe76 run ci/cd again 2024-09-05 15:38:49 -07:00
Ishaan Jaff
004a71b8dc use requirements txt 2024-09-05 15:34:17 -07:00
Ishaan Jaff
42b2566906 run ci/cd agaiin 2024-09-05 15:26:12 -07:00
Ishaan Jaff
5b1d207cf4 run test again 2024-09-05 15:20:21 -07:00
Ishaan Jaff
89649282ee run again 2024-09-05 15:17:17 -07:00
Ishaan Jaff
4e392ebdb2
Merge pull request #5484 from BerriAI/litellm_fix_redis
fix proxy server - always read redis for rate limiting logic
2024-09-04 12:39:15 -07:00
Krrish Dholakia
e7855329e8 style: ci/cd run again 2024-09-04 08:32:35 -07:00
Ishaan Jaff
f5122e0325 ci/cd run again 2024-09-04 07:41:08 -07:00
Krish Dholakia
be3c7b401e
LiteLLM Minor fixes + improvements (08/03/2024) (#5488)
* fix(internal_user_endpoints.py): set budget_reset_at for /user/update

* fix(vertex_and_google_ai_studio_gemini.py): handle accumulated json

Fixes https://github.com/BerriAI/litellm/issues/5479

* fix(vertex_ai_and_gemini.py): fix assistant message function call when content is not None

Fixes https://github.com/BerriAI/litellm/issues/5490

* fix(proxy_server.py): generic state uuid for okta sso

* fix(lago.py): improve debug logs

Debugging for https://github.com/BerriAI/litellm/issues/5477

* docs(bedrock.md): add bedrock cross-region inferencing to docs

* fix(azure.py): return azure response headers on aembedding call

* feat(azure.py): return azure response headers for `/audio/transcription`

* fix(types/utils.py): standardize deepseek / anthropic prompt caching usage information

Closes https://github.com/BerriAI/litellm/issues/5285

* docs(usage.md): add docs on litellm usage object

* test(test_completion.py): mark flaky test
2024-09-03 21:21:34 -07:00
Ishaan Jaff
666dd6aca2
Merge pull request #5478 from BerriAI/litellm_Add_ai21
[Feat] Add AI21 /chat API
2024-09-02 16:20:37 -07:00
Krish Dholakia
f9e6507cd1
LiteLLM Minor Fixes + Improvements (#5474)
* feat(proxy/_types.py): add lago billing to callbacks ui

Closes https://github.com/BerriAI/litellm/issues/5472

* fix(anthropic.py): return anthropic prompt caching information

Fixes https://github.com/BerriAI/litellm/issues/5364

* feat(bedrock/chat.py): support 'json_schema' for bedrock models

Closes https://github.com/BerriAI/litellm/issues/5434

* fix(bedrock/embed/embeddings.py): support async embeddings for amazon titan models

* fix: linting fixes

* fix: handle key errors

* fix(bedrock/chat.py): fix bedrock ai21 streaming object

* feat(bedrock/embed): support bedrock embedding optional params

* fix(databricks.py): fix usage chunk

* fix(internal_user_endpoints.py): apply internal user defaults, if user role updated

Fixes issue where user update wouldn't apply defaults

* feat(slack_alerting.py): provide multiple slack channels for a given alert type

multiple channels might be interested in receiving an alert for a given type

* docs(alerting.md): add multiple channel alerting to docs
2024-09-02 14:29:57 -07:00
Ishaan Jaff
d4f77832c0 docs add docs on supported params 2024-09-02 12:39:07 -07:00
Ishaan Jaff
93650db380 fix linting 2024-09-02 12:15:14 -07:00
Ishaan Jaff
263e283126 add ai21 model test 2024-09-02 12:14:13 -07:00
Ishaan Jaff
ae25c5695f add streaming test for ai21 2024-09-02 11:58:15 -07:00
Ishaan Jaff
0874885f04 add all ai21 params 2024-09-02 11:54:40 -07:00
Ishaan Jaff
0970475a3b test ai21 2024-09-02 11:47:48 -07:00
Ishaan Jaff
69dea613ca ci/cd run again 2024-08-30 14:03:00 -07:00
Ishaan Jaff
f1147696a3 ci/cd run again 2024-08-27 22:59:21 -07:00
Ishaan Jaff
c8fc5c44e1 ci/cd run again 2024-08-27 22:53:18 -07:00
Ishaan Jaff
3966c545b4 fix flaky tests 2024-08-27 22:47:11 -07:00
Krrish Dholakia
640e5d0dc9 test: fix test 2024-08-27 22:44:29 -07:00
Ishaan Jaff
1b04380540 fix palm api is deactivated by google 2024-08-27 21:08:49 -07:00
Krish Dholakia
415abc86c6
Merge pull request #5358 from BerriAI/litellm_fix_retry_after
fix retry after - cooldown individual models based on their specific 'retry-after' header
2024-08-27 11:50:14 -07:00
Krrish Dholakia
18b67a455e test: fix test 2024-08-27 10:46:57 -07:00
Krrish Dholakia
b0f01e5b95 test: fix test 2024-08-27 07:14:14 -07:00
Krrish Dholakia
d13d2e8a62 feat(vertex_httpx.py): support functions param for gemini google ai studio + vertex ai
Closes https://github.com/BerriAI/litellm/issues/5344
2024-08-26 10:59:01 -07:00
Ishaan Jaff
735eb041b9 ci/cd run again 2024-08-26 08:36:58 -07:00
Krrish Dholakia
8c5ff150f6 fix(huggingface_restapi.py): fix tests 2024-08-23 21:40:27 -07:00
Ishaan Jaff
18cd905a1f ci/cd run again 2024-08-22 18:41:10 -07:00
Krrish Dholakia
11bfc1dca7 fix(cohere_chat.py): support passing 'extra_headers'
Fixes https://github.com/BerriAI/litellm/issues/4709
2024-08-22 10:17:36 -07:00
Ishaan Jaff
9dbd1b089b fix tg ai -deprecated model 2024-08-21 12:14:39 -07:00
Krrish Dholakia
8406016096 test(test_completion.py): fix gemini rate limit error 2024-08-21 07:17:27 -07:00
Krrish Dholakia
a6ee1e04fe test(test_completion.py): fix test 2024-08-21 07:01:02 -07:00
Krish Dholakia
ad1e327208
Merge pull request #5284 from BerriAI/litellm_fix_azure_api_version
feat(azure.py): support dynamic azure api versions
2024-08-20 11:46:55 -07:00