Krish Dholakia
713d762411
LiteLLM Minor Fixes and Improvements (09/13/2024) ( #5689 )
...
* refactor: cleanup unused variables + fix pyright errors
* feat(health_check.py): Closes https://github.com/BerriAI/litellm/issues/5686
* fix(o1_reasoning.py): add stricter check for o-1 reasoning model
* refactor(mistral/): make it easier to see mistral transformation logic
* fix(openai.py): fix openai o-1 model param mapping
Fixes https://github.com/BerriAI/litellm/issues/5685
* feat(main.py): infer finetuned gemini model from base model
Fixes https://github.com/BerriAI/litellm/issues/5678
* docs(vertex.md): update docs to call finetuned gemini models
* feat(proxy_server.py): allow admin to hide proxy model aliases
Closes https://github.com/BerriAI/litellm/issues/5692
* docs(load_balancing.md): add docs on hiding alias models from proxy config
* fix(base.py): don't raise notimplemented error
* fix(user_api_key_auth.py): fix model max budget check
* fix(router.py): fix elif
* fix(user_api_key_auth.py): don't set team_id to empty str
* fix(team_endpoints.py): fix response type
* test(test_completion.py): handle predibase error
* test(test_proxy_server.py): fix test
* fix(o1_transformation.py): fix max_completion_token mapping
* test(test_image_generation.py): mark flaky test
2024-09-14 10:02:55 -07:00
Ishaan Jaff
d79321bd16
[Fix] Performance - use in memory cache when downloading images from a url ( #5657 )
...
* fix use in memory cache when getting images
* fix linting
* fix load testing
* fix load test size
* fix load test size
* trigger ci/cd again
2024-09-13 07:23:42 -07:00
Ishaan Jaff
5d660f239e
fix config.yml
2024-09-12 09:28:45 -07:00
Ishaan Jaff
1985b4a508
ci/cd run again
2024-09-12 08:42:34 -07:00
Ishaan Jaff
5469ebe53a
ci/cd run again
2024-09-12 08:31:17 -07:00
Ishaan Jaff
11099ff057
ci/cd run again
2024-09-10 19:21:13 -07:00
Krish Dholakia
09ca581620
LiteLLM Minor Fixes and Improvements (09/09/2024) ( #5602 )
...
* fix(main.py): pass default azure api version as alternative in completion call
Fixes api error caused due to api version
Closes https://github.com/BerriAI/litellm/issues/5584
* Fixed gemini-1.5-flash pricing (#5590 )
* add /key/list endpoint
* bump: version 1.44.21 → 1.44.22
* docs architecture
* Fixed gemini-1.5-flash pricing
---------
Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
* fix(bedrock/chat.py): fix converse api stop sequence param mapping
Fixes https://github.com/BerriAI/litellm/issues/5592
* fix(databricks/cost_calculator.py): handle databricks model name changes
Fixes https://github.com/BerriAI/litellm/issues/5597
* fix(azure.py): support azure api version 2024-08-01-preview
Closes https://github.com/BerriAI/litellm/issues/5377
* fix(proxy/_types.py): allow dev keys to call cohere /rerank endpoint
Fixes issue where only admin could call rerank endpoint
* fix(azure.py): check if model is gpt-4o
* fix(proxy/_types.py): support /v1/rerank on non-admin routes as well
* fix(cost_calculator.py): fix split on `/` logic in cost calculator
---------
Co-authored-by: F1bos <44951186+F1bos@users.noreply.github.com>
Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
2024-09-09 21:56:12 -07:00
Krish Dholakia
52849e6422
LiteLLM Minor Fixes and Improvements (09/07/2024) ( #5580 )
...
* fix(litellm_logging.py): set completion_start_time_float to end_time_float if none
Fixes https://github.com/BerriAI/litellm/issues/5500
* feat(_init_.py): add new 'openai_text_completion_compatible_providers' list
Fixes https://github.com/BerriAI/litellm/issues/5558
Handles correctly routing fireworks ai calls when done via text completions
* fix: fix linting errors
* fix: fix linting errors
* fix(openai.py): fix exception raised
* fix(openai.py): fix error handling
* fix(_redis.py): allow all supported arguments for redis cluster (#5554 )
* Revert "fix(_redis.py): allow all supported arguments for redis cluster (#5554 )" (#5583 )
This reverts commit f2191ef4cb
.
* fix(router.py): return model alias w/ underlying deployment on router.get_model_list()
Fixes https://github.com/BerriAI/litellm/issues/5524#issuecomment-2336410666
* test: handle flaky tests
---------
Co-authored-by: Jonas Dittrich <58814480+Kakadus@users.noreply.github.com>
2024-09-09 18:54:17 -07:00
Ishaan Jaff
3cbb85dfda
run ci cd again
2024-09-05 19:07:16 -07:00
Ishaan Jaff
dd7d93fd54
Merge branch 'main' into litellm_allow_internal_user_view_usage
2024-09-05 16:46:06 -07:00
Ishaan Jaff
a1736e900c
run ci - cd again
2024-09-05 15:42:40 -07:00
Ishaan Jaff
99130a02d9
run ci/cd on main
2024-09-05 15:39:30 -07:00
Ishaan Jaff
556e2eea3a
run ci/cd again
2024-09-05 15:38:49 -07:00
Ishaan Jaff
66b3585c14
use requirements txt
2024-09-05 15:34:17 -07:00
Ishaan Jaff
cd265d43a8
run ci/cd agaiin
2024-09-05 15:26:12 -07:00
Ishaan Jaff
6cb7d035a4
run test again
2024-09-05 15:20:21 -07:00
Ishaan Jaff
edac75b824
run again
2024-09-05 15:17:17 -07:00
Ishaan Jaff
935dba4470
Merge pull request #5484 from BerriAI/litellm_fix_redis
...
fix proxy server - always read redis for rate limiting logic
2024-09-04 12:39:15 -07:00
Krrish Dholakia
2eca308069
style: ci/cd run again
2024-09-04 08:32:35 -07:00
Ishaan Jaff
0fbf8f9b74
ci/cd run again
2024-09-04 07:41:08 -07:00
Krish Dholakia
8eb7cb5300
LiteLLM Minor fixes + improvements (08/03/2024) ( #5488 )
...
* fix(internal_user_endpoints.py): set budget_reset_at for /user/update
* fix(vertex_and_google_ai_studio_gemini.py): handle accumulated json
Fixes https://github.com/BerriAI/litellm/issues/5479
* fix(vertex_ai_and_gemini.py): fix assistant message function call when content is not None
Fixes https://github.com/BerriAI/litellm/issues/5490
* fix(proxy_server.py): generic state uuid for okta sso
* fix(lago.py): improve debug logs
Debugging for https://github.com/BerriAI/litellm/issues/5477
* docs(bedrock.md): add bedrock cross-region inferencing to docs
* fix(azure.py): return azure response headers on aembedding call
* feat(azure.py): return azure response headers for `/audio/transcription`
* fix(types/utils.py): standardize deepseek / anthropic prompt caching usage information
Closes https://github.com/BerriAI/litellm/issues/5285
* docs(usage.md): add docs on litellm usage object
* test(test_completion.py): mark flaky test
2024-09-03 21:21:34 -07:00
Ishaan Jaff
dc1b0ec182
Merge pull request #5478 from BerriAI/litellm_Add_ai21
...
[Feat] Add AI21 /chat API
2024-09-02 16:20:37 -07:00
Krish Dholakia
11f85d883f
LiteLLM Minor Fixes + Improvements ( #5474 )
...
* feat(proxy/_types.py): add lago billing to callbacks ui
Closes https://github.com/BerriAI/litellm/issues/5472
* fix(anthropic.py): return anthropic prompt caching information
Fixes https://github.com/BerriAI/litellm/issues/5364
* feat(bedrock/chat.py): support 'json_schema' for bedrock models
Closes https://github.com/BerriAI/litellm/issues/5434
* fix(bedrock/embed/embeddings.py): support async embeddings for amazon titan models
* fix: linting fixes
* fix: handle key errors
* fix(bedrock/chat.py): fix bedrock ai21 streaming object
* feat(bedrock/embed): support bedrock embedding optional params
* fix(databricks.py): fix usage chunk
* fix(internal_user_endpoints.py): apply internal user defaults, if user role updated
Fixes issue where user update wouldn't apply defaults
* feat(slack_alerting.py): provide multiple slack channels for a given alert type
multiple channels might be interested in receiving an alert for a given type
* docs(alerting.md): add multiple channel alerting to docs
2024-09-02 14:29:57 -07:00
Ishaan Jaff
d86abb4abe
docs add docs on supported params
2024-09-02 12:39:07 -07:00
Ishaan Jaff
93d556dbeb
fix linting
2024-09-02 12:15:14 -07:00
Ishaan Jaff
b1f3f27342
add ai21 model test
2024-09-02 12:14:13 -07:00
Ishaan Jaff
02c415a123
add streaming test for ai21
2024-09-02 11:58:15 -07:00
Ishaan Jaff
e1dacde1ec
add all ai21 params
2024-09-02 11:54:40 -07:00
Ishaan Jaff
59a32e56ce
test ai21
2024-09-02 11:47:48 -07:00
Ishaan Jaff
91c4338f01
ci/cd run again
2024-08-30 14:03:00 -07:00
Ishaan Jaff
d77d139e95
ci/cd run again
2024-08-27 22:59:21 -07:00
Ishaan Jaff
49f537a2ea
ci/cd run again
2024-08-27 22:53:18 -07:00
Ishaan Jaff
64398fe123
fix flaky tests
2024-08-27 22:47:11 -07:00
Krrish Dholakia
d8ecf55eca
test: fix test
2024-08-27 22:44:29 -07:00
Ishaan Jaff
54638adbe0
fix palm api is deactivated by google
2024-08-27 21:08:49 -07:00
Krish Dholakia
34eb1206c6
Merge pull request #5358 from BerriAI/litellm_fix_retry_after
...
fix retry after - cooldown individual models based on their specific 'retry-after' header
2024-08-27 11:50:14 -07:00
Krrish Dholakia
30bf8a4c1d
test: fix test
2024-08-27 10:46:57 -07:00
Krrish Dholakia
87644691be
test: fix test
2024-08-27 07:14:14 -07:00
Krrish Dholakia
aedc6652d4
feat(vertex_httpx.py): support functions param for gemini google ai studio + vertex ai
...
Closes https://github.com/BerriAI/litellm/issues/5344
2024-08-26 10:59:01 -07:00
Ishaan Jaff
150c226663
ci/cd run again
2024-08-26 08:36:58 -07:00
Krrish Dholakia
5a2e54c44d
fix(huggingface_restapi.py): fix tests
2024-08-23 21:40:27 -07:00
Ishaan Jaff
56402dca02
ci/cd run again
2024-08-22 18:41:10 -07:00
Krrish Dholakia
8f306f8e41
fix(cohere_chat.py): support passing 'extra_headers'
...
Fixes https://github.com/BerriAI/litellm/issues/4709
2024-08-22 10:17:36 -07:00
Ishaan Jaff
a39eff07df
fix tg ai -deprecated model
2024-08-21 12:14:39 -07:00
Krrish Dholakia
468768f19d
test(test_completion.py): fix gemini rate limit error
2024-08-21 07:17:27 -07:00
Krrish Dholakia
8487f1cbe4
test(test_completion.py): fix test
2024-08-21 07:01:02 -07:00
Krish Dholakia
b1e9017841
Merge pull request #5284 from BerriAI/litellm_fix_azure_api_version
...
feat(azure.py): support dynamic azure api versions
2024-08-20 11:46:55 -07:00
Krish Dholakia
969b724615
Merge pull request #5296 from BerriAI/litellm_azure_json_schema_support
...
feat(azure.py): support 'json_schema' for older models
2024-08-20 11:41:38 -07:00
Krish Dholakia
85c5d4f8c2
Merge branch 'main' into litellm_fix_azure_api_version
2024-08-20 11:40:53 -07:00
Ishaan Jaff
ca4b837062
ci/cd run again
2024-08-19 21:00:38 -07:00