Krish Dholakia
f2c0a31e3c
LiteLLM Minor Fixes & Improvements (10/05/2024) ( #6083 )
...
* docs(prompt_caching.md): add prompt caching cost calc example to docs
* docs(prompt_caching.md): add proxy examples to docs
* feat(utils.py): expose new helper `supports_prompt_caching()` to check if a model supports prompt caching
* docs(prompt_caching.md): add docs on checking model support for prompt caching
* build: fix invalid json
2024-10-05 18:59:11 -04:00
Krish Dholakia
2e5c46ef6d
LiteLLM Minor Fixes & Improvements (10/04/2024) ( #6064 )
...
* fix(litellm_logging.py): ensure cache hits are scrubbed if 'turn_off_message_logging' is enabled
* fix(sagemaker.py): fix streaming to raise error immediately
Fixes https://github.com/BerriAI/litellm/issues/6054
* (fixes) gcs bucket key based logging (#6044 )
* fixes for gcs bucket logging
* fix StandardCallbackDynamicParams
* fix - gcs logging when payload is not serializable
* add test_add_callback_via_key_litellm_pre_call_utils_gcs_bucket
* working success callbacks
* linting fixes
* fix linting error
* add type hints to functions
* fixes for dynamic success and failure logging
* fix for test_async_chat_openai_stream
* fix handle case when key based logging vars are set as os.environ/ vars
* fix prometheus track cooldown events on custom logger (#6060 )
* (docs) add 1k rps load test doc (#6059 )
* docs 1k rps load test
* docs load testing
* docs load testing litellm
* docs load testing
* clean up load test doc
* docs prom metrics for load testing
* docs using prometheus on load testing
* doc load testing with prometheus
* (fixes) docs + qa - gcs key based logging (#6061 )
* fixes for required values for gcs bucket
* docs gcs bucket logging
* bump: version 1.48.12 → 1.48.13
* ci/cd run again
* bump: version 1.48.13 → 1.48.14
* update load test doc
* (docs) router settings - on litellm config (#6037 )
* add yaml with all router settings
* add docs for router settings
* docs router settings litellm settings
* (feat) OpenAI prompt caching models to model cost map (#6063 )
* add prompt caching for latest models
* add cache_read_input_token_cost for prompt caching models
* fix(litellm_logging.py): check if param is iterable
Fixes https://github.com/BerriAI/litellm/issues/6025#issuecomment-2393929946
* fix(factory.py): support passing an 'assistant_continue_message' to prevent bedrock error
Fixes https://github.com/BerriAI/litellm/issues/6053
* fix(databricks/chat): handle streaming responses
* fix(factory.py): fix linting error
* fix(utils.py): unify anthropic + deepseek prompt caching information to openai format
Fixes https://github.com/BerriAI/litellm/issues/6069
* test: fix test
* fix(types/utils.py): support all openai roles
Fixes https://github.com/BerriAI/litellm/issues/6052
* test: fix test
---------
Co-authored-by: Ishaan Jaff <ishaanjaffer0324@gmail.com>
2024-10-04 21:28:53 -04:00
Ishaan Jaff
1973ae8fb8
[Feat] Allow setting supports_vision
for Custom OpenAI endpoints + Added testing ( #5821 )
...
* add test for using images with custom openai endpoints
* run all otel tests
* update name of test
* add custom openai model to test config
* add test for setting supports_vision=True for model
* fix test guardrails aporia
* docs supports vison
* fix yaml
* fix yaml
* docs supports vision
* fix bedrock guardrail test
* fix cohere rerank test
* update model_group doc string
* add better prints on test
2024-09-21 11:35:55 -07:00
Krish Dholakia
d46660ea0f
LiteLLM Minor Fixes & Improvements (09/18/2024) ( #5772 )
...
* fix(proxy_server.py): fix azure key vault logic to not require client id/secret
* feat(cost_calculator.py): support fireworks ai cost tracking
* build(docker-compose.yml): add lines for mounting config.yaml to docker compose
Closes https://github.com/BerriAI/litellm/issues/5739
* fix(input.md): update docs to clarify litellm supports content as a list of dictionaries
Fixes https://github.com/BerriAI/litellm/issues/5755
* fix(input.md): update input.md to include all message values
* fix(image_handling.py): follow image url redirects
Fixes https://github.com/BerriAI/litellm/issues/5763
* fix(router.py): Fix model key/base leak in error message
Fixes https://github.com/BerriAI/litellm/issues/5762
* fix(http_handler.py): fix linting error
* fix(azure.py): fix logging to show azure_ad_token being used
Fixes https://github.com/BerriAI/litellm/issues/5767
* fix(_redis.py): add redis sentinel support
Closes https://github.com/BerriAI/litellm/issues/4381
* feat(_redis.py): add redis sentinel support
Closes https://github.com/BerriAI/litellm/issues/4381
* test(test_completion_cost.py): fix test
* Databricks Integration: Integrate Databricks SDK as optional mechanism for fetching API base and token, if unspecified (#5746 )
* LiteLLM Minor Fixes & Improvements (09/16/2024) (#5723 )
* coverage (#5713 )
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* Move (#5714 )
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix(litellm_logging.py): fix logging client re-init (#5710 )
Fixes https://github.com/BerriAI/litellm/issues/5695
* fix(presidio.py): Fix logging_hook response and add support for additional presidio variables in guardrails config
Fixes https://github.com/BerriAI/litellm/issues/5682
* feat(o1_handler.py): fake streaming for openai o1 models
Fixes https://github.com/BerriAI/litellm/issues/5694
* docs: deprecated traceloop integration in favor of native otel (#5249 )
* fix: fix linting errors
* fix: fix linting errors
* fix(main.py): fix o1 import
---------
Signed-off-by: dbczumar <corey.zumar@databricks.com>
Co-authored-by: Corey Zumar <39497902+dbczumar@users.noreply.github.com>
Co-authored-by: Nir Gazit <nirga@users.noreply.github.com>
* feat(spend_management_endpoints.py): expose `/global/spend/refresh` endpoint for updating material view (#5730 )
* feat(spend_management_endpoints.py): expose `/global/spend/refresh` endpoint for updating material view
Supports having `MonthlyGlobalSpend` view be a material view, and exposes an endpoint to refresh it
* fix(custom_logger.py): reset calltype
* fix: fix linting errors
* fix: fix linting error
* fix
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix: fix import
* Fix
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* DB test
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* Coverage
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* progress
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix
Signed-off-by: dbczumar <corey.zumar@databricks.com>
* fix test name
Signed-off-by: dbczumar <corey.zumar@databricks.com>
---------
Signed-off-by: dbczumar <corey.zumar@databricks.com>
Co-authored-by: Krish Dholakia <krrishdholakia@gmail.com>
Co-authored-by: Nir Gazit <nirga@users.noreply.github.com>
* test: fix test
* test(test_databricks.py): fix test
* fix(databricks/chat.py): handle custom endpoint (e.g. sagemaker)
* Apply code scanning fix for clear-text logging of sensitive information
Co-authored-by: Copilot Autofix powered by AI <62310815+github-advanced-security[bot]@users.noreply.github.com>
* fix(__init__.py): fix known fireworks ai models
---------
Signed-off-by: dbczumar <corey.zumar@databricks.com>
Co-authored-by: Corey Zumar <39497902+dbczumar@users.noreply.github.com>
Co-authored-by: Nir Gazit <nirga@users.noreply.github.com>
Co-authored-by: Copilot Autofix powered by AI <62310815+github-advanced-security[bot]@users.noreply.github.com>
2024-09-19 13:25:29 -07:00
Ishaan Jaff
c220fc0e92
docs max_completion_tokens
2024-09-14 19:12:12 -07:00
Krish Dholakia
be3c7b401e
LiteLLM Minor fixes + improvements (08/03/2024) ( #5488 )
...
* fix(internal_user_endpoints.py): set budget_reset_at for /user/update
* fix(vertex_and_google_ai_studio_gemini.py): handle accumulated json
Fixes https://github.com/BerriAI/litellm/issues/5479
* fix(vertex_ai_and_gemini.py): fix assistant message function call when content is not None
Fixes https://github.com/BerriAI/litellm/issues/5490
* fix(proxy_server.py): generic state uuid for okta sso
* fix(lago.py): improve debug logs
Debugging for https://github.com/BerriAI/litellm/issues/5477
* docs(bedrock.md): add bedrock cross-region inferencing to docs
* fix(azure.py): return azure response headers on aembedding call
* feat(azure.py): return azure response headers for `/audio/transcription`
* fix(types/utils.py): standardize deepseek / anthropic prompt caching usage information
Closes https://github.com/BerriAI/litellm/issues/5285
* docs(usage.md): add docs on litellm usage object
* test(test_completion.py): mark flaky test
2024-09-03 21:21:34 -07:00
Krrish Dholakia
9df0588c2c
docs(json_mode.md): update docs
2024-09-02 22:41:17 -07:00
Krrish Dholakia
0c26b36d9d
docs(input.md): update docs on together ai response_format params support
2024-08-23 21:34:18 -07:00
Beltrán Aceves
5e583e0bf2
Fixed code snippet import typo in Structured Output docs
2024-08-20 23:01:22 +02:00
Krrish Dholakia
b8e4ef0abf
docs(json_mode.md): add azure openai models to doc
2024-08-19 07:19:23 -07:00
Zbigniew Łukasiak
963c921c5a
Mismatch in example fixed
2024-08-14 15:07:10 +02:00
Krrish Dholakia
fdd9a07051
fix(utils.py): Break out of infinite streaming loop
...
Fixes https://github.com/BerriAI/litellm/issues/5158
2024-08-12 14:00:43 -07:00
Krrish Dholakia
0ea056971c
docs(prefix.md): add prefix support to docs
2024-08-10 13:55:47 -07:00
Krrish Dholakia
2dd27a4e12
feat(utils.py): support validating json schema client-side if user opts in
2024-08-06 19:35:33 -07:00
Krrish Dholakia
0c88cc4153
docs(json_mode.md): add example of calling openai with pydantic model via litellm
2024-08-06 18:27:06 -07:00
Krrish Dholakia
f3a0eb8eb9
docs(json_mode.md): update json mode docs to show structured output responses
...
Relevant issue - https://github.com/BerriAI/litellm/issues/5074
2024-08-06 17:01:41 -07:00
Krrish Dholakia
696e75d69c
docs: add github provider to docs
2024-08-03 09:20:23 -07:00
Krrish Dholakia
f1b7d2318c
docs(input.md): update docs to show ollama tool calling
2024-07-30 09:56:24 -07:00
Krrish Dholakia
6d741a5424
docs(json_mode.md): add json mode to docs
2024-07-18 17:20:19 -07:00
Krrish Dholakia
ba334ff8b9
refactor(provider_specific_params.md): create separate doc for provider-specific param
...
Make it easier for people to know, how litellm handles provider-specific params.
2024-07-09 12:23:46 -07:00
berkecanrizai
40940cd606
fix: typo in vision docs
2024-07-05 13:31:12 +03:00
Krrish Dholakia
be8a6377f6
docs(input.md): add vertex ai json mode to mapped input params
2024-06-29 11:51:52 -07:00
Krrish Dholakia
a6aee18012
docs(token_usage.md): add response cost to usage docs
2024-06-26 18:05:47 -07:00
Krrish Dholakia
09f4eb7617
docs(reliable_completions.md): improve headers for easier searching
2024-06-26 08:09:31 -07:00
Krrish Dholakia
d6ed8c10b2
docs(function_call.md): cleanup
2024-06-25 18:26:34 -07:00
Krrish Dholakia
e96326a211
docs(input.md): update docs with parallel_tool_calls
2024-06-20 21:01:49 -07:00
Krrish Dholakia
3feaf231ac
docs(drop_params.md): drop unsupported params
2024-06-20 17:43:07 -07:00
Krrish Dholakia
f86dcbb109
docs(input.md): clarify meaning of 'drop_params'
2024-06-19 10:04:58 -07:00
Ishaan Jaff
69a20c94fd
docs - fix doc build time errors
2024-06-15 14:58:02 -07:00
Krrish Dholakia
162f9400d2
feat(utils.py): support dynamically setting 'drop_params'
...
Allows user to turn this on/off for individual calls by passing in as a completion arg
2024-06-05 08:44:04 -07:00
Krrish Dholakia
3db30ecb4c
docs(batching.md): add batch completion fastest response on proxy to docs
2024-05-28 22:14:22 -07:00
Krrish Dholakia
2ee599b848
docs(batching.md): add batch completion to docs
2024-05-28 22:08:06 -07:00
Krrish Dholakia
9698fc77fd
docs(input.md): add clarifai supported input params to docs
2024-05-24 08:57:50 -07:00
Krrish Dholakia
65c4d6be39
docs(databricks.md): add databricks api support to docs
2024-05-23 19:22:09 -07:00
Krrish Dholakia
af1d209f8f
docs(input.md): add anthropic tool choice support to docs
2024-05-21 17:56:21 -07:00
Krrish Dholakia
7fa203c810
docs(input.md): add mistral to input param docs
2024-05-13 13:50:49 -07:00
Ishaan Jaff
62276fc221
docs link to litellm batch completions
2024-05-11 13:45:32 -07:00
Ishaan Jaff
e7e54772ae
docs include stream_options
param
2024-05-08 21:57:25 -07:00
Ishaan Jaff
d968dedd77
Merge pull request #1530 from TanaroSch/main
...
change max_tokens type to int
2024-05-04 12:47:15 -07:00
Christian Privitelli
3449a5e446
update docs, allow use with encode/decode
2024-05-02 14:56:38 +10:00
Ishaan Jaff
cf4a38d361
docs - viewing if model supports vision
2024-04-15 18:15:15 -07:00
lucca
be265fbb15
initial
2024-04-04 16:58:51 -03:00
Krrish Dholakia
d8a6b8216d
docs(input.md): add docs on 'get_supported_openai_params'
2024-03-08 23:54:13 -08:00
ishaan-jaff
76e7f8831f
(docs) check if supports function calling
2024-02-28 17:41:54 -08:00
ishaan-jaff
c9567c8652
(docs) passing extra_headers to OpenAI
2024-02-23 09:27:48 -08:00
TanaroSch
b69cb2c773
change max_tokens float to int
2024-02-06 11:19:28 +01:00
Krrish Dholakia
1a1b929a4e
docs(token_usage.md): add LITELLM_LOCAL_MODEL_COST_MAP to docs
2024-02-01 18:05:13 -08:00
ishaan-jaff
248e5f3d92
(chore) remove deprecated completion_with_config() tests
2024-01-09 17:13:06 +05:30
ishaan-jaff
112ffa3596
(docs) add logprobs, top_logprobs
2023-12-26 14:53:30 +05:30
ishaan-jaff
f3844b309f
(docs) input params - add logprobs, top_logprobs
2023-12-26 14:41:33 +05:30