Commit graph

1690 commits

Author SHA1 Message Date
Ishaan Jaff
f5ab34e246 docs - deploy litellm on gcp cloud run 2024-05-06 08:10:35 -07:00
Ishaan Jaff
0b069a314f docs - add using vertex embedding models 2024-05-06 07:56:17 -07:00
Ishaan Jaff
b4358cc007 docs - set retry policy 2024-05-04 17:52:01 -07:00
Ishaan Jaff
1a96bbea64 Merge pull request #1530 from TanaroSch/main
change max_tokens type to int
2024-05-04 12:47:15 -07:00
Rajan Paneru
22ac1976c2 Improve the document of Traceloop
If we follow exact step I was getting two errors, which this Pull-Request will fix:
* Fixed the yaml snippet spacing issue: it was throwing error when running copy+paste code
* Added api_key: my-fake-key as to run litellm --config config.yaml --debug it will be required, otherwise it will fail

Need for Improvements:
* The traceloop is ambigous, because it's a company that maintains OpenLLMetry, and has observability solution and the sdk name is aslo traceloop-sdk
* The doc was missing several other observability solutions
* The steps were not quite obvious, added one more steps to make things clear
2024-05-04 22:42:53 +09:30
Krrish Dholakia
001e7976da docs(exception_mapping.md): update exception mapping docs with 'should_retry' 2024-05-03 16:51:26 -07:00
Ishaan Jaff
d281d74067 docs - titan embeddings v2 2024-05-03 15:59:02 -07:00
Ishaan Jaff
532866a92d docs - simplify prod docs 2024-05-03 15:40:05 -07:00
Krish Dholakia
7e04447159 Merge pull request #3393 from Priva28/main
Add Llama3 tokenizer and allow custom tokenizers.
2024-05-02 16:32:41 -07:00
greenscale-nandesh
02f0c6bfc8 Merge branch 'main' into main 2024-05-02 08:22:11 -07:00
Christian Privitelli
f461f6b746 update docs, allow use with encode/decode 2024-05-02 14:56:38 +10:00
Krrish Dholakia
0b1382aef4 build(ui): enable adding openmeter via proxy ui 2024-05-01 21:16:23 -07:00
Krrish Dholakia
49a97cbf6b docs(openmeter.md): add openmeter to docs 2024-05-01 18:31:45 -07:00
Krrish Dholakia
846e971f5e docs(deploy.md): cleanup docker quick start docs 2024-05-01 10:00:49 -07:00
Krrish Dholakia
faa139ff4c docs(routing.md): add docs on lowest latency routing buffer 2024-04-30 22:41:50 -07:00
Ishaan Jaff
63e5b8764e docs - slack alerting 2024-04-29 21:33:03 -07:00
Ishaan Jaff
a13fb2244e docs example logging to langfuse 2024-04-29 21:26:27 -07:00
Ishaan Jaff
bf01cc535b docs logging to langfuse on proxy 2024-04-29 21:26:15 -07:00
Ishaan Jaff
d3d62387f0 docs - fix kub.yaml config on docs 2024-04-29 21:20:29 -07:00
Krrish Dholakia
701be55872 feat(utils.py): json logs for raw request sent by litellm
make it easier to view verbose logs in datadog
2024-04-29 19:21:19 -07:00
Krrish Dholakia
a048bcf88f docs(langfuse_integration.md): add 'existing_trace_id' to langfuse docs 2024-04-29 16:40:38 -07:00
Krrish Dholakia
4f9e8d5aa1 docs(load_test.md): cleanup docs 2024-04-29 16:27:58 -07:00
Krrish Dholakia
963a59a786 docs(load_test.md): add multi-instance router load test to docs 2024-04-29 16:25:56 -07:00
Krrish Dholakia
e11a92c735 docs(load_test.md): simplify doc 2024-04-29 16:00:02 -07:00
Krrish Dholakia
0bae8911f8 docs(load_test.md): formatting 2024-04-29 15:58:41 -07:00
Krrish Dholakia
5b56a0856e docs(load_test.md): load test multiple instances of the proxy w/ tpm/rpm limits on deployments 2024-04-29 15:58:14 -07:00
Krrish Dholakia
8ba4b73573 docs(routing.md): add max parallel requests to router docs 2024-04-29 15:37:48 -07:00
Ishaan Jaff
5dc911cba9 docs - track cost custom callbacks 2024-04-29 13:15:08 -07:00
Krrish Dholakia
27f76e263e docs(vllm.md): update docs to tell people to check openai-compatible endpoint docs for vllm 2024-04-28 09:48:03 -07:00
Krish Dholakia
7ed64709aa Merge branch 'main' into litellm_common_auth_params 2024-04-28 08:38:06 -07:00
Ishaan Jaff
37d2960c35 docs - alerting 2024-04-27 15:33:21 -07:00
Ishaan Jaff
48d590c428 Merge pull request #3330 from BerriAI/litellm_rdct_msgs
[Feat] Redact Logging Messages/Response content on Logging Providers with `litellm.turn_off_message_logging=True`
2024-04-27 11:25:09 -07:00
Ishaan Jaff
c3435ff8bb sentry redact messages 2024-04-27 11:23:46 -07:00
Ishaan Jaff
99a2cb7ec3 docs - langfuse redact messages 2024-04-27 11:23:35 -07:00
Krrish Dholakia
ee68c0aaaa feat(utils.py): unify common auth params across azure/vertex_ai/bedrock/watsonx 2024-04-27 11:06:18 -07:00
Simon S. Viloria
2c8e19c576 (docs) fixed typo in url for watsonx colab notebook 2024-04-27 16:43:49 +02:00
Krish Dholakia
b7beab2e39 Merge pull request #3270 from simonsanvil/feature/watsonx-integration
(feat) add IBM watsonx.ai as an llm provider
2024-04-27 05:48:34 -07:00
Krrish Dholakia
1c619e794b docs(mistral.md): update to show tool calling example 2024-04-26 17:16:25 -07:00
Ishaan Jaff
d003946eb2 docs - setting up litellm-database 2024-04-26 10:23:15 -07:00
Ishaan Jaff
a0aafb19a0 docs - fix routing 2024-04-25 13:43:51 -07:00
Ishaan Jaff
d4d81dce01 Merge pull request #3289 from andeplane/patch-1
Fixed typo in token_usage.md
2024-04-25 09:08:53 -07:00
Krrish Dholakia
c05464d3ff docs(replicate.md): add proxy examples to docs 2024-04-25 08:46:21 -07:00
Anders Hafreager
992ee6c139 Fixed typo in token_usage.md 2024-04-25 09:47:25 +02:00
Krrish Dholakia
1988ce3247 feat(router.py): support mock testing fallbacks flag
allow user to test if fallbacks work as expected with a `mock_testing_fallbacks = True` flag set during a call
2024-04-24 20:13:10 -07:00
Nandesh Guru
30d1fe7fe3 Merge branch 'main' of https://github.com/greenscale-ai/litellm 2024-04-24 13:46:46 -07:00
Nandesh Guru
3a5edcc321 Greenscale Documentation 2024-04-24 13:46:43 -07:00
Krish Dholakia
157099dd9e Merge pull request #3098 from greenscale-ai/main
Support for Greenscale AI logging
2024-04-24 13:09:03 -07:00
Krrish Dholakia
840bb5a18d fix(proxy_server.py): add new flag for disable sharing master key on ui 2024-04-24 10:06:01 -07:00
Simon Sanchez Viloria
1bb9754068 (docs) added watsonx.ai provider documentation 2024-04-24 17:22:17 +02:00
Krrish Dholakia
0011b4c09e docs(anthropic.md): add parallel tool calling example for anthropic 2024-04-23 20:11:03 -07:00