Ishaan Jaff
f5ab34e246
docs - deploy litellm on gcp cloud run
2024-05-06 08:10:35 -07:00
Ishaan Jaff
0b069a314f
docs - add using vertex embedding models
2024-05-06 07:56:17 -07:00
Ishaan Jaff
b4358cc007
docs - set retry policy
2024-05-04 17:52:01 -07:00
Ishaan Jaff
1a96bbea64
Merge pull request #1530 from TanaroSch/main
...
change max_tokens type to int
2024-05-04 12:47:15 -07:00
Rajan Paneru
22ac1976c2
Improve the document of Traceloop
...
If we follow exact step I was getting two errors, which this Pull-Request will fix:
* Fixed the yaml snippet spacing issue: it was throwing error when running copy+paste code
* Added api_key: my-fake-key as to run litellm --config config.yaml --debug it will be required, otherwise it will fail
Need for Improvements:
* The traceloop is ambigous, because it's a company that maintains OpenLLMetry, and has observability solution and the sdk name is aslo traceloop-sdk
* The doc was missing several other observability solutions
* The steps were not quite obvious, added one more steps to make things clear
2024-05-04 22:42:53 +09:30
Krrish Dholakia
001e7976da
docs(exception_mapping.md): update exception mapping docs with 'should_retry'
2024-05-03 16:51:26 -07:00
Ishaan Jaff
d281d74067
docs - titan embeddings v2
2024-05-03 15:59:02 -07:00
Ishaan Jaff
532866a92d
docs - simplify prod docs
2024-05-03 15:40:05 -07:00
Krish Dholakia
7e04447159
Merge pull request #3393 from Priva28/main
...
Add Llama3 tokenizer and allow custom tokenizers.
2024-05-02 16:32:41 -07:00
greenscale-nandesh
02f0c6bfc8
Merge branch 'main' into main
2024-05-02 08:22:11 -07:00
Christian Privitelli
f461f6b746
update docs, allow use with encode/decode
2024-05-02 14:56:38 +10:00
Krrish Dholakia
0b1382aef4
build(ui): enable adding openmeter via proxy ui
2024-05-01 21:16:23 -07:00
Krrish Dholakia
49a97cbf6b
docs(openmeter.md): add openmeter to docs
2024-05-01 18:31:45 -07:00
Krrish Dholakia
846e971f5e
docs(deploy.md): cleanup docker quick start docs
2024-05-01 10:00:49 -07:00
Krrish Dholakia
faa139ff4c
docs(routing.md): add docs on lowest latency routing buffer
2024-04-30 22:41:50 -07:00
Ishaan Jaff
63e5b8764e
docs - slack alerting
2024-04-29 21:33:03 -07:00
Ishaan Jaff
a13fb2244e
docs example logging to langfuse
2024-04-29 21:26:27 -07:00
Ishaan Jaff
bf01cc535b
docs logging to langfuse on proxy
2024-04-29 21:26:15 -07:00
Ishaan Jaff
d3d62387f0
docs - fix kub.yaml config on docs
2024-04-29 21:20:29 -07:00
Krrish Dholakia
701be55872
feat(utils.py): json logs for raw request sent by litellm
...
make it easier to view verbose logs in datadog
2024-04-29 19:21:19 -07:00
Krrish Dholakia
a048bcf88f
docs(langfuse_integration.md): add 'existing_trace_id' to langfuse docs
2024-04-29 16:40:38 -07:00
Krrish Dholakia
4f9e8d5aa1
docs(load_test.md): cleanup docs
2024-04-29 16:27:58 -07:00
Krrish Dholakia
963a59a786
docs(load_test.md): add multi-instance router load test to docs
2024-04-29 16:25:56 -07:00
Krrish Dholakia
e11a92c735
docs(load_test.md): simplify doc
2024-04-29 16:00:02 -07:00
Krrish Dholakia
0bae8911f8
docs(load_test.md): formatting
2024-04-29 15:58:41 -07:00
Krrish Dholakia
5b56a0856e
docs(load_test.md): load test multiple instances of the proxy w/ tpm/rpm limits on deployments
2024-04-29 15:58:14 -07:00
Krrish Dholakia
8ba4b73573
docs(routing.md): add max parallel requests to router docs
2024-04-29 15:37:48 -07:00
Ishaan Jaff
5dc911cba9
docs - track cost custom callbacks
2024-04-29 13:15:08 -07:00
Krrish Dholakia
27f76e263e
docs(vllm.md): update docs to tell people to check openai-compatible endpoint docs for vllm
2024-04-28 09:48:03 -07:00
Krish Dholakia
7ed64709aa
Merge branch 'main' into litellm_common_auth_params
2024-04-28 08:38:06 -07:00
Ishaan Jaff
37d2960c35
docs - alerting
2024-04-27 15:33:21 -07:00
Ishaan Jaff
48d590c428
Merge pull request #3330 from BerriAI/litellm_rdct_msgs
...
[Feat] Redact Logging Messages/Response content on Logging Providers with `litellm.turn_off_message_logging=True`
2024-04-27 11:25:09 -07:00
Ishaan Jaff
c3435ff8bb
sentry redact messages
2024-04-27 11:23:46 -07:00
Ishaan Jaff
99a2cb7ec3
docs - langfuse redact messages
2024-04-27 11:23:35 -07:00
Krrish Dholakia
ee68c0aaaa
feat(utils.py): unify common auth params across azure/vertex_ai/bedrock/watsonx
2024-04-27 11:06:18 -07:00
Simon S. Viloria
2c8e19c576
(docs) fixed typo in url for watsonx colab notebook
2024-04-27 16:43:49 +02:00
Krish Dholakia
b7beab2e39
Merge pull request #3270 from simonsanvil/feature/watsonx-integration
...
(feat) add IBM watsonx.ai as an llm provider
2024-04-27 05:48:34 -07:00
Krrish Dholakia
1c619e794b
docs(mistral.md): update to show tool calling example
2024-04-26 17:16:25 -07:00
Ishaan Jaff
d003946eb2
docs - setting up litellm-database
2024-04-26 10:23:15 -07:00
Ishaan Jaff
a0aafb19a0
docs - fix routing
2024-04-25 13:43:51 -07:00
Ishaan Jaff
d4d81dce01
Merge pull request #3289 from andeplane/patch-1
...
Fixed typo in token_usage.md
2024-04-25 09:08:53 -07:00
Krrish Dholakia
c05464d3ff
docs(replicate.md): add proxy examples to docs
2024-04-25 08:46:21 -07:00
Anders Hafreager
992ee6c139
Fixed typo in token_usage.md
2024-04-25 09:47:25 +02:00
Krrish Dholakia
1988ce3247
feat(router.py): support mock testing fallbacks flag
...
allow user to test if fallbacks work as expected with a `mock_testing_fallbacks = True` flag set during a call
2024-04-24 20:13:10 -07:00
Nandesh Guru
30d1fe7fe3
Merge branch 'main' of https://github.com/greenscale-ai/litellm
2024-04-24 13:46:46 -07:00
Nandesh Guru
3a5edcc321
Greenscale Documentation
2024-04-24 13:46:43 -07:00
Krish Dholakia
157099dd9e
Merge pull request #3098 from greenscale-ai/main
...
Support for Greenscale AI logging
2024-04-24 13:09:03 -07:00
Krrish Dholakia
840bb5a18d
fix(proxy_server.py): add new flag for disable sharing master key on ui
2024-04-24 10:06:01 -07:00
Simon Sanchez Viloria
1bb9754068
(docs) added watsonx.ai provider documentation
2024-04-24 17:22:17 +02:00
Krrish Dholakia
0011b4c09e
docs(anthropic.md): add parallel tool calling example for anthropic
2024-04-23 20:11:03 -07:00