Commit graph

1705 commits

Author SHA1 Message Date
Ishaan Jaff
223e386737
Merge pull request #3511 from BerriAI/litellm_router_send_exceptions_slack
[Feat] litellm.Router / litellm.completion -  send llm exceptions to slack
2024-05-07 19:09:44 -07:00
Ishaan Jaff
d46544d2bc docs setup alerting on router 2024-05-07 18:26:45 -07:00
Ishaan Jaff
17787db973
Merge pull request #3503 from paul-gauthier/deepseek
Added "deepseek/" as a supported provider (openai compatible)
2024-05-07 15:15:47 -07:00
Ishaan Jaff
d5f93048cc docs - lowest cost routing 2024-05-07 13:15:30 -07:00
Ishaan Jaff
4c909194c7 docs - lowest - latency routing 2024-05-07 12:43:44 -07:00
Paul Gauthier
90eb0ea022 Added support for the deepseek api 2024-05-07 11:44:03 -07:00
Krish Dholakia
93e5fb49d3
Merge pull request #3500 from ghaemisr/main
Added support for JWT auth with PEM cert public keys
2024-05-07 11:07:30 -07:00
Ishaan Jaff
ee1b1fe4f8
Merge pull request #3464 from sepiatone/main
update langchain documentation to reflect refactor
2024-05-07 08:42:23 -07:00
Ishaan Jaff
8b15507add
Merge pull request #3471 from paneru-rajan/Issue-3466-update-doc
Synced the doc with the Mistral
2024-05-07 08:41:18 -07:00
Sara Ghaemi
66b2b5fab9 made audience optional and updated docs 2024-05-07 11:37:04 -04:00
sepiatone
1518a7b42d update langchain documentation to reflect refactor 2024-05-07 17:01:55 +05:30
Krish Dholakia
30003afbf8
Merge pull request #3459 from alexanderepstein/langfuse_improvements
Update support for langfuse metadata
2024-05-06 21:56:29 -07:00
Ishaan Jaff
ec63a30095 docs - deploy litellm on gcp cloud run 2024-05-06 08:10:35 -07:00
Ishaan Jaff
e0001a9121 docs - add using vertex embedding models 2024-05-06 07:56:17 -07:00
Rajan Paneru
2347cb2247 bring back the to be depricated mistral-medium 2024-05-07 00:19:19 +09:30
Rajan Paneru
d0d063e25a Synced the model with the Mistral
* Ordered them based on previous order
* Excluded to be depricated models
* Use the model name, instead of the slug in the first column as referred by the name of the column
2024-05-06 13:16:24 +09:30
Alex Epstein
b82162832a Update support for langfuse metadata
- Added ability to set trace release, version, metadata
- Added ability to update fields during a trace continuation
- Added ability to update input and output during a trace continuation
- Wrote new test for verifying metadata is set correctly
- Small improvement to setting secret boolean, prevent unnecessary literal_eval
- Small improvements to langfuse tests
2024-05-05 14:56:36 -04:00
Ishaan Jaff
bbf5d79069 docs - set retry policy 2024-05-04 17:52:01 -07:00
Ishaan Jaff
d968dedd77
Merge pull request #1530 from TanaroSch/main
change max_tokens type to int
2024-05-04 12:47:15 -07:00
Rajan Paneru
7d9377f18a Improve the document of Traceloop
If we follow exact step I was getting two errors, which this Pull-Request will fix:
* Fixed the yaml snippet spacing issue: it was throwing error when running copy+paste code
* Added api_key: my-fake-key as to run litellm --config config.yaml --debug it will be required, otherwise it will fail

Need for Improvements:
* The traceloop is ambigous, because it's a company that maintains OpenLLMetry, and has observability solution and the sdk name is aslo traceloop-sdk
* The doc was missing several other observability solutions
* The steps were not quite obvious, added one more steps to make things clear
2024-05-04 22:42:53 +09:30
Krrish Dholakia
f7eee60943 docs(exception_mapping.md): update exception mapping docs with 'should_retry' 2024-05-03 16:51:26 -07:00
Ishaan Jaff
bf048ecda4 docs - titan embeddings v2 2024-05-03 15:59:02 -07:00
Ishaan Jaff
209baaca02 docs - simplify prod docs 2024-05-03 15:40:05 -07:00
Krish Dholakia
2200900ca2
Merge pull request #3393 from Priva28/main
Add Llama3 tokenizer and allow custom tokenizers.
2024-05-02 16:32:41 -07:00
greenscale-nandesh
6009632705
Merge branch 'main' into main 2024-05-02 08:22:11 -07:00
Christian Privitelli
3449a5e446 update docs, allow use with encode/decode 2024-05-02 14:56:38 +10:00
Krrish Dholakia
cdd3e1eef3 build(ui): enable adding openmeter via proxy ui 2024-05-01 21:16:23 -07:00
Krrish Dholakia
61d680143f docs(openmeter.md): add openmeter to docs 2024-05-01 18:31:45 -07:00
Krrish Dholakia
b187deb787 docs(deploy.md): cleanup docker quick start docs 2024-05-01 10:00:49 -07:00
Krrish Dholakia
6a2ddc2791 docs(routing.md): add docs on lowest latency routing buffer 2024-04-30 22:41:50 -07:00
Ishaan Jaff
81df36b298 docs - slack alerting 2024-04-29 21:33:03 -07:00
Ishaan Jaff
b1e888edad docs example logging to langfuse 2024-04-29 21:26:27 -07:00
Ishaan Jaff
0cad58f5c6 docs logging to langfuse on proxy 2024-04-29 21:26:15 -07:00
Ishaan Jaff
0c99ae9451 docs - fix kub.yaml config on docs 2024-04-29 21:20:29 -07:00
Krrish Dholakia
b46db8b891 feat(utils.py): json logs for raw request sent by litellm
make it easier to view verbose logs in datadog
2024-04-29 19:21:19 -07:00
Krrish Dholakia
bd79e8b516 docs(langfuse_integration.md): add 'existing_trace_id' to langfuse docs 2024-04-29 16:40:38 -07:00
Krrish Dholakia
77f155d158 docs(load_test.md): cleanup docs 2024-04-29 16:27:58 -07:00
Krrish Dholakia
af6a21f27c docs(load_test.md): add multi-instance router load test to docs 2024-04-29 16:25:56 -07:00
Krrish Dholakia
8f830bd948 docs(load_test.md): simplify doc 2024-04-29 16:00:02 -07:00
Krrish Dholakia
fcb83781ec docs(load_test.md): formatting 2024-04-29 15:58:41 -07:00
Krrish Dholakia
5fe0f38558 docs(load_test.md): load test multiple instances of the proxy w/ tpm/rpm limits on deployments 2024-04-29 15:58:14 -07:00
Krrish Dholakia
cef2d95bb4 docs(routing.md): add max parallel requests to router docs 2024-04-29 15:37:48 -07:00
Ishaan Jaff
8d26030b99 docs - track cost custom callbacks 2024-04-29 13:15:08 -07:00
Krrish Dholakia
f74a43aa78 docs(vllm.md): update docs to tell people to check openai-compatible endpoint docs for vllm 2024-04-28 09:48:03 -07:00
Krish Dholakia
1841b74f49
Merge branch 'main' into litellm_common_auth_params 2024-04-28 08:38:06 -07:00
Ishaan Jaff
3b8126cf51 docs - alerting 2024-04-27 15:33:21 -07:00
Ishaan Jaff
6762d07c7f
Merge pull request #3330 from BerriAI/litellm_rdct_msgs
[Feat] Redact Logging Messages/Response content on Logging Providers with `litellm.turn_off_message_logging=True`
2024-04-27 11:25:09 -07:00
Ishaan Jaff
f55838d185 sentry redact messages 2024-04-27 11:23:46 -07:00
Ishaan Jaff
01478c9148 docs - langfuse redact messages 2024-04-27 11:23:35 -07:00
Krrish Dholakia
48f19cf839 feat(utils.py): unify common auth params across azure/vertex_ai/bedrock/watsonx 2024-04-27 11:06:18 -07:00