Commit graph

2434 commits

Author SHA1 Message Date
Ishaan Jaff
107254ebd3 fix endpoint name on router 2024-08-16 12:46:43 -07:00
Ishaan Jaff
cd7f9c250c docs add example on setting temp=0 for sagemaker 2024-08-16 12:04:35 -07:00
Ishaan Jaff
8f106eca68 docs sagemaker - add example using with proxy 2024-08-16 11:47:13 -07:00
Ishaan Jaff
05c5c34a13 docs cleanup 2024-08-16 11:38:53 -07:00
Krrish Dholakia
0771a004fd docs(bedrock.md): add guardrails on config.yaml to docs 2024-08-14 22:11:19 -07:00
Krrish Dholakia
bcebb334b5 docs(team_logging.md): add key-based logging to docs 2024-08-14 21:49:55 -07:00
Krrish Dholakia
8ff001f241 docs(pass_through.md): add doc on using langfuse client sdk w/ litellm proxy 2024-08-14 21:43:31 -07:00
Ishaan Jaff
d3bdfac7a2 Merge pull request #5210 from BerriAI/litellm_add_prompt_caching_support
[Feat] Add Anthropic API Prompt Caching Support
2024-08-14 17:43:01 -07:00
Ishaan Jaff
43738c7d9e docs using proxy with context caaching anthropic 2024-08-14 17:42:48 -07:00
Ishaan Jaff
2019682a97 docs add examples with litellm proxy 2024-08-14 17:13:26 -07:00
Ishaan Jaff
f186da28be docs add examples doing context caching anthropic sdk 2024-08-14 17:07:51 -07:00
Ishaan Jaff
0290b15333 add test for caching tool calls 2024-08-14 16:19:14 -07:00
Ishaan Jaff
6f864ca7c7 docs Caching - Continuing Multi-Turn Convo 2024-08-14 15:26:25 -07:00
Ishaan Jaff
c14a432607 test amnthropic prompt caching 2024-08-14 14:59:46 -07:00
Krrish Dholakia
1ff93ed664 docs(model_management.md): add section on adding additional model information to proxy config 2024-08-14 14:39:48 -07:00
Ishaan Jaff
38868a0a45 use litellm_ prefix for new deployment metrics 2024-08-14 09:08:14 -07:00
Krrish Dholakia
4de5bc35a2 docs(sidebar.js): cleanup docs 2024-08-14 09:04:52 -07:00
Zbigniew Łukasiak
060c1677bf Mismatch in example fixed 2024-08-14 15:07:10 +02:00
Ishaan Jaff
ff673e6305 Merge pull request #5191 from BerriAI/litellm_load_config_from_s3
[Feat] Allow loading LiteLLM config from s3 buckets
2024-08-13 21:19:16 -07:00
Ishaan Jaff
d24903d833 docs - set litellm config as s3 object 2024-08-13 20:26:29 -07:00
Keith Stevens
c19681b9a4 Improving the proxy docs for configuring with vllm 2024-08-13 16:07:41 -07:00
Ishaan Jaff
11e1fdc69f Merge pull request #5180 from BerriAI/litellm_allow_controlling_logged_tags_langfuse
[Feat-Proxy+langfuse] LiteLLM-specific Tags on Langfuse - `cache_hit`, `cache_key`
2024-08-13 13:50:01 -07:00
Krrish Dholakia
bc45b82555 docs(user_keys.md): cleanup instructor docs 2024-08-13 13:15:46 -07:00
Krrish Dholakia
170e733c3d docs(user_keys.md): cleanup docs 2024-08-13 13:14:27 -07:00
Ishaan Jaff
c3edd3bc52 docs control langfuse specific tags 2024-08-13 12:48:42 -07:00
Krrish Dholakia
638b5698f6 docs(team_logging.md): cleanup docs 2024-08-12 19:53:05 -07:00
Krrish Dholakia
f4c984878d fix(utils.py): Break out of infinite streaming loop
Fixes https://github.com/BerriAI/litellm/issues/5158
2024-08-12 14:00:43 -07:00
Ishaan Jaff
9301f85046 docs mark oidc as beta 2024-08-12 09:01:36 -07:00
Krrish Dholakia
93d7634ea3 docs(perplexity.md): show how to get 'return_citations' 2024-08-12 09:01:14 -07:00
Ishaan Jaff
00443aa0f9 Merge pull request #5154 from BerriAI/litellm_send_prometheus_fallbacks_from_slack
[Feat-Proxy] send prometheus fallbacks stats to slack
2024-08-10 17:14:01 -07:00
Ishaan Jaff
ecec37e220 doc new prometheus metrics 2024-08-10 17:13:36 -07:00
Ishaan Jaff
0fc717fd0a add fallback_reports as slack alert 2024-08-10 15:26:32 -07:00
Krrish Dholakia
5ad72419d2 docs(prefix.md): add prefix support to docs 2024-08-10 13:55:47 -07:00
Krrish Dholakia
0368ab1fba docs(custom_llm_server.md): clarify what to use for modifying incoming/outgoing calls 2024-08-10 12:58:43 -07:00
Ishaan Jaff
200d798692 docs clean sidebar 2024-08-09 18:09:11 -07:00
Ishaan Jaff
44e463e209 docs migration policy 2024-08-09 18:06:37 -07:00
Ishaan Jaff
098da5b94b docs add migration policy 2024-08-09 18:03:37 -07:00
Ishaan Jaff
b4aa0776d2 docs prometheus metrics 2024-08-09 09:07:31 -07:00
Ishaan Jaff
122d8ab2f4 doc Grounding vertex ai 2024-08-09 08:31:38 -07:00
Ishaan Jaff
00aea2734c docs fix typo 2024-08-09 08:17:36 -07:00
Krrish Dholakia
16b84d8b16 docs(self_serve.md): add internal_user_budget_duration to docs 2024-08-08 23:54:26 -07:00
Ishaan Jaff
59cd3cb543 docs vertex context caching 2024-08-08 17:18:12 -07:00
Ishaan Jaff
91a46366cb docs use (LLM Gateway) in some places 2024-08-08 17:00:52 -07:00
Ishaan Jaff
505e70eb56 docs vertex ai 2024-08-08 16:12:36 -07:00
Ishaan Jaff
e24e892af1 Merge pull request #5119 from BerriAI/litellm_add_gemini_context_caching_litellm
[Feat-Proxy] Add Support for VertexAI context caching
2024-08-08 16:08:58 -07:00
Ishaan Jaff
45d2b01856 docs vertex 2024-08-08 16:07:14 -07:00
Ishaan Jaff
d148a7f115 docs cachedContent endpoint 2024-08-08 16:06:23 -07:00
Ishaan Jaff
a87ece0a01 doc on using litellm proxy with vertex ai content caching 2024-08-08 11:45:46 -07:00
Krrish Dholakia
6f0b9a8775 docs(scheduler.md): cleanup docs to use /chat/completion endpoint 2024-08-07 21:49:06 -07:00
Krish Dholakia
6f0c7f8730 Merge pull request #5099 from BerriAI/litellm_personal_user_budgets
fix(user_api_key_auth.py): respect team budgets over user budget, if key belongs to team
2024-08-07 20:00:16 -07:00