Ishaan Jaff
|
05c5c34a13
|
docs cleanup
|
2024-08-16 11:38:53 -07:00 |
|
Krrish Dholakia
|
0771a004fd
|
docs(bedrock.md): add guardrails on config.yaml to docs
|
2024-08-14 22:11:19 -07:00 |
|
Krrish Dholakia
|
bcebb334b5
|
docs(team_logging.md): add key-based logging to docs
|
2024-08-14 21:49:55 -07:00 |
|
Krrish Dholakia
|
8ff001f241
|
docs(pass_through.md): add doc on using langfuse client sdk w/ litellm proxy
|
2024-08-14 21:43:31 -07:00 |
|
Ishaan Jaff
|
d3bdfac7a2
|
Merge pull request #5210 from BerriAI/litellm_add_prompt_caching_support
[Feat] Add Anthropic API Prompt Caching Support
|
2024-08-14 17:43:01 -07:00 |
|
Ishaan Jaff
|
43738c7d9e
|
docs using proxy with context caaching anthropic
|
2024-08-14 17:42:48 -07:00 |
|
Ishaan Jaff
|
2019682a97
|
docs add examples with litellm proxy
|
2024-08-14 17:13:26 -07:00 |
|
Ishaan Jaff
|
f186da28be
|
docs add examples doing context caching anthropic sdk
|
2024-08-14 17:07:51 -07:00 |
|
Ishaan Jaff
|
0290b15333
|
add test for caching tool calls
|
2024-08-14 16:19:14 -07:00 |
|
Ishaan Jaff
|
6f864ca7c7
|
docs Caching - Continuing Multi-Turn Convo
|
2024-08-14 15:26:25 -07:00 |
|
Ishaan Jaff
|
c14a432607
|
test amnthropic prompt caching
|
2024-08-14 14:59:46 -07:00 |
|
Krrish Dholakia
|
1ff93ed664
|
docs(model_management.md): add section on adding additional model information to proxy config
|
2024-08-14 14:39:48 -07:00 |
|
Ishaan Jaff
|
38868a0a45
|
use litellm_ prefix for new deployment metrics
|
2024-08-14 09:08:14 -07:00 |
|
Krrish Dholakia
|
4de5bc35a2
|
docs(sidebar.js): cleanup docs
|
2024-08-14 09:04:52 -07:00 |
|
Zbigniew Łukasiak
|
060c1677bf
|
Mismatch in example fixed
|
2024-08-14 15:07:10 +02:00 |
|
Ishaan Jaff
|
ff673e6305
|
Merge pull request #5191 from BerriAI/litellm_load_config_from_s3
[Feat] Allow loading LiteLLM config from s3 buckets
|
2024-08-13 21:19:16 -07:00 |
|
Ishaan Jaff
|
d24903d833
|
docs - set litellm config as s3 object
|
2024-08-13 20:26:29 -07:00 |
|
Keith Stevens
|
c19681b9a4
|
Improving the proxy docs for configuring with vllm
|
2024-08-13 16:07:41 -07:00 |
|
Ishaan Jaff
|
11e1fdc69f
|
Merge pull request #5180 from BerriAI/litellm_allow_controlling_logged_tags_langfuse
[Feat-Proxy+langfuse] LiteLLM-specific Tags on Langfuse - `cache_hit`, `cache_key`
|
2024-08-13 13:50:01 -07:00 |
|
Krrish Dholakia
|
bc45b82555
|
docs(user_keys.md): cleanup instructor docs
|
2024-08-13 13:15:46 -07:00 |
|
Krrish Dholakia
|
170e733c3d
|
docs(user_keys.md): cleanup docs
|
2024-08-13 13:14:27 -07:00 |
|
Ishaan Jaff
|
c3edd3bc52
|
docs control langfuse specific tags
|
2024-08-13 12:48:42 -07:00 |
|
Krrish Dholakia
|
638b5698f6
|
docs(team_logging.md): cleanup docs
|
2024-08-12 19:53:05 -07:00 |
|
Krrish Dholakia
|
f4c984878d
|
fix(utils.py): Break out of infinite streaming loop
Fixes https://github.com/BerriAI/litellm/issues/5158
|
2024-08-12 14:00:43 -07:00 |
|
Ishaan Jaff
|
9301f85046
|
docs mark oidc as beta
|
2024-08-12 09:01:36 -07:00 |
|
Krrish Dholakia
|
93d7634ea3
|
docs(perplexity.md): show how to get 'return_citations'
|
2024-08-12 09:01:14 -07:00 |
|
Ishaan Jaff
|
00443aa0f9
|
Merge pull request #5154 from BerriAI/litellm_send_prometheus_fallbacks_from_slack
[Feat-Proxy] send prometheus fallbacks stats to slack
|
2024-08-10 17:14:01 -07:00 |
|
Ishaan Jaff
|
ecec37e220
|
doc new prometheus metrics
|
2024-08-10 17:13:36 -07:00 |
|
Ishaan Jaff
|
0fc717fd0a
|
add fallback_reports as slack alert
|
2024-08-10 15:26:32 -07:00 |
|
Krrish Dholakia
|
5ad72419d2
|
docs(prefix.md): add prefix support to docs
|
2024-08-10 13:55:47 -07:00 |
|
Krrish Dholakia
|
0368ab1fba
|
docs(custom_llm_server.md): clarify what to use for modifying incoming/outgoing calls
|
2024-08-10 12:58:43 -07:00 |
|
Ishaan Jaff
|
200d798692
|
docs clean sidebar
|
2024-08-09 18:09:11 -07:00 |
|
Ishaan Jaff
|
44e463e209
|
docs migration policy
|
2024-08-09 18:06:37 -07:00 |
|
Ishaan Jaff
|
098da5b94b
|
docs add migration policy
|
2024-08-09 18:03:37 -07:00 |
|
Ishaan Jaff
|
b4aa0776d2
|
docs prometheus metrics
|
2024-08-09 09:07:31 -07:00 |
|
Ishaan Jaff
|
122d8ab2f4
|
doc Grounding vertex ai
|
2024-08-09 08:31:38 -07:00 |
|
Ishaan Jaff
|
00aea2734c
|
docs fix typo
|
2024-08-09 08:17:36 -07:00 |
|
Krrish Dholakia
|
16b84d8b16
|
docs(self_serve.md): add internal_user_budget_duration to docs
|
2024-08-08 23:54:26 -07:00 |
|
Ishaan Jaff
|
59cd3cb543
|
docs vertex context caching
|
2024-08-08 17:18:12 -07:00 |
|
Ishaan Jaff
|
91a46366cb
|
docs use (LLM Gateway) in some places
|
2024-08-08 17:00:52 -07:00 |
|
Ishaan Jaff
|
505e70eb56
|
docs vertex ai
|
2024-08-08 16:12:36 -07:00 |
|
Ishaan Jaff
|
e24e892af1
|
Merge pull request #5119 from BerriAI/litellm_add_gemini_context_caching_litellm
[Feat-Proxy] Add Support for VertexAI context caching
|
2024-08-08 16:08:58 -07:00 |
|
Ishaan Jaff
|
45d2b01856
|
docs vertex
|
2024-08-08 16:07:14 -07:00 |
|
Ishaan Jaff
|
d148a7f115
|
docs cachedContent endpoint
|
2024-08-08 16:06:23 -07:00 |
|
Ishaan Jaff
|
a87ece0a01
|
doc on using litellm proxy with vertex ai content caching
|
2024-08-08 11:45:46 -07:00 |
|
Krrish Dholakia
|
6f0b9a8775
|
docs(scheduler.md): cleanup docs to use /chat/completion endpoint
|
2024-08-07 21:49:06 -07:00 |
|
Krish Dholakia
|
6f0c7f8730
|
Merge pull request #5099 from BerriAI/litellm_personal_user_budgets
fix(user_api_key_auth.py): respect team budgets over user budget, if key belongs to team
|
2024-08-07 20:00:16 -07:00 |
|
Krish Dholakia
|
7d28b6ebc3
|
Merge branch 'main' into litellm_personal_user_budgets
|
2024-08-07 19:59:50 -07:00 |
|
Krrish Dholakia
|
a197afc0d5
|
docs(self_serve.md): cleanup docs on how to onboard new users + teams
|
2024-08-07 19:58:36 -07:00 |
|
Krrish Dholakia
|
07e5847e65
|
feat(router.py): allow using .acompletion() for request prioritization
allows /chat/completion endpoint to work for request prioritization calls
|
2024-08-07 16:43:12 -07:00 |
|