Commit graph

16578 commits

Author SHA1 Message Date
Ishaan Jaff
8cd1963c11 feat - guardrails v2 2024-08-19 18:24:20 -07:00
Ishaan Jaff
7721b9b176 docs fix typo 2024-08-19 15:44:15 -07:00
Ishaan Jaff
916e0655e7 doc update typo 2024-08-19 15:43:40 -07:00
Ishaan Jaff
d2f4535cf5 docs updates based on feedback 2024-08-19 15:36:10 -07:00
Ishaan Jaff
b7d4031f89 doc aporia_w_litellm 2024-08-19 14:36:55 -07:00
Ishaan Jaff
a0361b0e76 run during_call_hook 2024-08-19 12:07:46 -07:00
Ishaan Jaff
613bd1babd feat - return applied guardrails in response headers 2024-08-19 11:56:20 -07:00
Ishaan Jaff
4685b9909a feat - allow accessing data post success call 2024-08-19 11:35:33 -07:00
Ishaan Jaff
8cb62213e1 feat run aporia as post call success hook 2024-08-19 11:25:31 -07:00
Ishaan Jaff
601be5cb44
Merge pull request #5280 from lowjiansheng/js/litellm-help-0.2.3
Update helm repo to v0.2.3
2024-08-19 10:01:31 -07:00
lowjiansheng
f1c39510cb add 0.2.3 helm 2024-08-19 23:59:58 +08:00
Krrish Dholakia
04d69464e2 fix(ollama.py): fix ollama embeddings - pass optional params
Fixes https://github.com/BerriAI/litellm/issues/5267
2024-08-19 08:45:26 -07:00
Krrish Dholakia
cc42f96d6a fix(ollama_chat.py): fix sync tool calling
Fixes https://github.com/BerriAI/litellm/issues/5245
2024-08-19 08:31:46 -07:00
Krrish Dholakia
b8e4ef0abf docs(json_mode.md): add azure openai models to doc 2024-08-19 07:19:23 -07:00
Ishaan Jaff
398295116f inly write model tpm/rpm tracking when user set it 2024-08-18 09:58:09 -07:00
Krish Dholakia
f42ac2c9d8
Merge pull request #5264 from BerriAI/litellm_bedrock_pass_through
feat: Bedrock pass-through endpoint support (All endpoints)
2024-08-18 09:55:22 -07:00
Ishaan Jaff
69afb07dea sleep before checi g 2024-08-17 19:50:37 -07:00
Krrish Dholakia
663a0c1b83 feat(Support-pass-through-for-bedrock-endpoints): Allows pass-through support for bedrock endpoints 2024-08-17 17:57:43 -07:00
Ishaan Jaff
5adb7e29b9 fix test pass through 2024-08-17 17:42:51 -07:00
Ishaan Jaff
0bc67761dc docs access groups 2024-08-17 17:38:28 -07:00
Ishaan Jaff
3cba235109 docs virtual key access groups 2024-08-17 17:37:23 -07:00
Ishaan Jaff
83515e88ce
Merge pull request #5263 from BerriAI/litellm_support_access_groups
[Feat-Proxy] Use model access groups for teams
2024-08-17 17:11:11 -07:00
Ishaan Jaff
6fee350938 feat add model access groups for teams 2024-08-17 17:10:10 -07:00
Krrish Dholakia
f7a2e04426 feat(pass_through_endpoints.py): add pass-through support for all cohere endpoints 2024-08-17 16:57:55 -07:00
Ishaan Jaff
9b239111f4 fix test update tpm / rpm limits for a key 2024-08-17 16:57:23 -07:00
Ishaan Jaff
08db691dec use model access groups for teams 2024-08-17 16:45:53 -07:00
Ishaan Jaff
d9c91838ce docs cleanup 2024-08-17 15:59:23 -07:00
Ishaan Jaff
eff874bf05 fix proxy all models test 2024-08-17 15:54:51 -07:00
Ishaan Jaff
78d30990a3 docs clean up virtual key access 2024-08-17 15:39:50 -07:00
Ishaan Jaff
2a18a65f9e bump: version 1.43.17 → 1.43.18 2024-08-17 15:27:50 -07:00
Ishaan Jaff
b83fa87880 update tpm / rpm limit per model 2024-08-17 15:26:12 -07:00
Krrish Dholakia
db54b66457 style(vertex_httpx.py): make vertex error string more helpful 2024-08-17 15:09:55 -07:00
Ishaan Jaff
671663abe6 docs rate limits per model per api key 2024-08-17 14:50:17 -07:00
Krish Dholakia
be37310e94
Merge pull request #5232 from Penagwin/fix_anthropic_tool_streaming_index
Fixes the `tool_use` indexes not being correctly mapped
2024-08-17 14:33:50 -07:00
Ishaan Jaff
a60fc3ad70
Merge pull request #5261 from BerriAI/litellm_set_model_rpm_tpm_limit
[Feat-Proxy] set rpm/tpm limits per api key per model
2024-08-17 14:30:54 -07:00
Ishaan Jaff
653d2e6ce0 fix parallel request limiter tests 2024-08-17 14:21:59 -07:00
Ishaan Jaff
221e5b829b fix parallel request limiter 2024-08-17 14:14:12 -07:00
Krish Dholakia
5731287f1b
Merge pull request #5221 from kiriloman/adjust-pricing-file
[PRICING] Use specific llama2 and llama3 model names in Ollama
2024-08-17 14:03:20 -07:00
Krish Dholakia
1a3b686580
Merge pull request #5219 from dhlidongming/fix-messages-length-check
Fix incorrect message length check in cost calculator
2024-08-17 14:01:59 -07:00
Krish Dholakia
ff6ff133ee
Merge pull request #5260 from BerriAI/google_ai_studio_pass_through
Pass-through endpoints for Gemini - Google AI Studio
2024-08-17 13:51:51 -07:00
Krrish Dholakia
0df41653f3 docs(google_ai_studio.md): add docs on google ai studio pass through endpoints 2024-08-17 13:47:05 -07:00
Ishaan Jaff
b35b09ea93 docs clean up emojis 2024-08-17 13:30:11 -07:00
Ishaan Jaff
9b0bd54571 docs cleanup - reduce emojis 2024-08-17 13:28:34 -07:00
Ishaan Jaff
68b54bed85 add tpm limits per api key per model 2024-08-17 13:20:55 -07:00
Krrish Dholakia
fd44cf8d26 feat(pass_through_endpoints.py): support streaming requests 2024-08-17 12:46:57 -07:00
Ishaan Jaff
fa96610bbc fix async_pre_call_hook in parallel request limiter 2024-08-17 12:42:28 -07:00
Ishaan Jaff
feb8c3c5b4
Merge pull request #5259 from BerriAI/litellm_return_remaining_tokens_in_header
[Feat] return `x-litellm-key-remaining-requests-{model}`: 1, `x-litellm-key-remaining-tokens-{model}: None` in response headers
2024-08-17 12:41:16 -07:00
Ishaan Jaff
ee0f772b5c feat return rmng tokens for model for api key 2024-08-17 12:35:10 -07:00
Krrish Dholakia
bc0023a409 feat(google_ai_studio_endpoints.py): support pass-through endpoint for all google ai studio requests
New Feature
2024-08-17 10:46:59 -07:00
Ishaan Jaff
5985c7e933 feat - use commong helper for getting model group 2024-08-17 10:46:04 -07:00