Ishaan Jaff
|
b718bf9b0a
|
feat add model access groups for teams
|
2024-08-17 17:10:10 -07:00 |
|
Ishaan Jaff
|
7171efc729
|
use model access groups for teams
|
2024-08-17 16:45:53 -07:00 |
|
Ishaan Jaff
|
ec671b491d
|
fix proxy all models test
|
2024-08-17 15:54:51 -07:00 |
|
Ishaan Jaff
|
a2178c026b
|
update tpm / rpm limit per model
|
2024-08-17 15:26:12 -07:00 |
|
Krrish Dholakia
|
5dc52aedc9
|
style(vertex_httpx.py): make vertex error string more helpful
|
2024-08-17 15:09:55 -07:00 |
|
Krish Dholakia
|
4db37fbf02
|
Merge pull request #5232 from Penagwin/fix_anthropic_tool_streaming_index
Fixes the `tool_use` indexes not being correctly mapped
|
2024-08-17 14:33:50 -07:00 |
|
Ishaan Jaff
|
7854652696
|
Merge pull request #5261 from BerriAI/litellm_set_model_rpm_tpm_limit
[Feat-Proxy] set rpm/tpm limits per api key per model
|
2024-08-17 14:30:54 -07:00 |
|
Ishaan Jaff
|
45cc7291f5
|
fix parallel request limiter tests
|
2024-08-17 14:21:59 -07:00 |
|
Ishaan Jaff
|
cf807d1541
|
fix parallel request limiter
|
2024-08-17 14:14:12 -07:00 |
|
Krish Dholakia
|
c7627a7de6
|
Merge pull request #5219 from dhlidongming/fix-messages-length-check
Fix incorrect message length check in cost calculator
|
2024-08-17 14:01:59 -07:00 |
|
Krish Dholakia
|
5e6700f985
|
Merge pull request #5260 from BerriAI/google_ai_studio_pass_through
Pass-through endpoints for Gemini - Google AI Studio
|
2024-08-17 13:51:51 -07:00 |
|
Ishaan Jaff
|
2c5f5996f3
|
add tpm limits per api key per model
|
2024-08-17 13:20:55 -07:00 |
|
Krrish Dholakia
|
b2ffa564d1
|
feat(pass_through_endpoints.py): support streaming requests
|
2024-08-17 12:46:57 -07:00 |
|
Ishaan Jaff
|
8578301116
|
fix async_pre_call_hook in parallel request limiter
|
2024-08-17 12:42:28 -07:00 |
|
Ishaan Jaff
|
db8f789318
|
Merge pull request #5259 from BerriAI/litellm_return_remaining_tokens_in_header
[Feat] return `x-litellm-key-remaining-requests-{model}`: 1, `x-litellm-key-remaining-tokens-{model}: None` in response headers
|
2024-08-17 12:41:16 -07:00 |
|
Ishaan Jaff
|
9f6630912d
|
feat return rmng tokens for model for api key
|
2024-08-17 12:35:10 -07:00 |
|
Krrish Dholakia
|
29bedae79f
|
feat(google_ai_studio_endpoints.py): support pass-through endpoint for all google ai studio requests
New Feature
|
2024-08-17 10:46:59 -07:00 |
|
Ishaan Jaff
|
a62277a6aa
|
feat - use commong helper for getting model group
|
2024-08-17 10:46:04 -07:00 |
|
Ishaan Jaff
|
2dd098f384
|
show correct metric
|
2024-08-17 10:12:23 -07:00 |
|
Ishaan Jaff
|
03196742d2
|
add litellm-key-remaining-tokens on prometheus
|
2024-08-17 10:02:20 -07:00 |
|
Ishaan Jaff
|
8ae626b31f
|
feat add settings for rpm/tpm limits for a model
|
2024-08-17 09:16:01 -07:00 |
|
Krrish Dholakia
|
668ea6cbc7
|
fix(pass_through_endpoints.py): fix returned response headers for pass-through endpoitns
|
2024-08-17 09:00:00 -07:00 |
|
Krrish Dholakia
|
3b9eb7ca1e
|
docs(vertex_ai.md): cleanup docs
|
2024-08-17 08:38:01 -07:00 |
|
Krrish Dholakia
|
b3dd0bb738
|
bump: version 1.43.16 → 1.43.17
|
2024-08-16 21:34:35 -07:00 |
|
Krish Dholakia
|
88fccb2427
|
Merge branch 'main' into litellm_log_model_price_information
|
2024-08-16 19:34:16 -07:00 |
|
Krish Dholakia
|
0916197c9d
|
Merge pull request #5244 from BerriAI/litellm_better_error_logging_sentry
refactor: replace .error() with .exception() logging for better debugging on sentry
|
2024-08-16 19:16:20 -07:00 |
|
Krish Dholakia
|
1844e01133
|
Merge pull request #5251 from Manouchehri/oidc-improvements-20240816
(oidc): Add support for loading tokens via a file, env var, and path in env var
|
2024-08-16 19:15:31 -07:00 |
|
Ishaan Jaff
|
824ea32452
|
track rpm/tpm usage per key+model
|
2024-08-16 18:28:58 -07:00 |
|
Ishaan Jaff
|
dbc9c9e8d8
|
user api key auth rpm_limit_per_model
|
2024-08-16 18:22:35 -07:00 |
|
Krish Dholakia
|
6fe21d6dd4
|
Merge pull request #5255 from BerriAI/litellm_fix_token_counter
fix(utils.py): fix get_image_dimensions to handle more image types
|
2024-08-16 17:27:27 -07:00 |
|
Krrish Dholakia
|
9efe9982f5
|
fix(health_check.py): return 'missing mode' error message, if error with health check, and mode is missing
|
2024-08-16 17:24:29 -07:00 |
|
Ishaan Jaff
|
937471223a
|
fix databricks streaming test
|
2024-08-16 16:56:08 -07:00 |
|
Krrish Dholakia
|
9609505d0c
|
fix(litellm_logging.py): fix price information logging to s3
|
2024-08-16 16:42:38 -07:00 |
|
Krrish Dholakia
|
ef51f8600d
|
feat(litellm_logging.py): support logging model price information to s3 logs
|
2024-08-16 16:21:34 -07:00 |
|
Ishaan Jaff
|
31e698e180
|
fix predictions image generation response
|
2024-08-16 15:06:12 -07:00 |
|
Ishaan Jaff
|
5790275de5
|
skip InternalServerError on vertex test
|
2024-08-16 14:51:51 -07:00 |
|
Ishaan Jaff
|
6de7785442
|
Merge pull request #5243 from BerriAI/litellm_add_bedrock_traces_in_response
[Feat] Add bedrock Guardrail `traces ` in response when trace=enabled
|
2024-08-16 14:49:20 -07:00 |
|
Ishaan Jaff
|
55df861291
|
docs oauh 2.0 enterprise feature
|
2024-08-16 14:00:24 -07:00 |
|
Ishaan Jaff
|
9a9710b8a1
|
add debugging for oauth2.0
|
2024-08-16 13:40:32 -07:00 |
|
Ishaan Jaff
|
8745e1608a
|
allow using oauth2 checks for logging into proxy
|
2024-08-16 13:36:29 -07:00 |
|
Ishaan Jaff
|
d2be2d6e23
|
add init commit for oauth 2 checks
|
2024-08-16 13:30:22 -07:00 |
|
David Manouchehri
|
f24e986534
|
(oidc): Add support for loading tokens via a file, environment variable, and from a file path set in an env var.
|
2024-08-16 20:13:07 +00:00 |
|
Ishaan Jaff
|
f2569740fa
|
ui new build
|
2024-08-16 12:53:23 -07:00 |
|
Krrish Dholakia
|
3e42ee1bbb
|
fix(utils.py): fix get_image_dimensions to handle more image types
Fixes https://github.com/BerriAI/litellm/issues/5205
|
2024-08-16 12:00:04 -07:00 |
|
Ishaan Jaff
|
b82d120c47
|
add provider_specific_fields to GenericStreamingChunk
|
2024-08-16 11:38:22 -07:00 |
|
Ishaan Jaff
|
262bf14917
|
return traces in bedrock guardrails when enabled
|
2024-08-16 11:35:43 -07:00 |
|
Krrish Dholakia
|
83ed174059
|
fix(__init__.py): fix models_by_provider to include cohere_chat models
Fixes https://github.com/BerriAI/litellm/issues/5201
|
2024-08-16 11:33:23 -07:00 |
|
Krrish Dholakia
|
2874b94fb1
|
refactor: replace .error() with .exception() logging for better debugging on sentry
|
2024-08-16 09:22:47 -07:00 |
|
Ishaan Jaff
|
98c9191f84
|
pass trace through for bedrock guardrails
|
2024-08-16 09:10:56 -07:00 |
|
Ishaan Jaff
|
15334cfae3
|
Merge pull request #5173 from gitravin/rn/sagemaker-zero-temp
Allow zero temperature for Sagemaker models based on config
|
2024-08-16 08:45:44 -07:00 |
|