Ishaan Jaff
|
8a8f538329
|
Merge pull request #1829 from BerriAI/litellm_add_semantic_cache
[Feat] Add Semantic Caching to litellm💰
|
2024-02-06 13:18:59 -08:00 |
|
Krrish Dholakia
|
22913945b0
|
fix(utils.py): round max tokens to be int always
|
2024-02-06 13:10:52 -08:00 |
|
Ishaan Jaff
|
7cb69c72c8
|
Merge branch 'main' into litellm_add_semantic_cache
|
2024-02-06 11:18:43 -08:00 |
|
ishaan-jaff
|
a3b1e3bc84
|
(feat) redis-semantic cache
|
2024-02-06 08:54:36 -08:00 |
|
Krrish Dholakia
|
c2a523b954
|
fix(utils.py): use print_verbose for statements, so debug can be seen when running sdk
|
2024-02-06 07:30:26 -08:00 |
|
Krrish Dholakia
|
2e3748e6eb
|
fix(ollama_chat.py): fix ollama chat completion token counting
|
2024-02-06 07:30:26 -08:00 |
|
Krrish Dholakia
|
a1bbb16ab2
|
fix(langfuse.py): support logging failed llm api calls to langfuse
|
2024-02-05 16:16:15 -08:00 |
|
Krrish Dholakia
|
77fe71ee08
|
fix(utils.py): support together ai function calling
|
2024-02-05 15:30:44 -08:00 |
|
Krrish Dholakia
|
1bdb332454
|
fix(utils.py): handle count response tokens false case token counting
|
2024-02-05 08:47:10 -08:00 |
|
Krish Dholakia
|
640572647a
|
Merge pull request #1805 from BerriAI/litellm_cost_tracking_image_gen
feat(utils.py): support cost tracking for openai/azure image gen models
|
2024-02-03 22:23:22 -08:00 |
|
Krrish Dholakia
|
25a0e15727
|
fix(utils.py): support time based pricing for openai-compatible together ai
|
2024-02-03 19:59:32 -08:00 |
|
Krrish Dholakia
|
efb6123d28
|
fix(utils.py): support get_secret("TOGETHER_AI_TOKEN")
|
2024-02-03 19:35:09 -08:00 |
|
Krrish Dholakia
|
c49c88c8e5
|
fix(utils.py): route together ai calls to openai client
together ai is now openai-compatible
n
|
2024-02-03 19:22:48 -08:00 |
|
Krrish Dholakia
|
c2f674ebe0
|
fix(utils.py): fix conditional check
|
2024-02-03 18:58:58 -08:00 |
|
Krish Dholakia
|
9ab59045a3
|
Merge branch 'main' into litellm_embedding_caching_updates
|
2024-02-03 18:08:47 -08:00 |
|
Krrish Dholakia
|
ef0171e063
|
feat(utils.py): support cost tracking for openai/azure image gen models
|
2024-02-03 17:09:54 -08:00 |
|
Krish Dholakia
|
1d241b4001
|
Merge pull request #1802 from BerriAI/litellm_vertex_ai_high_traffic_fix
fix(vertex_ai.py): treat vertex ai high-traffic error as a rate limit error - allows user-controlled backoff logic to work here
|
2024-02-03 15:37:05 -08:00 |
|
Krish Dholakia
|
6408af11b6
|
Merge pull request #1799 from BerriAI/litellm_bedrock_stable_diffusion_support
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:59:00 -08:00 |
|
Krrish Dholakia
|
e617ef981d
|
fix(vertex_ai.py): treat vertex ai high-traffic error as a rate limit error - allows backoff logic to work here
|
2024-02-03 12:58:16 -08:00 |
|
Krrish Dholakia
|
36416360c4
|
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:08:38 -08:00 |
|
Krrish Dholakia
|
d9ba8668f4
|
feat(vertex_ai.py): vertex ai gecko text embedding support
|
2024-02-03 09:48:29 -08:00 |
|
Krrish Dholakia
|
1ed6842009
|
fix(utils.py): fix cost tracking for cache hits (should be 0)
|
2024-02-02 17:33:39 -08:00 |
|
ishaan-jaff
|
d69edac11b
|
(feat) use same id to log on s3, langfuse
|
2024-02-02 16:26:59 -08:00 |
|
Krrish Dholakia
|
31b19a420f
|
fix(helicone.py): fix logging
|
2024-02-02 14:30:01 -08:00 |
|
Krrish Dholakia
|
a11b505721
|
fix(utils.py): dynamic switch langfuse client based on keys
|
2024-02-02 14:03:46 -08:00 |
|
Krish Dholakia
|
93fb0134e5
|
Merge pull request #1761 from BerriAI/litellm_fix_dynamic_callbacks
fix(utils.py): override default success callbacks with dynamic callbacks if set
|
2024-02-02 13:06:55 -08:00 |
|
ishaan-jaff
|
aae54e0305
|
(feat) improve exceptions for OpenAI compatible -
|
2024-02-02 10:14:26 -08:00 |
|
Krrish Dholakia
|
127db122ef
|
test(test_keys.py): add more logging
|
2024-02-02 10:08:02 -08:00 |
|
ishaan-jaff
|
a55e0a9689
|
(feat) show correct provider in exceptions
|
2024-02-02 08:37:42 -08:00 |
|
Krrish Dholakia
|
61576fa6a9
|
fix(utils.py): logging fixes
|
2024-02-02 08:31:17 -08:00 |
|
Krrish Dholakia
|
5f9ea8a94f
|
fix(utils.py): add track cost callback to callback list for team id callbacks
|
2024-02-02 08:18:06 -08:00 |
|
Krrish Dholakia
|
92058cbcd4
|
fix(utils.py): override default success callbacks with dynamic callbacks if set
|
2024-02-02 06:21:43 -08:00 |
|
Krish Dholakia
|
7fc03bf745
|
Merge pull request #1749 from BerriAI/litellm_vertex_ai_model_garden
feat(vertex_ai.py): vertex ai model garden support
|
2024-02-01 21:52:12 -08:00 |
|
Krrish Dholakia
|
bbb5bd19e8
|
test: testing fixes
|
2024-02-01 21:20:46 -08:00 |
|
Krrish Dholakia
|
a301d8aa4b
|
feat(utils.py): support dynamic langfuse params and team settings on proxy
|
2024-02-01 21:08:24 -08:00 |
|
Krrish Dholakia
|
b79a6607b2
|
feat(utils.py): support dynamic success callbacks
|
2024-02-01 19:42:01 -08:00 |
|
Krrish Dholakia
|
245ec2430e
|
fix(utils.py): fix azure exception mapping
|
2024-02-01 19:05:20 -08:00 |
|
Krrish Dholakia
|
241f0aad5e
|
fix(utils.py): fix deepinfra streaming
|
2024-02-01 18:46:50 -08:00 |
|
Krrish Dholakia
|
d46df34ff5
|
fix(utils.py): fix streaming chunks to not return role, unless set
|
2024-02-01 09:55:56 -08:00 |
|
Krrish Dholakia
|
9593df23c4
|
test: add more logging
|
2024-01-31 19:31:07 -08:00 |
|
Krrish Dholakia
|
a32639fa79
|
fix(utils.py): support max token adjustment for sagemaker
|
2024-01-31 19:09:54 -08:00 |
|
Krrish Dholakia
|
93a52a2d35
|
fix(utils.py): set call_type at the top of the function
|
2024-01-31 18:23:44 -08:00 |
|
Krrish Dholakia
|
9dc972de70
|
fix(utils.py): fix get for dict
|
2024-01-31 18:15:47 -08:00 |
|
Krrish Dholakia
|
a0daac212d
|
fix(utils.py): support checking if user defined max tokens exceeds model limit
|
2024-01-31 17:01:46 -08:00 |
|
Krrish Dholakia
|
dd47c9d854
|
test(test_amazing_vertex_completion.py): fix test to handle api instability
|
2024-01-31 08:34:01 -08:00 |
|
Krish Dholakia
|
ce415a243d
|
Merge branch 'main' into litellm_no_store_cache_control
|
2024-01-30 21:44:57 -08:00 |
|
Krrish Dholakia
|
f9acad87dc
|
feat(proxy_server.py): enable cache controls per key + no-store cache flag
|
2024-01-30 20:46:50 -08:00 |
|
Krrish Dholakia
|
60cc3d6b30
|
fix(utils.py): fix streaming delta content being none edge-case
|
2024-01-30 15:35:23 -08:00 |
|
Krrish Dholakia
|
b01f31c41f
|
fix(utils.py): check if delta is none
|
2024-01-30 15:34:38 -08:00 |
|
Krrish Dholakia
|
d170e6e83a
|
fix(utils.py): fix custom model with provider to also support time based pricing
|
2024-01-30 13:00:43 -08:00 |
|