ishaan-jaff
|
6267026e9c
|
(fix) importing enterprise features
|
2024-02-15 17:24:27 -08:00 |
|
ishaan-jaff
|
85eefcb4f4
|
(chore) debug sys path docker error
|
2024-02-15 17:23:07 -08:00 |
|
Krrish Dholakia
|
36ec6ed4d1
|
fix(utils.py): fix local tiktoken package import
|
2024-02-15 16:18:54 -08:00 |
|
ishaan-jaff
|
8b1a2a8518
|
(feat) fix api callback imports
|
2024-02-15 16:15:36 -08:00 |
|
ishaan-jaff
|
bbface2a01
|
(feat) log with generic logger
|
2024-02-15 13:43:16 -08:00 |
|
Krish Dholakia
|
f565949fd5
|
Merge pull request #1947 from sorokine/main
Update utils.py: switched from pkg_resources to importlib.resources for `filename` stored in the `os.environ(TIKTOKEN_CACHE_DIR)`
|
2024-02-14 22:02:35 -08:00 |
|
Krish Dholakia
|
eda9fa300e
|
Merge branch 'main' into litellm_aioboto3_sagemaker
|
2024-02-14 21:46:58 -08:00 |
|
Rena Lu
|
f5971a373e
|
handle optional params
|
2024-02-14 23:27:16 +00:00 |
|
ishaan-jaff
|
6c1c2e8c7d
|
(feat) add moderation on router
|
2024-02-14 11:00:09 -08:00 |
|
kan-bayashi
|
eb6d181ba6
|
fix the issue when using multiple tools in gemini
|
2024-02-14 13:09:35 +09:00 |
|
Krrish Dholakia
|
1a57b21fd6
|
fix(utils.py): fix streaming rule calling
|
2024-02-12 22:36:32 -08:00 |
|
Krrish Dholakia
|
fd6f64a4ae
|
feat(utils.py): enable post call rules for streaming
|
2024-02-12 22:08:04 -08:00 |
|
Krrish Dholakia
|
5de569fcb1
|
feat(sagemaker.py): aioboto3 streaming support
|
2024-02-12 21:18:34 -08:00 |
|
Alexandre Sorokine
|
acfbda7b6b
|
Update utils.py: switched from pkg_resources to importlib.resources for filename stored in the os/environ(TIKTOKEN_CACHE_DIR)
|
2024-02-12 16:28:38 -05:00 |
|
ishaan-jaff
|
e71009e4b4
|
(feat) fix litellm verbose logger
|
2024-02-09 19:01:21 -08:00 |
|
ishaan-jaff
|
1884c24d31
|
(bedrock) raise timeout error
|
2024-02-09 14:37:34 -08:00 |
|
Krrish Dholakia
|
07aa05bf17
|
fix(test_parallel_request_limiter.py): use mock responses for streaming
|
2024-02-08 21:45:38 -08:00 |
|
Ishaan Jaff
|
82f10ed322
|
Merge pull request #1892 from BerriAI/litellm_speed_up_s3_logging
[FEAT] 76 % Faster s3 logging Proxy / litellm.acompletion / router.acompletion 🚀
|
2024-02-08 11:48:36 -08:00 |
|
ishaan-jaff
|
6328baac6a
|
(feat) speed up s3 logging
|
2024-02-08 10:59:54 -08:00 |
|
ishaan-jaff
|
798509a067
|
(feat) fix s3 cache
|
2024-02-08 10:01:56 -08:00 |
|
ishaan-jaff
|
6aef843911
|
(fix) azure cost calc
|
2024-02-07 17:33:10 -08:00 |
|
ishaan-jaff
|
7bed984179
|
(fix) cost tracking
|
2024-02-07 17:06:05 -08:00 |
|
ishaan-jaff
|
67305fcb20
|
(feat) use base_model for azure response_cost
|
2024-02-07 16:33:07 -08:00 |
|
Krrish Dholakia
|
ef95b70b87
|
fix(utils.py): fix ollama stop sequence mapping
|
2024-02-07 13:14:03 -08:00 |
|
ishaan-jaff
|
e2ccdb7a1b
|
(feat) redis-semantic cache
|
2024-02-06 13:30:55 -08:00 |
|
Krrish Dholakia
|
659a460923
|
fix(utils.py): round max tokens to be int always
|
2024-02-06 13:17:26 -08:00 |
|
Krrish Dholakia
|
32639bf398
|
fix(utils.py): return finish reason for last vertex ai chunk
|
2024-02-06 09:21:03 -08:00 |
|
Krrish Dholakia
|
9b2a2e6c8b
|
fix(utils.py): use print_verbose for statements, so debug can be seen when running sdk
|
2024-02-06 07:30:26 -08:00 |
|
Krrish Dholakia
|
9a526b6cd4
|
fix(ollama_chat.py): fix ollama chat completion token counting
|
2024-02-06 07:30:26 -08:00 |
|
Krrish Dholakia
|
1b975bbe28
|
fix(langfuse.py): support logging failed llm api calls to langfuse
|
2024-02-05 16:16:15 -08:00 |
|
Krrish Dholakia
|
2e3c7088d1
|
fix(utils.py): support together ai function calling
|
2024-02-05 15:30:44 -08:00 |
|
Krrish Dholakia
|
1c8a3973ab
|
fix(utils.py): handle count response tokens false case token counting
|
2024-02-05 08:47:10 -08:00 |
|
Krish Dholakia
|
92d50616c2
|
Merge pull request #1805 from BerriAI/litellm_cost_tracking_image_gen
feat(utils.py): support cost tracking for openai/azure image gen models
|
2024-02-03 22:23:22 -08:00 |
|
Krrish Dholakia
|
f9aa05c268
|
fix(utils.py): support time based pricing for openai-compatible together ai
|
2024-02-03 19:59:32 -08:00 |
|
Krrish Dholakia
|
740eda84ee
|
fix(utils.py): support get_secret("TOGETHER_AI_TOKEN")
|
2024-02-03 19:35:09 -08:00 |
|
Krrish Dholakia
|
85a3515d83
|
fix(utils.py): route together ai calls to openai client
together ai is now openai-compatible
n
|
2024-02-03 19:22:48 -08:00 |
|
Krrish Dholakia
|
8195bfd286
|
fix(utils.py): fix conditional check
|
2024-02-03 18:58:58 -08:00 |
|
Krish Dholakia
|
45cbb3cf3d
|
Merge branch 'main' into litellm_embedding_caching_updates
|
2024-02-03 18:08:47 -08:00 |
|
Krrish Dholakia
|
96c2c64dce
|
feat(utils.py): support cost tracking for openai/azure image gen models
|
2024-02-03 17:09:54 -08:00 |
|
Krish Dholakia
|
39305f1e07
|
Merge pull request #1802 from BerriAI/litellm_vertex_ai_high_traffic_fix
fix(vertex_ai.py): treat vertex ai high-traffic error as a rate limit error - allows user-controlled backoff logic to work here
|
2024-02-03 15:37:05 -08:00 |
|
Krish Dholakia
|
dbaad8ae56
|
Merge pull request #1799 from BerriAI/litellm_bedrock_stable_diffusion_support
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:59:00 -08:00 |
|
Krrish Dholakia
|
16a43224ba
|
fix(vertex_ai.py): treat vertex ai high-traffic error as a rate limit error - allows backoff logic to work here
|
2024-02-03 12:58:16 -08:00 |
|
Krrish Dholakia
|
5994c1e7ef
|
feat(bedrock.py): add stable diffusion image generation support
|
2024-02-03 12:08:38 -08:00 |
|
Krrish Dholakia
|
3f23b18dad
|
feat(vertex_ai.py): vertex ai gecko text embedding support
|
2024-02-03 09:48:29 -08:00 |
|
Krrish Dholakia
|
7699f48930
|
fix(utils.py): fix cost tracking for cache hits (should be 0)
|
2024-02-02 17:33:39 -08:00 |
|
ishaan-jaff
|
2e4b1bcf4d
|
(feat) use same id to log on s3, langfuse
|
2024-02-02 16:26:59 -08:00 |
|
Krrish Dholakia
|
34886b0261
|
fix(helicone.py): fix logging
|
2024-02-02 14:30:01 -08:00 |
|
Krrish Dholakia
|
af1319bd9f
|
fix(utils.py): dynamic switch langfuse client based on keys
|
2024-02-02 14:03:46 -08:00 |
|
Krish Dholakia
|
6824014c5f
|
Merge pull request #1761 from BerriAI/litellm_fix_dynamic_callbacks
fix(utils.py): override default success callbacks with dynamic callbacks if set
|
2024-02-02 13:06:55 -08:00 |
|
ishaan-jaff
|
fdddb69dea
|
(feat) improve exceptions for OpenAI compatible -
|
2024-02-02 10:14:26 -08:00 |
|