Commit graph

930 commits

Author SHA1 Message Date
Krrish Dholakia
f66af6ccb2 fix(utils.py): support checking if user defined max tokens exceeds model limit 2024-01-31 17:01:46 -08:00
Krrish Dholakia
13110b184d test(test_amazing_vertex_completion.py): fix test to handle api instability 2024-01-31 08:34:01 -08:00
Krish Dholakia
a9059eef10 Merge branch 'main' into litellm_no_store_cache_control 2024-01-30 21:44:57 -08:00
Krrish Dholakia
f5c88018dd feat(proxy_server.py): enable cache controls per key + no-store cache flag 2024-01-30 20:46:50 -08:00
Krrish Dholakia
90791e754d fix(utils.py): fix streaming delta content being none edge-case 2024-01-30 15:35:23 -08:00
Krrish Dholakia
456ec3add7 fix(utils.py): check if delta is none 2024-01-30 15:34:38 -08:00
Krrish Dholakia
85cbe5fb4c fix(utils.py): fix custom model with provider to also support time based pricing 2024-01-30 13:00:43 -08:00
Krrish Dholakia
e937662e8a build(ui/litellm_dashboard_v_2): allow app owner to create keys and view their keys 2024-01-27 13:48:35 -08:00
Krish Dholakia
e73e3be842 Merge pull request #1646 from BerriAI/litellm_image_gen_cost_tracking_proxy
Litellm image gen cost tracking proxy
2024-01-26 22:30:14 -08:00
Krrish Dholakia
6b1c1376b9 fix(utils.py): fix logging 2024-01-26 22:25:20 -08:00
Krrish Dholakia
d9fdddc301 fix(utils.py): enable cost tracking for image gen models on proxy 2024-01-26 20:51:13 -08:00
Krish Dholakia
a9a9969d6d Merge pull request #1641 from BerriAI/litellm_bedrock_region_based_pricing
feat(utils.py): support region based pricing for bedrock + use bedrock's token counts if given
2024-01-26 20:28:16 -08:00
Krrish Dholakia
e1beaf0945 feat(utils.py): support region based pricing for bedrock + use bedrock's token counts if given 2024-01-26 14:53:58 -08:00
ishaan-jaff
92779f6950 (feat) support dimensions param 2024-01-26 10:55:38 -08:00
Krish Dholakia
2abb430180 Merge pull request #1618 from BerriAI/litellm_sagemaker_cost_tracking_fixes
fix(utils.py): fix sagemaker cost tracking for streaming
2024-01-25 19:01:57 -08:00
Krrish Dholakia
4a34084ee6 fix(utils.py): completion_cost support for image gen models 2024-01-25 18:08:18 -08:00
Krrish Dholakia
e45ac58a11 refactor: trigger new bump 2024-01-25 16:40:23 -08:00
Krrish Dholakia
7591aba27e fix(main.py): allow vertex ai project and location to be set in completion() call 2024-01-25 16:40:23 -08:00
Krrish Dholakia
bbfed59e9c refactor: trigger new bump 2024-01-25 16:06:01 -08:00
Krrish Dholakia
d88e190304 fix(main.py): fix logging event loop for async logging but sync streaming 2024-01-25 15:59:53 -08:00
Krrish Dholakia
dbc929fddb fix(main.py): allow vertex ai project and location to be set in completion() call 2024-01-25 15:00:51 -08:00
Krrish Dholakia
806eef02dd fix(main.py): fix order of assembly for streaming chunks 2024-01-25 14:51:08 -08:00
Krrish Dholakia
402235dc5d fix(utils.py): fix sagemaker async logging for sync streaming
https://github.com/BerriAI/litellm/issues/1592
2024-01-25 12:49:45 -08:00
Krish Dholakia
f1d309d700 Merge branch 'main' into litellm_global_spend_updates 2024-01-24 20:20:15 -08:00
Krrish Dholakia
327ceb33b7 fix(ollama_chat.py): fix default token counting for ollama chat 2024-01-24 20:09:17 -08:00
Krrish Dholakia
d536374be0 fix(proxy_server.py): track cost for global proxy 2024-01-24 16:06:10 -08:00
Krish Dholakia
89e420b243 Merge branch 'main' into litellm_reset_key_budget 2024-01-23 18:10:32 -08:00
Krrish Dholakia
503ce7020b test(test_keys.py): use correct model name for token counting 2024-01-23 17:46:14 -08:00
Krish Dholakia
2ba8863f75 Merge pull request #1574 from BerriAI/litellm_fix_streaming_spend_tracking
[WIP] fix(utils.py): fix proxy streaming spend tracking
2024-01-23 17:07:40 -08:00
Krrish Dholakia
a5e53271d3 fix(utils.py): fix double hashing issue on spend logs, streaming usage metadata logging iss
ue for spend logs
2024-01-23 16:14:01 -08:00
Krrish Dholakia
344e232549 fix(utils.py): fix proxy streaming spend tracking 2024-01-23 15:59:03 -08:00
Krrish Dholakia
88486a3123 fix(utils.py): fix streaming cost tracking 2024-01-23 14:39:45 -08:00
ishaan-jaff
6e51c213e2 (fix) same response_id across chunk 2024-01-23 12:56:03 -08:00
ishaan-jaff
28f6a69dbf (fix) sagemaker streaming support 2024-01-23 12:31:16 -08:00
Krrish Dholakia
e04a4a7439 fix(utils.py): fix content policy violation check for streaming 2024-01-23 06:55:04 -08:00
Ishaan Jaff
97dd61a6cb Merge pull request #1561 from BerriAI/litellm_sagemaker_streaming
[Feat] Add REAL Sagemaker streaming
2024-01-22 22:10:20 -08:00
ishaan-jaff
09dd1ed68b v0 sagemaker_stream 2024-01-22 21:53:16 -08:00
Krrish Dholakia
29fe97b6a9 fix(router.py): fix order of dereferenced dictionaries 2024-01-22 21:42:25 -08:00
Krish Dholakia
3eaae0e73c Merge pull request #1557 from BerriAI/litellm_emit_spend_logs
feat(utils.py): emit response cost as part of logs
2024-01-22 21:02:40 -08:00
Krrish Dholakia
579dfc3013 test: fix tests 2024-01-22 20:20:17 -08:00
Krrish Dholakia
db2b7bfd4e fix(openai.py): fix linting issue 2024-01-22 18:20:15 -08:00
Krish Dholakia
8647f2a665 Merge pull request #1556 from BerriAI/litellm_importlib_issue
fix(utils.py): move from pkg_resources to importlib
2024-01-22 15:56:07 -08:00
Krrish Dholakia
2ea18785ca feat(utils.py): emit response cost as part of logs 2024-01-22 15:53:04 -08:00
Krrish Dholakia
737a5a7b38 fix(utils.py): fix debug log 2024-01-22 15:15:34 -08:00
Krrish Dholakia
70b0d0307c refactor(utils.py): fix linting errors 2024-01-22 15:15:34 -08:00
Krrish Dholakia
e423aeff85 fix: support streaming custom cost completion tracking 2024-01-22 15:15:34 -08:00
Krrish Dholakia
85b9ad7def fix(main.py): support custom pricing for embedding calls 2024-01-22 15:15:34 -08:00
Krrish Dholakia
480c3d3991 feat(utils.py): support custom cost tracking per second
https://github.com/BerriAI/litellm/issues/1374
2024-01-22 15:15:34 -08:00
Krrish Dholakia
78308ddf91 fix(utils.py): move from pkg_resources to importlib 2024-01-22 15:05:09 -08:00
Krrish Dholakia
b55dd5aa57 fix(utils.py): fix async/sync streaming logging 2024-01-22 13:54:51 -08:00