Commit graph

596 commits

Author SHA1 Message Date
ishaan-jaff
70fc5afb5d (fix) streaming + function / tool calling 2023-11-18 16:23:29 -08:00
Krrish Dholakia
478bd7def6 fix(router.py): introducing usage-based-routing 2023-11-17 17:56:26 -08:00
ishaan-jaff
88200432b0 (feat) support parallel function calling 2023-11-17 15:51:27 -08:00
ishaan-jaff
7de87c845b (feat) improve logging - show model_call_details 2023-11-17 15:51:27 -08:00
Krrish Dholakia
0ab6b2451d fix(acompletion): support client side timeouts + raise exceptions correctly for async calls 2023-11-17 15:39:47 -08:00
ishaan-jaff
7abb65d53f (feat) completion: add response_format, seed, tools, tool_choice 2023-11-17 13:59:57 -08:00
ishaan-jaff
bd82559553 (v1.0+ breaking change) get_max_tokens -> return int 2023-11-17 10:38:50 -08:00
Krrish Dholakia
9bd1f4ebd0 fix(utils.py): improve exception mapping for vertex ai 2023-11-16 22:02:26 -08:00
Krrish Dholakia
7456c26940 docs(routing.md): update tutorial on deploying router 2023-11-16 21:46:43 -08:00
Krrish Dholakia
51bf637656 feat: global client for sync + async calls (openai + Azure only) 2023-11-16 14:44:13 -08:00
Krrish Dholakia
a94c09c13c fix(openai.py): handling extra headers 2023-11-16 12:48:21 -08:00
ishaan-jaff
9e072f87bd (fix) bedrock meta llama optional params 2023-11-16 12:38:27 -08:00
ishaan-jaff
04971674b4 (fix) only decode chunk when it's not a str 2023-11-16 12:24:31 -08:00
Krrish Dholakia
bf0f8b824c fix(azure.py): use openai client sdk for handling sync+async calling 2023-11-16 12:08:12 -08:00
ishaan-jaff
da8c2f4a4a (fix) HF api + streaming 2023-11-16 11:59:56 -08:00
ishaan-jaff
c84fe1af5e (fix) langfuse logging + openai streaming when chunk = [DONE} 2023-11-16 10:45:35 -08:00
Krrish Dholakia
bb51216846 fix(openai.py): supporting openai client sdk for handling sync + async calls (incl. for openai-compatible apis) 2023-11-16 10:35:03 -08:00
Ishaan Jaff
d6ad62d793
Merge pull request #811 from dchristian3188/bedrock-llama
Bedrock llama
2023-11-16 07:57:50 -08:00
Ishaan Jaff
3f95fd2da5
Merge pull request #826 from rodneyxr/ollama-fixes
Fix typo for initial_prompt_value and too many values to unpack error
2023-11-16 07:55:53 -08:00
David Christian
45ae4a76e4 updated utils for bedrock.meta streaming 2023-11-16 07:12:27 -08:00
Krrish Dholakia
9fd7c5b343 test: set request timeout at request level 2023-11-15 17:42:31 -08:00
Rodney Rodriguez
5d39279ac0 bugfixes for ollama 2023-11-15 19:27:06 -06:00
Krrish Dholakia
bcea28e2e4 fix(utils): fixing exception mapping 2023-11-15 15:51:17 -08:00
Krrish Dholakia
f84db3ce14 feat(get_max_tokens): get max tokens for huggingface hub models 2023-11-15 15:25:40 -08:00
Krrish Dholakia
1a705bfbcb refactor(huggingface_restapi.py): moving async completion + streaming to real async calls 2023-11-15 15:14:21 -08:00
Krrish Dholakia
c8b73d90f1 fix(utils.py): fix langfuse integration 2023-11-15 14:05:40 -08:00
Krrish Dholakia
60d2dff74a fix(utils.py): check for none params 2023-11-15 13:39:09 -08:00
Krrish Dholakia
4348fd6435 fix(utils.py): azure streaming initial format 2023-11-15 13:30:08 -08:00
Krrish Dholakia
8c104e9c6a fix(azure.py-+-proxy_server.py): fix function calling response object + support router on proxy 2023-11-15 13:15:16 -08:00
Krrish Dholakia
e07bf0a8de fix(utils.py): await async function in client wrapper 2023-11-14 22:07:28 -08:00
Krrish Dholakia
59eaeba92a fix(router.py): enabling retrying with expo backoff (without tenacity) for router 2023-11-14 20:57:51 -08:00
ishaan-jaff
465f427465 (feat) debug POST logs 2023-11-14 18:16:45 -08:00
ishaan-jaff
c7780cbc40 (fix) debugging with POST request 2023-11-14 18:05:34 -08:00
ishaan-jaff
cb5669f679 (feat) improve logging of raw POST curl command 2023-11-14 17:54:09 -08:00
ishaan-jaff
7c317b78eb (feat) add ability to view POST requests from litellm.completion() 2023-11-14 17:27:20 -08:00
Krrish Dholakia
a7222f257c fix(main.py): keep client consistent across calls + exponential backoff retry on ratelimit errors 2023-11-14 16:26:05 -08:00
Krrish Dholakia
e9e86cac79 fix(palm.py): exception mapping bad requests / filtered responses 2023-11-14 11:53:13 -08:00
Jack Collins
223ec8933d Provide response to ServiceUnavailableError where needed 2023-11-13 21:20:40 -08:00
Krrish Dholakia
dd925d3de3 fix(utils.py): streaming 2023-11-13 18:15:14 -08:00
Krrish Dholakia
afc19caf0f test(utils.py): adding logging for azure streaming 2023-11-13 17:53:15 -08:00
Krrish Dholakia
661b78f4a3 test(utils.py): test logging 2023-11-13 17:41:45 -08:00
Krrish Dholakia
b722594792 test(utils.py): add logging and fix azure streaming 2023-11-13 17:24:13 -08:00
Krrish Dholakia
c489f41964 test(utils.py): additional logging 2023-11-13 17:13:41 -08:00
Krrish Dholakia
681da80e55 test(utils.py): additional logging 2023-11-13 17:06:24 -08:00
Krrish Dholakia
c1a0411186 test(utils.py): additional logging 2023-11-13 16:59:04 -08:00
Krrish Dholakia
05d720075b test(utils.py): adding more logging for streaming test 2023-11-13 16:54:16 -08:00
Krrish Dholakia
38ff412b9a fix(utils.py): fix response object mapping 2023-11-13 15:58:25 -08:00
David Christian
fe7e3ff038 added support for bedrock llama models 2023-11-13 15:41:21 -08:00
Krrish Dholakia
9d8f872f38 fix(promptlayer.py): fixing promptlayer logging integration 2023-11-13 15:04:15 -08:00
Krrish Dholakia
f20820fd00 fix(main.py): fix linting errors 2023-11-13 14:52:37 -08:00