Commit graph

564 commits

Author SHA1 Message Date
Krrish Dholakia
9bb2c7ee0f fix(utils.py): add param mapping for perplexity, anyscale, deepinfra
n

n
2023-11-22 10:04:27 -08:00
Krrish Dholakia
efc2bfe295 fix(utils.py): add response ms for async calls 2023-11-21 19:59:00 -08:00
Krrish Dholakia
c0e3b2ece9 fix(utils.py): fix pre call rules 2023-11-21 07:10:04 -08:00
Krrish Dholakia
8e1dcc540f fix(main.py): revert model alias map change 2023-11-20 21:07:52 -08:00
Krrish Dholakia
bdd9a933ad fix(utils.py): fix rules calling 2023-11-20 21:06:36 -08:00
Krrish Dholakia
d83c2b9ee8 fix(main.py): fix model alias map logic 2023-11-20 20:49:10 -08:00
Krrish Dholakia
35e5a757b0 fix(openai.py-+-azure.py): fix linting issues 2023-11-20 19:29:23 -08:00
Krrish Dholakia
c7e2cbd995 fix(utils.py): adding support for rules + mythomax/alpaca prompt template 2023-11-20 18:58:15 -08:00
ishaan-jaff
2f1180418b (fix) linting error 2023-11-20 18:32:43 -08:00
ishaan-jaff
372b4654c3 (fix) pydantic errors openai usage 2023-11-20 18:28:19 -08:00
ishaan-jaff
bb4ee4be0a (fix) completion - always map finish_reason 2023-11-20 17:24:16 -08:00
ishaan-jaff
756f356897 (fix) completion: max_retries using OpenAI client 2023-11-20 16:57:37 -08:00
Krrish Dholakia
7472be1529 fix(routing.py): update token usage on streaming 2023-11-20 14:19:25 -08:00
Krrish Dholakia
854b749535 fix(utils.py): expanding exception mapping coverage for vertex ai 2023-11-18 20:05:40 -08:00
ishaan-jaff
0fabd4caf8 (fix) streaming completion azure 2023-11-18 19:04:41 -08:00
ishaan-jaff
70dc8441f6 (fix) streaming ensure response obj is initialized 2023-11-18 17:31:58 -08:00
ishaan-jaff
e527a45ffc (feat) print_verbose Raw openai chunk 2023-11-18 17:12:49 -08:00
ishaan-jaff
8f402e04c9 (fix) streaming openai + function calling 2023-11-18 17:01:46 -08:00
ishaan-jaff
edf98cabae (fix) streaming + function / tool calling 2023-11-18 16:23:29 -08:00
Krrish Dholakia
cf0a9f591c fix(router.py): introducing usage-based-routing 2023-11-17 17:56:26 -08:00
ishaan-jaff
0ba90475c9 (feat) support parallel function calling 2023-11-17 15:51:27 -08:00
ishaan-jaff
698f47c226 (feat) improve logging - show model_call_details 2023-11-17 15:51:27 -08:00
Krrish Dholakia
02ed97d0b2 fix(acompletion): support client side timeouts + raise exceptions correctly for async calls 2023-11-17 15:39:47 -08:00
ishaan-jaff
ef8d82a54c (feat) completion: add response_format, seed, tools, tool_choice 2023-11-17 13:59:57 -08:00
ishaan-jaff
e9f6741b0b (v1.0+ breaking change) get_max_tokens -> return int 2023-11-17 10:38:50 -08:00
Krrish Dholakia
1e0560e4d2 fix(utils.py): improve exception mapping for vertex ai 2023-11-16 22:02:26 -08:00
Krrish Dholakia
d9123ea2e8 docs(routing.md): update tutorial on deploying router 2023-11-16 21:46:43 -08:00
Krrish Dholakia
48a508bab6 feat: global client for sync + async calls (openai + Azure only) 2023-11-16 14:44:13 -08:00
Krrish Dholakia
a6e9f147d3 fix(openai.py): handling extra headers 2023-11-16 12:48:21 -08:00
ishaan-jaff
56838ee815 (fix) bedrock meta llama optional params 2023-11-16 12:38:27 -08:00
ishaan-jaff
55a054f3f6 (fix) only decode chunk when it's not a str 2023-11-16 12:24:31 -08:00
Krrish Dholakia
e54056f0ed fix(azure.py): use openai client sdk for handling sync+async calling 2023-11-16 12:08:12 -08:00
ishaan-jaff
aa84ca04d8 (fix) HF api + streaming 2023-11-16 11:59:56 -08:00
ishaan-jaff
fb2d398d2c (fix) langfuse logging + openai streaming when chunk = [DONE} 2023-11-16 10:45:35 -08:00
Krrish Dholakia
9c7cc84eb0 fix(openai.py): supporting openai client sdk for handling sync + async calls (incl. for openai-compatible apis) 2023-11-16 10:35:03 -08:00
Ishaan Jaff
da9a0ab928 Merge pull request #811 from dchristian3188/bedrock-llama
Bedrock llama
2023-11-16 07:57:50 -08:00
Ishaan Jaff
d6d0cbd63c Merge pull request #826 from rodneyxr/ollama-fixes
Fix typo for initial_prompt_value and too many values to unpack error
2023-11-16 07:55:53 -08:00
David Christian
461115330b updated utils for bedrock.meta streaming 2023-11-16 07:12:27 -08:00
Krrish Dholakia
ef4e5b9636 test: set request timeout at request level 2023-11-15 17:42:31 -08:00
Rodney Rodriguez
f2d8bfd40d bugfixes for ollama 2023-11-15 19:27:06 -06:00
Krrish Dholakia
b42cf80585 fix(utils): fixing exception mapping 2023-11-15 15:51:17 -08:00
Krrish Dholakia
0ede0e836e feat(get_max_tokens): get max tokens for huggingface hub models 2023-11-15 15:25:40 -08:00
Krrish Dholakia
e35ce15a89 refactor(huggingface_restapi.py): moving async completion + streaming to real async calls 2023-11-15 15:14:21 -08:00
Krrish Dholakia
04ce14e404 fix(utils.py): fix langfuse integration 2023-11-15 14:05:40 -08:00
Krrish Dholakia
e324388520 fix(utils.py): check for none params 2023-11-15 13:39:09 -08:00
Krrish Dholakia
8eaa1eb37f fix(utils.py): azure streaming initial format 2023-11-15 13:30:08 -08:00
Krrish Dholakia
e5929f2f7e fix(azure.py-+-proxy_server.py): fix function calling response object + support router on proxy 2023-11-15 13:15:16 -08:00
Krrish Dholakia
29a0c29eb3 fix(utils.py): await async function in client wrapper 2023-11-14 22:07:28 -08:00
Krrish Dholakia
0f6713993d fix(router.py): enabling retrying with expo backoff (without tenacity) for router 2023-11-14 20:57:51 -08:00
ishaan-jaff
9585856b9f (feat) debug POST logs 2023-11-14 18:16:45 -08:00