Commit graph

1275 commits

Author SHA1 Message Date
WilliamEspegren
4bf7272885 rebuild chunks to openAI response
Rebuild the cunks, but does not include the "usage"
2023-09-17 13:07:54 +02:00
WilliamEspegren
684b11b456 add stream_chunk_builder function 2023-09-17 10:26:34 +02:00
Krrish Dholakia
d6dc825fa7 bump version and clean print statements 2023-09-16 20:59:24 -07:00
Krrish Dholakia
6430bdb636 fix cohere streaming 2023-09-16 19:36:41 -07:00
Krrish Dholakia
e7974d95d5 fix streaming formatting for non-openai models 2023-09-16 19:20:07 -07:00
Krrish Dholakia
d80fe482ce except custom openai proxy 2023-09-16 16:15:44 -07:00
Krrish Dholakia
925ec2ded9 fixes to testing 2023-09-16 11:40:01 -07:00
Krrish Dholakia
0a1f3af840 ensure streaming format is exactly the same as openai 2023-09-16 10:35:20 -07:00
ishaan-jaff
9ee3faed5b streaming for amazon titan bedrock 2023-09-16 09:57:16 -07:00
Toni Engelhardt
d181ab86b0 fix setting mock response 2023-09-16 17:03:32 +01:00
ishaan-jaff
75de6a3516 bump version 2023-09-15 14:15:09 -07:00
ishaan-jaff
595dd46dbc tests 2023-09-15 13:38:26 -07:00
ishaan-jaff
ee9e2af570 add new tg ai llm check 2023-09-15 09:29:46 -07:00
ishaan-jaff
486fe00ba6 add bedrock to main and init 2023-09-14 13:51:09 -07:00
Krrish Dholakia
22fb574009 move cohere to http endpoint 2023-09-14 11:17:38 -07:00
Krrish Dholakia
52a02c719a fixes to mock completion 2023-09-14 10:03:57 -07:00
Krish Dholakia
1e9097d997 Merge pull request #371 from promptmetheus/simplify-mock-logic
Simplify mock logic
2023-09-14 09:23:45 -07:00
Krrish Dholakia
bd01f70257 adding support for nlp cloud 2023-09-14 09:19:34 -07:00
ishaan-jaff
3d0b1428d0 pass api_type as an arg in ChatCompl 2023-09-14 08:48:59 -07:00
Toni Engelhardt
878dcb551d simplify mock logic
Adds shortcut for the mock_completion method.
2023-09-14 16:13:31 +01:00
Toni Engelhardt
90df061ffb add kwargs to mock_completion
Allows replacing `completion` with `mock_completion` without encountering errors.
2023-09-14 10:27:19 +01:00
Krrish Dholakia
3496bebdc5 map finish reason 2023-09-13 19:22:38 -07:00
Krrish Dholakia
f493779a97 huggingface conversational task support 2023-09-13 13:45:23 -07:00
ishaan-jaff
0bd5822a97 remove verify_access_key from main 2023-09-12 11:50:30 -07:00
Ishaan Jaff
5cec386ae2 Merge branch 'main' into main 2023-09-12 11:43:03 -07:00
Krrish Dholakia
3f6b194b71 raise better exception if llm provider isn't passed in or inferred 2023-09-12 11:28:50 -07:00
ishaan-jaff
56fee1b2a1 fix azure logging 2023-09-12 10:35:47 -07:00
ishaan-jaff
78b7bcb49d use api_version for azure completion call 2023-09-12 10:35:24 -07:00
William Espegren
09022ad61d util: verify_access_key
Verify that the user provided a valid openai token by creating a request to the openAI endpoint.
2023-09-12 18:48:23 +02:00
Ishaan Jaff
029c3cb3b1 Merge pull request #330 from Taik/main
Add support for overriding API type for Azure calls
2023-09-11 17:04:25 -07:00
Krrish Dholakia
cd12e95526 code cleanup 2023-09-11 16:32:17 -07:00
Krrish Dholakia
f944eaee4b mock responses for streaming 2023-09-11 16:30:29 -07:00
Thinh Nguyen
92a1402663 Add support for overriding API type for Azure calls 2023-09-11 16:28:00 -07:00
Krrish Dholakia
9f7c1d3885 add mock request to docs 2023-09-11 12:19:13 -07:00
ishaan-jaff
c4f58695d4 begin using litellm.api_key 2023-09-11 07:32:20 -07:00
Krrish Dholakia
1dd77ff176 update testing 2023-09-09 16:35:38 -07:00
Krrish Dholakia
4319b1011c add api manager 2023-09-09 15:55:38 -07:00
ishaan-jaff
3ed14cf8f8 add support for calling openai proxy using lite 2023-09-09 15:47:29 -07:00
ishaan-jaff
b8538fea34 olla upgrades, fix streaming, add non streaming resp 2023-09-09 14:07:13 -07:00
ishaan-jaff
990651c8c0 add cache params 2023-09-08 20:18:12 -07:00
ishaan-jaff
028fed5bf2 updating caching tests 2023-09-08 20:15:15 -07:00
ishaan-jaff
da87721a20 caching updates 2023-09-08 18:06:47 -07:00
ishaan-jaff
c1a7853da2 raise cohere, vertex, tenacity import errors 2023-09-08 15:26:09 -07:00
ishaan-jaff
847e9f47cc comments about using hosted vllm 2023-09-08 14:06:52 -07:00
ishaan-jaff
9eae0dac26 hosted vllm usage 2023-09-08 13:58:06 -07:00
ishaan-jaff
9bdd27df1d custom base with openai completion 2023-09-08 13:21:43 -07:00
ishaan-jaff
27b8145c56 async streaming generator 2023-09-07 13:53:40 -07:00
Krrish Dholakia
e6a65695eb fix linting issues 2023-09-06 20:43:59 -07:00
Krrish Dholakia
14fa57c185 batch completions for vllm now works too 2023-09-06 19:26:19 -07:00
Krrish Dholakia
7290a972e5 adding support for vllm 2023-09-06 18:07:44 -07:00