Commit graph

388 commits

Author SHA1 Message Date
Krrish Dholakia
4e828ff541 fix(health.md): add background health check details to docs 2023-12-16 10:31:59 -08:00
ishaan-jaff
5ee6b87f2e (fix) vertexai - gemini 2023-12-16 22:15:41 +05:30
ishaan-jaff
764f31c970 (feat) add async, async+stream for gemini 2023-12-16 18:58:12 +05:30
ishaan-jaff
efe8b75200 (fix) use litellm.vertex_vision_models 2023-12-16 18:39:40 +05:30
ishaan-jaff
774a725ccb (feat) add vertex ai gemini-pro-vision 2023-12-16 18:31:03 +05:30
ishaan-jaff
20b5505476 (feat) show POST request for HF embeddings 2023-12-16 13:09:49 +05:30
ishaan-jaff
287633887e (feat) add ollama/llava 2023-12-16 10:35:27 +05:30
Krrish Dholakia
add153d110 fix(huggingface_restapi.py): add support for additional hf embedding formats 2023-12-15 21:02:41 -08:00
Krrish Dholakia
4791dda66f feat(proxy_server.py): enable infinite retries on rate limited requests 2023-12-15 20:03:41 -08:00
Krrish Dholakia
e5268fa6bc fix(router.py): support openai-compatible endpoints 2023-12-15 14:47:54 -08:00
Krrish Dholakia
edb88e31e4 fix(together_ai.py): return empty tgai responses 2023-12-15 10:46:35 -08:00
Krrish Dholakia
a09a6f24a4 fix(together_ai.py): additional logging for together ai encoding prompt 2023-12-15 10:39:23 -08:00
Krrish Dholakia
cab870f73a fix(ollama.py): fix ollama async streaming for /completions calls 2023-12-15 09:28:32 -08:00
ishaan-jaff
85a3c67574 (feat) - acompletion, correct exception mapping 2023-12-15 08:28:12 +05:30
Krrish Dholakia
804d58eb20 bump: version 1.14.4 → 1.14.5.dev1 2023-12-14 15:23:52 -08:00
Krrish Dholakia
1608dd7e0b fix(main.py): support async streaming for text completions endpoint 2023-12-14 13:56:32 -08:00
Krish Dholakia
a6e78497b5
Merge pull request #1122 from emsi/main
Fix #1119, no content when streaming.
2023-12-14 10:01:00 -08:00
Krrish Dholakia
e678009695 fix(vertex_ai.py): add exception mapping for acompletion calls 2023-12-13 16:35:50 -08:00
Krrish Dholakia
7b8851cce5 fix(ollama.py): fix async completion calls for ollama 2023-12-13 13:10:25 -08:00
Mariusz Woloszyn
1feb6317f6 Fix #1119, no content when streaming. 2023-12-13 21:42:35 +01:00
Krrish Dholakia
75bcb37cb2 fix(factory.py): fix tgai rendering template 2023-12-13 12:27:31 -08:00
Krrish Dholakia
69c29f8f86 fix(vertex_ai.py): add support for real async streaming + completion calls 2023-12-13 11:53:55 -08:00
Krrish Dholakia
07015843ac fix(vertex_ai.py): support optional params + enable async calls for gemini 2023-12-13 11:01:23 -08:00
Krrish Dholakia
ef7a6e3ae1 feat(vertex_ai.py): adds support for gemini-pro on vertex ai 2023-12-13 10:26:30 -08:00
ishaan-jaff
86e626edab (feat) pass vertex_ai/ as custom_llm_provider 2023-12-13 19:02:24 +03:00
Krrish Dholakia
a64bd2ca1e fix(sagemaker.py): filter out templated prompt if in model response 2023-12-13 07:43:33 -08:00
Krrish Dholakia
82d28a8825 fix(factory.py): safely fail prompt template get requests for together ai 2023-12-12 17:28:22 -08:00
Krrish Dholakia
8e7116635f fix(ollama.py): add support for async streaming 2023-12-12 16:44:20 -08:00
Krrish Dholakia
8b07a6c046 fix(main.py): pass user_id + encoding_format for logging + to openai/azure 2023-12-12 15:46:44 -08:00
ishaan-jaff
a251a52717 (chore) remove junk tkinter import 2023-12-12 13:54:50 -08:00
ishaan-jaff
99b48eff17 (fix) tkinter import 2023-12-12 12:18:25 -08:00
Krrish Dholakia
9cf5ab468f fix(router.py): deepcopy initial model list, don't mutate it 2023-12-12 09:54:06 -08:00
Krrish Dholakia
2c1c75fdf0 fix(ollama.py): enable parallel ollama completion calls 2023-12-11 23:18:37 -08:00
Krrish Dholakia
ad39afc0ad test(test_custom_callback_input.py): embedding callback tests for azure, openai, bedrock 2023-12-11 15:32:46 -08:00
Krrish Dholakia
b09ecb986e test(test_custom_callback_input.py): add bedrock testing
n

n
2023-12-11 13:00:01 -08:00
Krrish Dholakia
ea89a8a938 test(test_custom_callback_unit.py): adding unit tests for custom callbacks + fixing related bugs 2023-12-11 11:44:09 -08:00
Krish Dholakia
4ffe6a4296
Merge pull request #1054 from James4Ever0/main
Update factory.py to fix issue when calling from write-the -> langchain -> litellm served ollama
2023-12-11 07:18:02 -08:00
Krish Dholakia
bbbc5db104
Merge pull request #1080 from nbaldwin98/fixing-replicate-sys-prompt
fix replicate system prompt: forgot to add **optional_params to input data
2023-12-11 07:11:52 -08:00
chabala98
c5ce11541b adfix: ded **optional_params in input data when system prompt is available (allows to papass other optional params apart from sys prompt) 2023-12-11 14:42:05 +01:00
James4Ever0
69fc2694bb
Update factory.py
Fixing issue when calling from write-the -> langchain -> litellm served ollama
2023-12-08 02:58:28 +08:00
ishaan-jaff
d2eee342fb (feat) vertex ai - better debugging 2023-12-07 09:38:37 -08:00
Krrish Dholakia
c1e95740b0 fix(bedrock.py): fix output format for cohere embeddings 2023-12-06 22:47:01 -08:00
Krrish Dholakia
ac7d0a1632 fix(together_ai.py): improve together ai custom prompt templating 2023-12-06 19:34:49 -08:00
Krrish Dholakia
fff0228c20 fix(factory.py): support togethercomputer codellama pt 2023-12-06 19:02:58 -08:00
Krrish Dholakia
0295509b3b fix(factory.py): fix claude 2.1 prompt template to handle system, assistant, user prompt 2023-12-06 18:02:06 -08:00
Krrish Dholakia
f1c1ec8523 fix(bedrock.py): fix embeddings call 2023-12-06 14:16:00 -08:00
Krrish Dholakia
b24c9b4cbf refactor: fix linting 2023-12-06 13:27:40 -08:00
Krrish Dholakia
d962d5d4c0 fix(bedrock.py): adding support for cohere embeddings 2023-12-06 13:25:18 -08:00
Krrish Dholakia
102de97960 refactor: fix linting errors 2023-12-06 11:46:15 -08:00
Krrish Dholakia
94f065f83c feat(sagemaker.py): support huggingface embedding models 2023-12-06 11:41:38 -08:00