ishaan-jaff
b15682bc1f
(feat) set default openrouter configs
2023-12-18 08:55:51 +05:30
Joel Eriksson
e214e6ab47
Fix bug when iterating over lines in ollama response
...
async for line in resp.content.iter_any() will return
incomplete lines when the lines are long, and that
results in an exception being thrown by json.loads()
when it tries to parse the incomplete JSON
The default behavior of the stream reader for aiohttp
response objects is to iterate over lines, so just
removing .iter_any() fixes the bug
2023-12-17 20:23:26 +02:00
Krrish Dholakia
a3c7a340a5
fix(ollama.py): fix sync ollama streaming
2023-12-16 21:23:21 -08:00
Krrish Dholakia
13d088b72e
feat(main.py): add support for image generation endpoint
2023-12-16 21:07:29 -08:00
Krrish Dholakia
7c2fad2d57
fix(azure.py): fix azure streaming logging
2023-12-16 18:06:08 -08:00
Krrish Dholakia
3923c389fd
build(Dockerfile): fixing build requirements
2023-12-16 17:52:30 -08:00
Krrish Dholakia
4e828ff541
fix(health.md): add background health check details to docs
2023-12-16 10:31:59 -08:00
ishaan-jaff
5ee6b87f2e
(fix) vertexai - gemini
2023-12-16 22:15:41 +05:30
ishaan-jaff
764f31c970
(feat) add async, async+stream for gemini
2023-12-16 18:58:12 +05:30
ishaan-jaff
efe8b75200
(fix) use litellm.vertex_vision_models
2023-12-16 18:39:40 +05:30
ishaan-jaff
774a725ccb
(feat) add vertex ai gemini-pro-vision
2023-12-16 18:31:03 +05:30
ishaan-jaff
20b5505476
(feat) show POST request for HF embeddings
2023-12-16 13:09:49 +05:30
ishaan-jaff
287633887e
(feat) add ollama/llava
2023-12-16 10:35:27 +05:30
Krrish Dholakia
add153d110
fix(huggingface_restapi.py): add support for additional hf embedding formats
2023-12-15 21:02:41 -08:00
Krrish Dholakia
4791dda66f
feat(proxy_server.py): enable infinite retries on rate limited requests
2023-12-15 20:03:41 -08:00
Krrish Dholakia
e5268fa6bc
fix(router.py): support openai-compatible endpoints
2023-12-15 14:47:54 -08:00
Krrish Dholakia
edb88e31e4
fix(together_ai.py): return empty tgai responses
2023-12-15 10:46:35 -08:00
Krrish Dholakia
a09a6f24a4
fix(together_ai.py): additional logging for together ai encoding prompt
2023-12-15 10:39:23 -08:00
Krrish Dholakia
cab870f73a
fix(ollama.py): fix ollama async streaming for /completions calls
2023-12-15 09:28:32 -08:00
ishaan-jaff
85a3c67574
(feat) - acompletion, correct exception mapping
2023-12-15 08:28:12 +05:30
Krrish Dholakia
804d58eb20
bump: version 1.14.4 → 1.14.5.dev1
2023-12-14 15:23:52 -08:00
Krrish Dholakia
1608dd7e0b
fix(main.py): support async streaming for text completions endpoint
2023-12-14 13:56:32 -08:00
Krish Dholakia
a6e78497b5
Merge pull request #1122 from emsi/main
...
Fix #1119 , no content when streaming.
2023-12-14 10:01:00 -08:00
Krrish Dholakia
e678009695
fix(vertex_ai.py): add exception mapping for acompletion calls
2023-12-13 16:35:50 -08:00
Krrish Dholakia
7b8851cce5
fix(ollama.py): fix async completion calls for ollama
2023-12-13 13:10:25 -08:00
Mariusz Woloszyn
1feb6317f6
Fix #1119 , no content when streaming.
2023-12-13 21:42:35 +01:00
Krrish Dholakia
75bcb37cb2
fix(factory.py): fix tgai rendering template
2023-12-13 12:27:31 -08:00
Krrish Dholakia
69c29f8f86
fix(vertex_ai.py): add support for real async streaming + completion calls
2023-12-13 11:53:55 -08:00
Krrish Dholakia
07015843ac
fix(vertex_ai.py): support optional params + enable async calls for gemini
2023-12-13 11:01:23 -08:00
Krrish Dholakia
ef7a6e3ae1
feat(vertex_ai.py): adds support for gemini-pro on vertex ai
2023-12-13 10:26:30 -08:00
ishaan-jaff
86e626edab
(feat) pass vertex_ai/ as custom_llm_provider
2023-12-13 19:02:24 +03:00
Krrish Dholakia
a64bd2ca1e
fix(sagemaker.py): filter out templated prompt if in model response
2023-12-13 07:43:33 -08:00
Krrish Dholakia
82d28a8825
fix(factory.py): safely fail prompt template get requests for together ai
2023-12-12 17:28:22 -08:00
Krrish Dholakia
8e7116635f
fix(ollama.py): add support for async streaming
2023-12-12 16:44:20 -08:00
Krrish Dholakia
8b07a6c046
fix(main.py): pass user_id + encoding_format for logging + to openai/azure
2023-12-12 15:46:44 -08:00
ishaan-jaff
a251a52717
(chore) remove junk tkinter import
2023-12-12 13:54:50 -08:00
ishaan-jaff
99b48eff17
(fix) tkinter import
2023-12-12 12:18:25 -08:00
Krrish Dholakia
9cf5ab468f
fix(router.py): deepcopy initial model list, don't mutate it
2023-12-12 09:54:06 -08:00
Krrish Dholakia
2c1c75fdf0
fix(ollama.py): enable parallel ollama completion calls
2023-12-11 23:18:37 -08:00
Krrish Dholakia
ad39afc0ad
test(test_custom_callback_input.py): embedding callback tests for azure, openai, bedrock
2023-12-11 15:32:46 -08:00
Krrish Dholakia
b09ecb986e
test(test_custom_callback_input.py): add bedrock testing
...
n
n
2023-12-11 13:00:01 -08:00
Krrish Dholakia
ea89a8a938
test(test_custom_callback_unit.py): adding unit tests for custom callbacks + fixing related bugs
2023-12-11 11:44:09 -08:00
Krish Dholakia
4ffe6a4296
Merge pull request #1054 from James4Ever0/main
...
Update factory.py to fix issue when calling from write-the -> langchain -> litellm served ollama
2023-12-11 07:18:02 -08:00
Krish Dholakia
bbbc5db104
Merge pull request #1080 from nbaldwin98/fixing-replicate-sys-prompt
...
fix replicate system prompt: forgot to add **optional_params to input data
2023-12-11 07:11:52 -08:00
chabala98
c5ce11541b
adfix: ded **optional_params in input data when system prompt is available (allows to papass other optional params apart from sys prompt)
2023-12-11 14:42:05 +01:00
James4Ever0
69fc2694bb
Update factory.py
...
Fixing issue when calling from write-the -> langchain -> litellm served ollama
2023-12-08 02:58:28 +08:00
ishaan-jaff
d2eee342fb
(feat) vertex ai - better debugging
2023-12-07 09:38:37 -08:00
Krrish Dholakia
c1e95740b0
fix(bedrock.py): fix output format for cohere embeddings
2023-12-06 22:47:01 -08:00
Krrish Dholakia
ac7d0a1632
fix(together_ai.py): improve together ai custom prompt templating
2023-12-06 19:34:49 -08:00
Krrish Dholakia
fff0228c20
fix(factory.py): support togethercomputer codellama pt
2023-12-06 19:02:58 -08:00