Krrish Dholakia
|
5b4ca42de6
|
docs(routing.md): add docs on using caching groups across deployments
|
2023-12-15 21:51:59 -08:00 |
|
Krrish Dholakia
|
84ad9f441e
|
feat(router.py): support caching groups
|
2023-12-15 21:45:51 -08:00 |
|
ishaan-jaff
|
287633887e
|
(feat) add ollama/llava
|
2023-12-16 10:35:27 +05:30 |
|
Krrish Dholakia
|
add153d110
|
fix(huggingface_restapi.py): add support for additional hf embedding formats
|
2023-12-15 21:02:41 -08:00 |
|
Krrish Dholakia
|
cab870f73a
|
fix(ollama.py): fix ollama async streaming for /completions calls
|
2023-12-15 09:28:32 -08:00 |
|
Krrish Dholakia
|
e45b491985
|
fix(main.py): log user for embedding calls
|
2023-12-14 15:14:37 -08:00 |
|
Krrish Dholakia
|
1608dd7e0b
|
fix(main.py): support async streaming for text completions endpoint
|
2023-12-14 13:56:32 -08:00 |
|
ishaan-jaff
|
7945664e61
|
(feat) add mistral api
|
2023-12-14 18:17:48 +05:30 |
|
ishaan-jaff
|
9526ee16c4
|
(fix) aembedding - don't pop aembedding out
|
2023-12-14 17:13:35 +05:30 |
|
Ishaan Jaff
|
cbc8cb08c8
|
Merge pull request #1112 from Undertone0809/add-cs-return-type-for-completion
refactor: add CustomStreamWrapper return type for completion
|
2023-12-14 16:52:11 +05:30 |
|
Krrish Dholakia
|
7b8851cce5
|
fix(ollama.py): fix async completion calls for ollama
|
2023-12-13 13:10:25 -08:00 |
|
Krrish Dholakia
|
69c29f8f86
|
fix(vertex_ai.py): add support for real async streaming + completion calls
|
2023-12-13 11:53:55 -08:00 |
|
Krrish Dholakia
|
07015843ac
|
fix(vertex_ai.py): support optional params + enable async calls for gemini
|
2023-12-13 11:01:23 -08:00 |
|
Krrish Dholakia
|
ef7a6e3ae1
|
feat(vertex_ai.py): adds support for gemini-pro on vertex ai
|
2023-12-13 10:26:30 -08:00 |
|
ishaan-jaff
|
86e626edab
|
(feat) pass vertex_ai/ as custom_llm_provider
|
2023-12-13 19:02:24 +03:00 |
|
zeeland
|
79ea466cf5
|
refactor: add CustomStreamWrapper return type for completion
|
2023-12-13 22:57:19 +08:00 |
|
Krrish Dholakia
|
8b07a6c046
|
fix(main.py): pass user_id + encoding_format for logging + to openai/azure
|
2023-12-12 15:46:44 -08:00 |
|
Krrish Dholakia
|
632d6e0bff
|
fix(utils.py): add more logging
|
2023-12-12 15:46:12 -08:00 |
|
Krrish Dholakia
|
d6669fe9e3
|
fix(utils.py): add more logging
|
2023-12-12 15:46:00 -08:00 |
|
ishaan-jaff
|
b6b88370ca
|
(fix) from re import T - junk import
|
2023-12-12 12:26:15 -08:00 |
|
Krrish Dholakia
|
9cf5ab468f
|
fix(router.py): deepcopy initial model list, don't mutate it
|
2023-12-12 09:54:06 -08:00 |
|
Krrish Dholakia
|
2c1c75fdf0
|
fix(ollama.py): enable parallel ollama completion calls
|
2023-12-11 23:18:37 -08:00 |
|
Krrish Dholakia
|
3b6099633c
|
test(test_custom_callback_router.py): add async azure testing for router
|
2023-12-11 16:40:35 -08:00 |
|
Krrish Dholakia
|
ad39afc0ad
|
test(test_custom_callback_input.py): embedding callback tests for azure, openai, bedrock
|
2023-12-11 15:32:46 -08:00 |
|
Krrish Dholakia
|
b09ecb986e
|
test(test_custom_callback_input.py): add bedrock testing
n
n
|
2023-12-11 13:00:01 -08:00 |
|
Krrish Dholakia
|
ea89a8a938
|
test(test_custom_callback_unit.py): adding unit tests for custom callbacks + fixing related bugs
|
2023-12-11 11:44:09 -08:00 |
|
ishaan-jaff
|
91f09365e8
|
(feat) access metadata in embedding kwargs()
|
2023-12-11 09:39:25 -08:00 |
|
ishaan-jaff
|
ee3c9d19a2
|
(feat) caching + stream - bedrock
|
2023-12-11 08:43:50 -08:00 |
|
Krish Dholakia
|
cc4a1d2603
|
Merge pull request #985 from estill01/patch-1
Enable setting default `model` value for `LiteLLM`, `Chat`, `Completions`
|
2023-12-09 13:59:00 -08:00 |
|
ishaan-jaff
|
c8b699c0aa
|
(feat) custom logger: async stream,assemble chunks
|
2023-12-09 10:10:48 -08:00 |
|
ishaan-jaff
|
14d57ec7de
|
(feat) proxy: log model_info + proxy_server request
|
2023-12-08 14:26:18 -08:00 |
|
ishaan-jaff
|
93b1df1c79
|
(feat) embedding - pass model_info, proxy_server request
|
2023-12-08 14:26:18 -08:00 |
|
ishaan-jaff
|
be94a8c478
|
(feat) pass model_info, proxy_server_request to callback
|
2023-12-08 14:26:18 -08:00 |
|
ishaan-jaff
|
762f28e4d7
|
(fix) make print_verbose non blocking
|
2023-12-07 17:31:32 -08:00 |
|
Krrish Dholakia
|
c1e95740b0
|
fix(bedrock.py): fix output format for cohere embeddings
|
2023-12-06 22:47:01 -08:00 |
|
ishaan-jaff
|
e3b24ec797
|
(feat) aembedding - add custom logging support
|
2023-12-06 19:09:06 -08:00 |
|
Krrish Dholakia
|
94f065f83c
|
feat(sagemaker.py): support huggingface embedding models
|
2023-12-06 11:41:38 -08:00 |
|
Krrish Dholakia
|
54d8a9df3f
|
fix(sagemaker.py): enable passing hf model name for prompt template
|
2023-12-05 16:31:59 -08:00 |
|
Krrish Dholakia
|
b4c78c7b9e
|
fix(utils.py): support sagemaker llama2 custom endpoints
|
2023-12-05 16:05:15 -08:00 |
|
Krrish Dholakia
|
69a4497550
|
fix(main.py): accept user in embedding()
|
2023-12-02 21:49:23 -08:00 |
|
estill01
|
737abbb0c1
|
fix
|
2023-12-03 05:37:57 +00:00 |
|
estill01
|
82fbbf67ca
|
Fix; persistent 'model' default value
|
2023-12-03 05:34:24 +00:00 |
|
Krrish Dholakia
|
bb4f82066a
|
fix(main.py): only send user if set
|
2023-12-02 20:36:30 -08:00 |
|
Krrish Dholakia
|
f72dd24ab9
|
fix(main.py): set user to none if not passed in
|
2023-12-02 20:08:25 -08:00 |
|
Krrish Dholakia
|
6c0eec4ff4
|
fix(main.py): fix pydantic warning for usage dict
|
2023-12-02 20:02:55 -08:00 |
|
estill01
|
56e95197c6
|
Enable setting default model value for Completions
add `model` arg to `Completions` class; if you provide a value, it will be used when you create new completions from an instance of the class.
|
2023-12-02 19:50:18 -08:00 |
|
Krrish Dholakia
|
82553e8aac
|
fix(azure.py): fix linting errors
|
2023-11-30 13:32:29 -08:00 |
|
Krrish Dholakia
|
4f07c8565a
|
feat(main.py): add support for azure-openai via cloudflare ai gateway
|
2023-11-30 13:19:49 -08:00 |
|
Krrish Dholakia
|
01c7e18f31
|
fix(utils.py): include system fingerprint in streaming response object
|
2023-11-30 08:45:52 -08:00 |
|
Krrish Dholakia
|
0d200cd8dc
|
feat(main.py): allow updating model cost via completion()
|
2023-11-29 20:14:39 -08:00 |
|