ishaan-jaff
|
f681f0f2b2
|
(feat) completion_cost - embeddings + raise Exception
|
2024-01-05 13:11:23 +05:30 |
|
ishaan-jaff
|
790dcff5e0
|
(feat) add xinference as an embedding provider
|
2024-01-02 15:32:26 +05:30 |
|
fatih
|
6566ebd815
|
update azure turbo namings
|
2024-01-01 13:03:08 +03:00 |
|
ishaan-jaff
|
037dcbbe10
|
(fix) use openai token counter for azure llms
|
2023-12-29 15:37:46 +05:30 |
|
ishaan-jaff
|
367e9913dc
|
(feat) v0 adding cloudflare
|
2023-12-29 09:32:29 +05:30 |
|
ishaan-jaff
|
95e6d2fbba
|
(feat) add voyage ai embeddings
|
2023-12-28 17:10:15 +05:30 |
|
Krrish Dholakia
|
e516cfe9f5
|
fix(utils.py): allow text completion input to be either model or engine
|
2023-12-27 17:24:16 +05:30 |
|
Krrish Dholakia
|
9ba520cc8b
|
fix(google_kms.py): support enums for key management system
|
2023-12-27 13:19:33 +05:30 |
|
Krrish Dholakia
|
2070a785a4
|
feat(utils.py): support google kms for secret management
https://github.com/BerriAI/litellm/issues/1235
|
2023-12-26 15:39:40 +05:30 |
|
ishaan-jaff
|
3e97a766a6
|
(feat) add ollama_chat as a provider
|
2023-12-25 23:04:17 +05:30 |
|
Krrish Dholakia
|
4905929de3
|
refactor: add black formatting
|
2023-12-25 14:11:20 +05:30 |
|
Krrish Dholakia
|
1262d89ab3
|
feat(gemini.py): add support for completion calls for gemini-pro (google ai studio)
|
2023-12-24 09:42:58 +05:30 |
|
Krrish Dholakia
|
13d088b72e
|
feat(main.py): add support for image generation endpoint
|
2023-12-16 21:07:29 -08:00 |
|
ishaan-jaff
|
0bf29a14e8
|
init vertex_vision_models
|
2023-12-16 18:37:00 +05:30 |
|
ishaan-jaff
|
9bdd6e73bb
|
(feat) proxy logs: dynamodb - set table name
|
2023-12-15 21:38:44 +05:30 |
|
ishaan-jaff
|
353433e5ce
|
(feat) add openai.NotFoundError
|
2023-12-15 10:18:02 +05:30 |
|
Krrish Dholakia
|
1608dd7e0b
|
fix(main.py): support async streaming for text completions endpoint
|
2023-12-14 13:56:32 -08:00 |
|
ishaan-jaff
|
9ee16bc962
|
(feat) caching - add supported call types
|
2023-12-14 22:27:14 +05:30 |
|
ishaan-jaff
|
c0cc78b943
|
(feat) mistral - add exception mapping
|
2023-12-14 18:57:39 +05:30 |
|
ishaan-jaff
|
7945664e61
|
(feat) add mistral api
|
2023-12-14 18:17:48 +05:30 |
|
Krrish Dholakia
|
8d688b6217
|
fix(utils.py): support caching for embedding + log cache hits
n
n
|
2023-12-13 18:37:30 -08:00 |
|
Krrish Dholakia
|
0f29cda8d9
|
test(test_amazing_vertex_completion.py): fix testing
|
2023-12-13 16:41:26 -08:00 |
|
Krrish Dholakia
|
ef7a6e3ae1
|
feat(vertex_ai.py): adds support for gemini-pro on vertex ai
|
2023-12-13 10:26:30 -08:00 |
|
Krrish Dholakia
|
4bf875d3ed
|
fix(router.py): fix least-busy routing
|
2023-12-08 20:29:49 -08:00 |
|
ishaan-jaff
|
ee70c4e822
|
(feat) router - add model_group_alias_map
|
2023-12-06 20:13:33 -08:00 |
|
ishaan-jaff
|
b3f039627e
|
(feat) litellm - add _async_failure_callback
|
2023-12-06 14:43:47 -08:00 |
|
Krrish Dholakia
|
d962d5d4c0
|
fix(bedrock.py): adding support for cohere embeddings
|
2023-12-06 13:25:18 -08:00 |
|
Frank Colson
|
95e5331090
|
Use litellm logging convention
|
2023-12-05 22:28:23 -07:00 |
|
Krrish Dholakia
|
e0ccb281d8
|
feat(utils.py): add async success callbacks for custom functions
|
2023-12-04 16:42:40 -08:00 |
|
Krrish Dholakia
|
eae5b3ce50
|
fix(__init__.py): fix linting error
|
2023-12-01 20:08:08 -08:00 |
|
Krrish Dholakia
|
328113a28e
|
fix(proxy_server.py): fix linting errors
|
2023-12-01 19:45:09 -08:00 |
|
Krrish Dholakia
|
368fee224e
|
feat: support for azure key vault
|
2023-12-01 19:36:06 -08:00 |
|
Krrish Dholakia
|
bb1267eb07
|
fix(router.py): fix exponential backoff to use retry-after if present in headers
|
2023-11-28 17:25:03 -08:00 |
|
ishaan-jaff
|
e9947be7b2
|
(fix) add timeout to __init__ litellm
|
2023-11-27 07:49:18 -08:00 |
|
Krrish Dholakia
|
e4f302a8e2
|
fix(proxy_server.py): expose a /health endpoint
|
2023-11-25 18:28:47 -08:00 |
|
Krrish Dholakia
|
2686894823
|
fix(router.py): fix retry logic
|
2023-11-24 13:27:44 -08:00 |
|
Krrish Dholakia
|
12dbdc4c15
|
docs(simple_proxy.md): add tutorial for doing fallbacks + retries + timeouts on the proxy
|
2023-11-24 12:20:38 -08:00 |
|
Krrish Dholakia
|
10fe16c965
|
fix(utils.py): add param mapping for perplexity, anyscale, deepinfra
n
n
|
2023-11-22 10:04:27 -08:00 |
|
Krrish Dholakia
|
b16646e584
|
refactor(proxy_server.py): using celery workers instead of rq for concurrency
|
2023-11-21 16:31:56 -08:00 |
|
Krrish Dholakia
|
a1f6b9b531
|
refactor(proxy_server.py): refactoring background rq worker
|
2023-11-21 13:47:09 -08:00 |
|
Krrish Dholakia
|
c6a4744947
|
feat(proxy_server.py): EXPERIMENTAL: adding queuing endpoints to openai proxy server
|
2023-11-21 12:06:23 -08:00 |
|
Krrish Dholakia
|
855964ed45
|
fix(utils.py): adding support for rules + mythomax/alpaca prompt template
|
2023-11-20 18:58:15 -08:00 |
|
Krrish Dholakia
|
e75c07f726
|
fix(init.py): exposing apiconnectionerror
|
2023-11-20 08:12:29 -08:00 |
|
ishaan-jaff
|
bd82559553
|
(v1.0+ breaking change) get_max_tokens -> return int
|
2023-11-17 10:38:50 -08:00 |
|
Krrish Dholakia
|
a23c0a2599
|
fix(openai.py): fix linting issues
|
2023-11-16 11:01:28 -08:00 |
|
Krrish Dholakia
|
fc8b54eb73
|
fix(openai.py): switch back to using requests instead of httpx
|
2023-11-15 18:25:21 -08:00 |
|
Krrish Dholakia
|
9fd7c5b343
|
test: set request timeout at request level
|
2023-11-15 17:42:31 -08:00 |
|
ishaan-jaff
|
2a81cc567d
|
(chore) removing bloat
|
2023-11-15 17:11:56 -08:00 |
|
Krrish Dholakia
|
1a705bfbcb
|
refactor(huggingface_restapi.py): moving async completion + streaming to real async calls
|
2023-11-15 15:14:21 -08:00 |
|
Krrish Dholakia
|
33d823a18d
|
docs(routing.md): add contributor s/o
|
2023-11-15 13:20:02 -08:00 |
|