Krish Dholakia
|
1e99ebfec3
|
Merge pull request #845 from canada4663/upstream-main
Added support for multiple embeddings via Bedrock
|
2023-11-21 14:00:06 -08:00 |
|
Krrish Dholakia
|
aad17a5479
|
fix(huggingface_restapi.pyu): fix linting errors
|
2023-11-21 10:05:35 -08:00 |
|
Krrish Dholakia
|
a1c342c3ad
|
feat(factory.py): add support for anthropic system prompts for claude 2.1
|
2023-11-21 09:57:26 -08:00 |
|
Krrish Dholakia
|
53d71c229f
|
fix(huggingface_restapi.py): fixing formatting
|
2023-11-21 09:57:26 -08:00 |
|
Krrish Dholakia
|
c48e775908
|
fix(huggingface_restapi.py): fix huggingface response format
|
2023-11-21 09:57:26 -08:00 |
|
Krrish Dholakia
|
c425fb7eeb
|
fix(huggingface_restapi.py): handle generate text output
|
2023-11-21 09:57:26 -08:00 |
|
Krrish Dholakia
|
35e5a757b0
|
fix(openai.py-+-azure.py): fix linting issues
|
2023-11-20 19:29:23 -08:00 |
|
Krrish Dholakia
|
c7e2cbd995
|
fix(utils.py): adding support for rules + mythomax/alpaca prompt template
|
2023-11-20 18:58:15 -08:00 |
|
ishaan-jaff
|
7bc28f3b1c
|
(fix) pydantic errors with response.time
|
2023-11-20 18:28:19 -08:00 |
|
ishaan-jaff
|
b1c0326cc5
|
(fix) linting
|
2023-11-20 17:00:20 -08:00 |
|
ishaan-jaff
|
756f356897
|
(fix) completion: max_retries using OpenAI client
|
2023-11-20 16:57:37 -08:00 |
|
canada4663
|
5108b00688
|
bedrock embedding changes pre-testing
|
2023-11-18 21:00:06 -08:00 |
|
ishaan-jaff
|
b494990045
|
(feat) openai improve logging post_call
|
2023-11-17 15:51:27 -08:00 |
|
Krrish Dholakia
|
aa6e6865cd
|
fix(main.py): fix linting issue
|
2023-11-17 15:45:00 -08:00 |
|
Krrish Dholakia
|
02ed97d0b2
|
fix(acompletion): support client side timeouts + raise exceptions correctly for async calls
|
2023-11-17 15:39:47 -08:00 |
|
Krrish Dholakia
|
81f6732fe5
|
fix(factory.py): for ollama models check if it's instruct or not before applying prompt template
|
2023-11-16 15:45:08 -08:00 |
|
Krrish Dholakia
|
48a508bab6
|
feat: global client for sync + async calls (openai + Azure only)
|
2023-11-16 14:44:13 -08:00 |
|
Krrish Dholakia
|
7a71385e53
|
fix(openai.py): fix linting issues
|
2023-11-16 12:57:53 -08:00 |
|
Krrish Dholakia
|
a6e9f147d3
|
fix(openai.py): handling extra headers
|
2023-11-16 12:48:21 -08:00 |
|
Krrish Dholakia
|
9c2e74b6ea
|
fix(azure.py): fix linting errors
|
2023-11-16 12:15:50 -08:00 |
|
Krrish Dholakia
|
e54056f0ed
|
fix(azure.py): use openai client sdk for handling sync+async calling
|
2023-11-16 12:08:12 -08:00 |
|
Krrish Dholakia
|
f14bd24b46
|
fix(openai.py): fix linting issues
|
2023-11-16 11:01:28 -08:00 |
|
Krrish Dholakia
|
9c7cc84eb0
|
fix(openai.py): supporting openai client sdk for handling sync + async calls (incl. for openai-compatible apis)
|
2023-11-16 10:35:03 -08:00 |
|
Ishaan Jaff
|
da9a0ab928
|
Merge pull request #811 from dchristian3188/bedrock-llama
Bedrock llama
|
2023-11-16 07:57:50 -08:00 |
|
Krrish Dholakia
|
a80c5841b2
|
fix(openai.py): using openai sdk for completion calls
|
2023-11-15 20:25:27 -08:00 |
|
Krrish Dholakia
|
6aef9d2db3
|
perf(openai.py): extending request timeout
|
2023-11-15 18:45:45 -08:00 |
|
Krrish Dholakia
|
41e1fcba96
|
fix(openai.py): fix linting errors
|
2023-11-15 18:33:01 -08:00 |
|
Krrish Dholakia
|
1c219106ba
|
fix(openai.py): move chat completions post call to requests
|
2023-11-15 18:29:07 -08:00 |
|
Krrish Dholakia
|
4bd471644e
|
fix(openai.py): switch back to using requests instead of httpx
|
2023-11-15 18:25:21 -08:00 |
|
Krrish Dholakia
|
081400ecd4
|
fix(openai.py): fix timeout errors
|
2023-11-15 18:03:14 -08:00 |
|
Krrish Dholakia
|
f2020c2284
|
test: cleanup testing
|
2023-11-15 17:57:14 -08:00 |
|
Krrish Dholakia
|
ef4e5b9636
|
test: set request timeout at request level
|
2023-11-15 17:42:31 -08:00 |
|
Krrish Dholakia
|
c4c24dafb8
|
fix(openai-+-azure): async calling
|
2023-11-15 17:03:02 -08:00 |
|
Krrish Dholakia
|
e28b9534e4
|
fix(huggingface_restapi.py): async implementation
|
2023-11-15 16:54:15 -08:00 |
|
Krrish Dholakia
|
b42cf80585
|
fix(utils): fixing exception mapping
|
2023-11-15 15:51:17 -08:00 |
|
Krrish Dholakia
|
6d1b8945dd
|
fix(huggingface_restapi.py): fix linting errors
|
2023-11-15 15:34:21 -08:00 |
|
Krrish Dholakia
|
e35ce15a89
|
refactor(huggingface_restapi.py): moving async completion + streaming to real async calls
|
2023-11-15 15:14:21 -08:00 |
|
Krrish Dholakia
|
e5929f2f7e
|
fix(azure.py-+-proxy_server.py): fix function calling response object + support router on proxy
|
2023-11-15 13:15:16 -08:00 |
|
Krrish Dholakia
|
c3f81b9e34
|
fix: fix linting errors
|
2023-11-14 21:51:39 -08:00 |
|
Krrish Dholakia
|
0f6713993d
|
fix(router.py): enabling retrying with expo backoff (without tenacity) for router
|
2023-11-14 20:57:51 -08:00 |
|
Krrish Dholakia
|
5d58bb9cd0
|
fix(main.py): misrouting ollama models to nlp cloud
|
2023-11-14 18:55:08 -08:00 |
|
ishaan-jaff
|
84e931ce0f
|
(feat) completion debug view HF POST request
|
2023-11-14 17:57:41 -08:00 |
|
ishaan-jaff
|
d25d4d26bd
|
(feat) debug ollama POST request
|
2023-11-14 17:53:48 -08:00 |
|
ishaan-jaff
|
c7fbbe8764
|
(feat) add ability to view POST requests from litellm.completion()
|
2023-11-14 17:27:20 -08:00 |
|
Krrish Dholakia
|
9b582b2c85
|
fix(main.py): keep client consistent across calls + exponential backoff retry on ratelimit errors
|
2023-11-14 16:26:05 -08:00 |
|
ishaan-jaff
|
e1ea6f2cc7
|
(fix) AI21 exception mapping - raise error when status !=200
|
2023-11-14 15:01:22 -08:00 |
|
ishaan-jaff
|
1bb99af134
|
(fix) azure_ad_token
|
2023-11-14 12:20:14 -08:00 |
|
ishaan-jaff
|
90e7f28522
|
(feat) add azure_ad_token
|
2023-11-14 12:12:14 -08:00 |
|
Krrish Dholakia
|
526eb99ade
|
fix(palm.py): exception mapping bad requests / filtered responses
|
2023-11-14 11:53:13 -08:00 |
|
Krrish Dholakia
|
b572e9fe3a
|
test(utils.py): add logging and fix azure streaming
|
2023-11-13 17:24:13 -08:00 |
|