ishaan-jaff
|
54b6a767f4
|
(docs) using proxy
|
2023-11-17 10:19:01 -08:00 |
|
ishaan-jaff
|
79e93ee7aa
|
(docs) proxy - aws sagemaker jumpstart
|
2023-11-17 10:04:35 -08:00 |
|
Krish Dholakia
|
107edc631c
|
Update README.md
|
2023-11-17 09:39:06 -08:00 |
|
Krish Dholakia
|
85b987741f
|
Update README.md
|
2023-11-17 09:38:26 -08:00 |
|
Ishaan Jaff
|
1a09b93214
|
Merge pull request #832 from BerriAI/dependabot/npm_and_yarn/docs/my-website/sharp-0.32.6
build(deps): bump sharp from 0.32.5 to 0.32.6 in /docs/my-website
|
2023-11-17 09:05:34 -08:00 |
|
Krrish Dholakia
|
3d45d8a58c
|
test: load test router
|
2023-11-17 08:23:44 -08:00 |
|
ishaan-jaff
|
c7aba49d83
|
(ci/cd) re run pipeline
|
2023-11-17 08:07:02 -08:00 |
|
Krrish Dholakia
|
a27c2961dc
|
docs(migration.md): new base case exception - apiconnectionerror
|
2023-11-16 22:04:14 -08:00 |
|
Krrish Dholakia
|
9bd1f4ebd0
|
fix(utils.py): improve exception mapping for vertex ai
|
2023-11-16 22:02:26 -08:00 |
|
ishaan-jaff
|
3e2750ffb9
|
(ci/cd) re run pipeline
|
2023-11-16 21:55:10 -08:00 |
|
Krrish Dholakia
|
75ef1d7eb4
|
fix(router.py): check if async response is coroutine
|
2023-11-16 21:53:35 -08:00 |
|
Krrish Dholakia
|
7456c26940
|
docs(routing.md): update tutorial on deploying router
|
2023-11-16 21:46:43 -08:00 |
|
Ishaan Jaff
|
61705f3467
|
(ci/cd) run again
|
2023-11-16 21:29:18 -08:00 |
|
Krrish Dholakia
|
7ef1014e59
|
fix(factory.py): for ollama models check if it's instruct or not before applying prompt template
|
2023-11-16 15:45:08 -08:00 |
|
ishaan-jaff
|
25b2bc6da9
|
(test) add --debug to cli tool
|
2023-11-16 14:46:26 -08:00 |
|
Krrish Dholakia
|
51bf637656
|
feat: global client for sync + async calls (openai + Azure only)
|
2023-11-16 14:44:13 -08:00 |
|
ishaan-jaff
|
5fd4376802
|
(docs) lm harness
|
2023-11-16 14:30:33 -08:00 |
|
ishaan-jaff
|
8d0f4bfc74
|
(docs) lm harness
|
2023-11-16 14:21:37 -08:00 |
|
Krrish Dholakia
|
d7f7694848
|
fix(openai.py): fix linting issues
|
2023-11-16 12:57:53 -08:00 |
|
Krrish Dholakia
|
a94c09c13c
|
fix(openai.py): handling extra headers
|
2023-11-16 12:48:21 -08:00 |
|
ishaan-jaff
|
9e072f87bd
|
(fix) bedrock meta llama optional params
|
2023-11-16 12:38:27 -08:00 |
|
ishaan-jaff
|
23d560071b
|
(linting) fix
|
2023-11-16 12:33:03 -08:00 |
|
ishaan-jaff
|
f8af5e0155
|
(fix) linting
|
2023-11-16 12:25:46 -08:00 |
|
ishaan-jaff
|
2dc411fdb3
|
(test) hf streaming
|
2023-11-16 12:24:31 -08:00 |
|
ishaan-jaff
|
04971674b4
|
(fix) only decode chunk when it's not a str
|
2023-11-16 12:24:31 -08:00 |
|
Krrish Dholakia
|
f582189cea
|
test(loadtest_router.py): commenting out of ci/cd
|
2023-11-16 12:17:25 -08:00 |
|
Krrish Dholakia
|
f99a161d98
|
fix(azure.py): fix linting errors
|
2023-11-16 12:15:50 -08:00 |
|
Krrish Dholakia
|
63104f4194
|
bump: version 1.0.3 → 1.0.4
|
2023-11-16 12:08:16 -08:00 |
|
Krrish Dholakia
|
bf0f8b824c
|
fix(azure.py): use openai client sdk for handling sync+async calling
|
2023-11-16 12:08:12 -08:00 |
|
ishaan-jaff
|
3285113d2d
|
(test) regular hf tests
|
2023-11-16 12:00:49 -08:00 |
|
ishaan-jaff
|
da8c2f4a4a
|
(fix) HF api + streaming
|
2023-11-16 11:59:56 -08:00 |
|
ishaan-jaff
|
a1cecbafe6
|
(fix) linting
|
2023-11-16 11:44:26 -08:00 |
|
ishaan-jaff
|
baf4e83738
|
(test) text_completion
|
2023-11-16 11:37:46 -08:00 |
|
ishaan-jaff
|
77468e0a70
|
(feat) text_completion add rules on when to use engine & model together
|
2023-11-16 11:37:31 -08:00 |
|
ishaan-jaff
|
a1223e1f55
|
(test) proxy cli
|
2023-11-16 11:19:09 -08:00 |
|
ishaan-jaff
|
b607e5eb2a
|
(test) proxy cli test
|
2023-11-16 11:13:39 -08:00 |
|
Krrish Dholakia
|
a23c0a2599
|
fix(openai.py): fix linting issues
|
2023-11-16 11:01:28 -08:00 |
|
dependabot[bot]
|
6ead640637
|
build(deps): bump sharp from 0.32.5 to 0.32.6 in /docs/my-website
Bumps [sharp](https://github.com/lovell/sharp) from 0.32.5 to 0.32.6.
- [Release notes](https://github.com/lovell/sharp/releases)
- [Changelog](https://github.com/lovell/sharp/blob/main/docs/changelog.md)
- [Commits](https://github.com/lovell/sharp/compare/v0.32.5...v0.32.6)
---
updated-dependencies:
- dependency-name: sharp
dependency-type: direct:production
...
Signed-off-by: dependabot[bot] <support@github.com>
|
2023-11-16 18:59:28 +00:00 |
|
ishaan-jaff
|
6b14c8d2de
|
(test) langfuse logger
|
2023-11-16 10:45:35 -08:00 |
|
ishaan-jaff
|
c84fe1af5e
|
(fix) langfuse logging + openai streaming when chunk = [DONE}
|
2023-11-16 10:45:35 -08:00 |
|
Krrish Dholakia
|
bb51216846
|
fix(openai.py): supporting openai client sdk for handling sync + async calls (incl. for openai-compatible apis)
|
2023-11-16 10:35:03 -08:00 |
|
ishaan-jaff
|
b8c64f16cd
|
(chore) linting fix
|
2023-11-16 09:57:40 -08:00 |
|
ishaan-jaff
|
d9fb4d494c
|
(feat) raise exception on proxy server
|
2023-11-16 09:51:17 -08:00 |
|
ishaan-jaff
|
b283dd2a07
|
(test) add proxy cli testing
|
2023-11-16 09:51:17 -08:00 |
|
Krrish Dholakia
|
fcc250b312
|
docs(simple_proxy.md): add tutorial on handling multiple instances of 1 model
|
2023-11-16 09:14:15 -08:00 |
|
Ishaan Jaff
|
d6ad62d793
|
Merge pull request #811 from dchristian3188/bedrock-llama
Bedrock llama
|
2023-11-16 07:57:50 -08:00 |
|
Ishaan Jaff
|
3f95fd2da5
|
Merge pull request #826 from rodneyxr/ollama-fixes
Fix typo for initial_prompt_value and too many values to unpack error
|
2023-11-16 07:55:53 -08:00 |
|
David Christian
|
45ae4a76e4
|
updated utils for bedrock.meta streaming
|
2023-11-16 07:12:27 -08:00 |
|
Krrish Dholakia
|
791a3877d9
|
bump: version 1.0.3.dev1 → 1.0.3
|
2023-11-15 20:46:48 -08:00 |
|
Krrish Dholakia
|
93aae8669d
|
fix(openai.py): using openai sdk for completion calls
|
2023-11-15 20:25:27 -08:00 |
|