Krrish Dholakia
|
d6e142428b
|
fix(openai-+-azure): async calling
|
2023-11-15 17:03:02 -08:00 |
|
ishaan-jaff
|
eaa55ac6fd
|
(chore) remove bloat deprecated config.json
|
2023-11-15 17:00:14 -08:00 |
|
ishaan-jaff
|
717d3995b5
|
(chore) remove bloat - deprecated gpt cache integration
|
2023-11-15 16:59:31 -08:00 |
|
ishaan-jaff
|
6d9d79d12b
|
(chore) remove bloat old testing.py file
|
2023-11-15 16:58:54 -08:00 |
|
ishaan-jaff
|
5b79abdfb6
|
(chore) rename litellm server
|
2023-11-15 16:57:42 -08:00 |
|
ishaan-jaff
|
08c182c57c
|
(chore) move litellm_server inside litellm
|
2023-11-15 16:56:35 -08:00 |
|
Krrish Dholakia
|
03efc9185e
|
fix(huggingface_restapi.py): async implementation
|
2023-11-15 16:54:15 -08:00 |
|
ishaan-jaff
|
cc955fca89
|
(docs) config
|
2023-11-15 16:48:01 -08:00 |
|
ishaan-jaff
|
9493aeff68
|
(docs) add example configs to proxy
|
2023-11-15 16:41:26 -08:00 |
|
ishaan-jaff
|
a18fe5a1fd
|
(fix) proxy raise exceptions from litellm.completion()
|
2023-11-15 16:26:04 -08:00 |
|
ishaan-jaff
|
e698f8218f
|
(feat) proxy - show loaded config for users
|
2023-11-15 16:26:04 -08:00 |
|
Krrish Dholakia
|
bcea28e2e4
|
fix(utils): fixing exception mapping
|
2023-11-15 15:51:17 -08:00 |
|
Krrish Dholakia
|
a59494571f
|
fix(huggingface_restapi.py): fix linting errors
|
2023-11-15 15:34:21 -08:00 |
|
Krrish Dholakia
|
f84db3ce14
|
feat(get_max_tokens): get max tokens for huggingface hub models
|
2023-11-15 15:25:40 -08:00 |
|
Krrish Dholakia
|
1a705bfbcb
|
refactor(huggingface_restapi.py): moving async completion + streaming to real async calls
|
2023-11-15 15:14:21 -08:00 |
|
ishaan-jaff
|
77394e7987
|
(docs) using litellm proxy + langfuse
|
2023-11-15 14:29:14 -08:00 |
|
Krrish Dholakia
|
c8b73d90f1
|
fix(utils.py): fix langfuse integration
|
2023-11-15 14:05:40 -08:00 |
|
ishaan-jaff
|
5625fe3d81
|
(docs) using langfuse with litellm
|
2023-11-15 13:53:12 -08:00 |
|
ishaan-jaff
|
e9dfea887c
|
(docs) using litellm proxy + Langfuse
|
2023-11-15 13:52:47 -08:00 |
|
ishaan-jaff
|
11872e0d45
|
(fix) proxy safe access router model names
|
2023-11-15 13:44:17 -08:00 |
|
ishaan-jaff
|
b7859f0dd1
|
(docs) show how to debug proxy logs
|
2023-11-15 13:44:17 -08:00 |
|
Krrish Dholakia
|
60d2dff74a
|
fix(utils.py): check for none params
|
2023-11-15 13:39:09 -08:00 |
|
ishaan-jaff
|
8cced8eabc
|
(docs) proxy cli settings
|
2023-11-15 13:33:29 -08:00 |
|
Krrish Dholakia
|
4348fd6435
|
fix(utils.py): azure streaming initial format
|
2023-11-15 13:30:08 -08:00 |
|
ishaan-jaff
|
ae7731b4dc
|
(docs) using litellm with curl, OpenAI v1.0.0+
|
2023-11-15 13:29:20 -08:00 |
|
Krrish Dholakia
|
33d823a18d
|
docs(routing.md): add contributor s/o
|
2023-11-15 13:20:02 -08:00 |
|
Krrish Dholakia
|
8c104e9c6a
|
fix(azure.py-+-proxy_server.py): fix function calling response object + support router on proxy
|
2023-11-15 13:15:16 -08:00 |
|
ishaan-jaff
|
8f8b35eb34
|
add ollama/zephyr to config
|
2023-11-15 13:04:34 -08:00 |
|
ishaan-jaff
|
f3bbedbbbc
|
(docs) proxy add example on using local hf llm
|
2023-11-15 13:00:43 -08:00 |
|
ishaan-jaff
|
6f923ea813
|
(fix) remove bloat from ci/cid
|
2023-11-15 12:39:51 -08:00 |
|
Krrish Dholakia
|
563a3d3016
|
refactor(router.py): renaming variable
|
2023-11-15 12:31:29 -08:00 |
|
Krrish Dholakia
|
03303033e5
|
feat(router.py): enable passing chat completion params for Router.chat.completion.create
|
2023-11-15 12:28:16 -08:00 |
|
Krish Dholakia
|
95f9c6779d
|
Merge pull request #820 from kfsone/main
need to re-attempt backoff and yaml imports if the first import attempt fails
|
2023-11-15 11:28:34 -08:00 |
|
Oliver Smith
|
8e4349d3dc
|
Correct pypi package name for yaml
|
2023-11-15 10:46:26 -08:00 |
|
Oliver Smith
|
7cf1846654
|
Ensure the imports needed by proxy_server are in requirements.txt to start with
|
2023-11-15 10:38:05 -08:00 |
|
Oliver Smith
|
4c980cb161
|
need to re-attempt backoff and yaml imports if the first import attempt fails. not sure which import is missing from requirements
|
2023-11-15 00:10:22 -08:00 |
|
Krrish Dholakia
|
f5abde775b
|
test(test_completion.py): fix palm test
|
2023-11-14 22:28:56 -08:00 |
|
Krrish Dholakia
|
aafd8a50d2
|
bump: version 1.0.2 → 1.0.3.dev1
|
2023-11-14 22:20:17 -08:00 |
|
Krrish Dholakia
|
e07bf0a8de
|
fix(utils.py): await async function in client wrapper
|
2023-11-14 22:07:28 -08:00 |
|
Krrish Dholakia
|
efe81032f4
|
fix: fix linting errors
|
2023-11-14 21:51:39 -08:00 |
|
Krrish Dholakia
|
59eaeba92a
|
fix(router.py): enabling retrying with expo backoff (without tenacity) for router
|
2023-11-14 20:57:51 -08:00 |
|
Krrish Dholakia
|
98c45f1b4e
|
bump: version 1.0.1 → 1.0.2
|
2023-11-14 18:55:20 -08:00 |
|
Krrish Dholakia
|
1738341dcb
|
fix(main.py): misrouting ollama models to nlp cloud
|
2023-11-14 18:55:08 -08:00 |
|
ishaan-jaff
|
465f427465
|
(feat) debug POST logs
|
2023-11-14 18:16:45 -08:00 |
|
ishaan-jaff
|
c7780cbc40
|
(fix) debugging with POST request
|
2023-11-14 18:05:34 -08:00 |
|
ishaan-jaff
|
f650be4fee
|
(feat) completion debug view HF POST request
|
2023-11-14 17:57:41 -08:00 |
|
ishaan-jaff
|
cb5669f679
|
(feat) improve logging of raw POST curl command
|
2023-11-14 17:54:09 -08:00 |
|
ishaan-jaff
|
e82b8ed7e2
|
(feat) debug ollama POST request
|
2023-11-14 17:53:48 -08:00 |
|
ishaan-jaff
|
7c317b78eb
|
(feat) add ability to view POST requests from litellm.completion()
|
2023-11-14 17:27:20 -08:00 |
|
Krrish Dholakia
|
a7222f257c
|
fix(main.py): keep client consistent across calls + exponential backoff retry on ratelimit errors
|
2023-11-14 16:26:05 -08:00 |
|