Commit graph

3465 commits

Author SHA1 Message Date
Krrish Dholakia
d6e142428b fix(openai-+-azure): async calling 2023-11-15 17:03:02 -08:00
ishaan-jaff
eaa55ac6fd (chore) remove bloat deprecated config.json 2023-11-15 17:00:14 -08:00
ishaan-jaff
717d3995b5 (chore) remove bloat - deprecated gpt cache integration 2023-11-15 16:59:31 -08:00
ishaan-jaff
6d9d79d12b (chore) remove bloat old testing.py file 2023-11-15 16:58:54 -08:00
ishaan-jaff
5b79abdfb6 (chore) rename litellm server 2023-11-15 16:57:42 -08:00
ishaan-jaff
08c182c57c (chore) move litellm_server inside litellm 2023-11-15 16:56:35 -08:00
Krrish Dholakia
03efc9185e fix(huggingface_restapi.py): async implementation 2023-11-15 16:54:15 -08:00
ishaan-jaff
cc955fca89 (docs) config 2023-11-15 16:48:01 -08:00
ishaan-jaff
9493aeff68 (docs) add example configs to proxy 2023-11-15 16:41:26 -08:00
ishaan-jaff
a18fe5a1fd (fix) proxy raise exceptions from litellm.completion() 2023-11-15 16:26:04 -08:00
ishaan-jaff
e698f8218f (feat) proxy - show loaded config for users 2023-11-15 16:26:04 -08:00
Krrish Dholakia
bcea28e2e4 fix(utils): fixing exception mapping 2023-11-15 15:51:17 -08:00
Krrish Dholakia
a59494571f fix(huggingface_restapi.py): fix linting errors 2023-11-15 15:34:21 -08:00
Krrish Dholakia
f84db3ce14 feat(get_max_tokens): get max tokens for huggingface hub models 2023-11-15 15:25:40 -08:00
Krrish Dholakia
1a705bfbcb refactor(huggingface_restapi.py): moving async completion + streaming to real async calls 2023-11-15 15:14:21 -08:00
ishaan-jaff
77394e7987 (docs) using litellm proxy + langfuse 2023-11-15 14:29:14 -08:00
Krrish Dholakia
c8b73d90f1 fix(utils.py): fix langfuse integration 2023-11-15 14:05:40 -08:00
ishaan-jaff
5625fe3d81 (docs) using langfuse with litellm 2023-11-15 13:53:12 -08:00
ishaan-jaff
e9dfea887c (docs) using litellm proxy + Langfuse 2023-11-15 13:52:47 -08:00
ishaan-jaff
11872e0d45 (fix) proxy safe access router model names 2023-11-15 13:44:17 -08:00
ishaan-jaff
b7859f0dd1 (docs) show how to debug proxy logs 2023-11-15 13:44:17 -08:00
Krrish Dholakia
60d2dff74a fix(utils.py): check for none params 2023-11-15 13:39:09 -08:00
ishaan-jaff
8cced8eabc (docs) proxy cli settings 2023-11-15 13:33:29 -08:00
Krrish Dholakia
4348fd6435 fix(utils.py): azure streaming initial format 2023-11-15 13:30:08 -08:00
ishaan-jaff
ae7731b4dc (docs) using litellm with curl, OpenAI v1.0.0+ 2023-11-15 13:29:20 -08:00
Krrish Dholakia
33d823a18d docs(routing.md): add contributor s/o 2023-11-15 13:20:02 -08:00
Krrish Dholakia
8c104e9c6a fix(azure.py-+-proxy_server.py): fix function calling response object + support router on proxy 2023-11-15 13:15:16 -08:00
ishaan-jaff
8f8b35eb34 add ollama/zephyr to config 2023-11-15 13:04:34 -08:00
ishaan-jaff
f3bbedbbbc (docs) proxy add example on using local hf llm 2023-11-15 13:00:43 -08:00
ishaan-jaff
6f923ea813 (fix) remove bloat from ci/cid 2023-11-15 12:39:51 -08:00
Krrish Dholakia
563a3d3016 refactor(router.py): renaming variable 2023-11-15 12:31:29 -08:00
Krrish Dholakia
03303033e5 feat(router.py): enable passing chat completion params for Router.chat.completion.create 2023-11-15 12:28:16 -08:00
Krish Dholakia
95f9c6779d
Merge pull request #820 from kfsone/main
need to re-attempt backoff and yaml imports if the first import attempt fails
2023-11-15 11:28:34 -08:00
Oliver Smith
8e4349d3dc Correct pypi package name for yaml 2023-11-15 10:46:26 -08:00
Oliver Smith
7cf1846654 Ensure the imports needed by proxy_server are in requirements.txt to start with 2023-11-15 10:38:05 -08:00
Oliver Smith
4c980cb161 need to re-attempt backoff and yaml imports if the first import attempt fails. not sure which import is missing from requirements 2023-11-15 00:10:22 -08:00
Krrish Dholakia
f5abde775b test(test_completion.py): fix palm test 2023-11-14 22:28:56 -08:00
Krrish Dholakia
aafd8a50d2 bump: version 1.0.2 → 1.0.3.dev1 2023-11-14 22:20:17 -08:00
Krrish Dholakia
e07bf0a8de fix(utils.py): await async function in client wrapper 2023-11-14 22:07:28 -08:00
Krrish Dholakia
efe81032f4 fix: fix linting errors 2023-11-14 21:51:39 -08:00
Krrish Dholakia
59eaeba92a fix(router.py): enabling retrying with expo backoff (without tenacity) for router 2023-11-14 20:57:51 -08:00
Krrish Dholakia
98c45f1b4e bump: version 1.0.1 → 1.0.2 2023-11-14 18:55:20 -08:00
Krrish Dholakia
1738341dcb fix(main.py): misrouting ollama models to nlp cloud 2023-11-14 18:55:08 -08:00
ishaan-jaff
465f427465 (feat) debug POST logs 2023-11-14 18:16:45 -08:00
ishaan-jaff
c7780cbc40 (fix) debugging with POST request 2023-11-14 18:05:34 -08:00
ishaan-jaff
f650be4fee (feat) completion debug view HF POST request 2023-11-14 17:57:41 -08:00
ishaan-jaff
cb5669f679 (feat) improve logging of raw POST curl command 2023-11-14 17:54:09 -08:00
ishaan-jaff
e82b8ed7e2 (feat) debug ollama POST request 2023-11-14 17:53:48 -08:00
ishaan-jaff
7c317b78eb (feat) add ability to view POST requests from litellm.completion() 2023-11-14 17:27:20 -08:00
Krrish Dholakia
a7222f257c fix(main.py): keep client consistent across calls + exponential backoff retry on ratelimit errors 2023-11-14 16:26:05 -08:00