Duc Pham
|
61f2e37349
|
Reverted error while refactoring
|
2023-11-10 01:35:41 +07:00 |
|
Duc Pham
|
c74e6f8cdd
|
Improved trimming logic and OpenAI token counter
|
2023-11-10 01:26:13 +07:00 |
|
Krrish Dholakia
|
e66373bd47
|
refactor(openai.py): moving openai text completion calls to http
|
2023-11-08 18:40:03 -08:00 |
|
Krrish Dholakia
|
decf86b145
|
refactor(openai.py): moving openai chat completion calls to http
|
2023-11-08 17:40:41 -08:00 |
|
Krrish Dholakia
|
17f5e46080
|
refactor(azure.py): moving azure openai calls to http calls
|
2023-11-08 16:52:18 -08:00 |
|
ishaan-jaff
|
11ee52207e
|
(feat) add streaming for text_completion
|
2023-11-08 11:58:07 -08:00 |
|
ishaan-jaff
|
106ccc2b94
|
(fix) text_completion don't pass echo to HF after translating
|
2023-11-08 11:45:05 -08:00 |
|
Krrish Dholakia
|
97c8b52bba
|
fix(utils.py): llmmonitor integration
|
2023-11-07 15:49:32 -08:00 |
|
ishaan-jaff
|
4d8d50d97e
|
(fix) HF round up temperature 0 -> 0.01
|
2023-11-06 14:35:06 -08:00 |
|
ishaan-jaff
|
b75a113e39
|
(fix) hf fix this error: Failed: Error occurred: HuggingfaceException - Input validation error: temperature must be strictly positive
|
2023-11-06 14:22:33 -08:00 |
|
ishaan-jaff
|
fdded281a9
|
(fix) bug fix: completion, text_completion, check if optional params are not None and pass to LLM
|
2023-11-06 13:17:19 -08:00 |
|
Krrish Dholakia
|
713c659d09
|
fix(utils.py): remove special characters from streaming output
|
2023-11-06 12:21:50 -08:00 |
|
ishaan-jaff
|
441ef48a54
|
(fix) improve litellm.set_verbose prints
|
2023-11-06 08:00:03 -08:00 |
|
Krrish Dholakia
|
10987304ba
|
bump: version 0.13.3.dev1 → 0.13.3.dev2
|
2023-11-06 06:44:15 -08:00 |
|
Krrish Dholakia
|
b8cc981db5
|
fix(utils.py): better exception raising if logging object is not able to get set
|
2023-11-06 06:34:27 -08:00 |
|
Krrish Dholakia
|
e633566253
|
feat(utils.py): adding additional states for custom logging
|
2023-11-04 17:07:20 -07:00 |
|
Krrish Dholakia
|
f7c5595a0d
|
fix(main.py): fixing print_verbose
|
2023-11-04 14:41:34 -07:00 |
|
ishaan-jaff
|
3477604c90
|
(fix) linting
|
2023-11-04 13:28:09 -07:00 |
|
ishaan-jaff
|
e53f5316d0
|
(fix) anyscale streaming detect [DONE] special char
|
2023-11-04 13:23:02 -07:00 |
|
Krrish Dholakia
|
d0b23a2722
|
refactor(all-files): removing all print statements; adding pre-commit + flake8 to prevent future regressions
|
2023-11-04 12:50:15 -07:00 |
|
ishaan-jaff
|
07f8fa65eb
|
(feat) add TextCompletionResponse
|
2023-11-03 22:14:07 -07:00 |
|
Krrish Dholakia
|
64b6b0155d
|
fix(bedrock.py): add exception mapping coverage for authentication scenarios
|
2023-11-03 18:25:34 -07:00 |
|
Krrish Dholakia
|
8bf8464fc2
|
fix(bedrock.py): fix bedrock exception mapping
|
2023-11-03 18:14:12 -07:00 |
|
Krrish Dholakia
|
fa24a61976
|
refactor(proxy_server.py): print statement showing how to add debug for logs
|
2023-11-03 17:41:14 -07:00 |
|
ishaan-jaff
|
5b76e12976
|
(fix) temp_top_logprobs
|
2023-11-03 16:45:10 -07:00 |
|
ishaan-jaff
|
5a5e6e0fac
|
(fix) remove errant print statements
|
2023-11-03 08:20:14 -07:00 |
|
Krrish Dholakia
|
127972a80b
|
build(litellm_server/utils.py): add support for general settings + num retries as a module variable
|
2023-11-02 20:56:41 -07:00 |
|
ishaan-jaff
|
fb94c7e00d
|
(fix) litellm utils
|
2023-11-02 17:03:46 -07:00 |
|
ishaan-jaff
|
3c22fbf637
|
(fix) vertexai detect code_chat and code_text llms as vertex
|
2023-11-02 16:31:13 -07:00 |
|
Krrish Dholakia
|
33c1118080
|
feat(completion()): enable setting prompt templates via completion()
|
2023-11-02 16:24:01 -07:00 |
|
ishaan-jaff
|
5262a3a2f7
|
(feat) add transform_logprobs for text_completion
|
2023-11-01 18:25:13 -07:00 |
|
ishaan-jaff
|
51060d1eea
|
(feat) text_completion add transform_logprobs
|
2023-11-01 18:25:13 -07:00 |
|
ishaan-jaff
|
a0ed669f25
|
(fix) improve litellm.set_verbose=True logging
|
2023-11-01 18:25:13 -07:00 |
|
ishaan-jaff
|
f66e9c6bce
|
(feat) detect amazon.titan-embed-text-v1 as bedrock embedding model
|
2023-11-01 14:46:33 -07:00 |
|
Krrish Dholakia
|
a951f7ff85
|
fix(utils.py): mapping stop sequences for palm
|
2023-11-01 14:00:45 -07:00 |
|
Krrish Dholakia
|
2cf06a3235
|
feat(utils.py): accept context window fallback dictionary
|
2023-10-31 22:32:36 -07:00 |
|
ishaan-jaff
|
c6229b7113
|
(feat) add bedrock.cohere streaming
|
2023-10-31 22:26:43 -07:00 |
|
ishaan-jaff
|
19177ae041
|
(feat) add support for echo for HF logprobs
|
2023-10-31 18:20:59 -07:00 |
|
ishaan-jaff
|
d57dc616b8
|
(feat) text_completion return raw response for davinci003
|
2023-10-31 15:32:04 -07:00 |
|
Krrish Dholakia
|
147d69f230
|
feat(main.py): add support for maritalk api
|
2023-10-30 17:36:51 -07:00 |
|
ishaan-jaff
|
32b6714a8b
|
(docs) completion_with_config
|
2023-10-30 14:29:40 -07:00 |
|
ishaan-jaff
|
ae376a9835
|
(docs) encode docstring
|
2023-10-30 14:10:29 -07:00 |
|
ishaan-jaff
|
fdef63439d
|
(docs) add docstring for validate_environment
|
2023-10-30 14:06:55 -07:00 |
|
ishaan-jaff
|
94542ae6be
|
def get_valid_models():
(docs) add docstring for
|
2023-10-30 14:05:09 -07:00 |
|
ishaan-jaff
|
bb5073fcbe
|
(docs) add doc string for check_valid_api_key
|
2023-10-30 14:02:31 -07:00 |
|
ishaan-jaff
|
0069b09caa
|
(feat) track cost for responses easily
|
2023-10-28 15:08:35 -07:00 |
|
ishaan-jaff
|
8803116ec6
|
(feat) set litellm_call_id in fallbacks
|
2023-10-27 18:09:00 -07:00 |
|
ishaan-jaff
|
ef449f155e
|
(fix) utils - remove bloat - deprecated completion_with_split_tests
|
2023-10-27 18:04:15 -07:00 |
|
ishaan-jaff
|
8ab127d6a3
|
(fix) only set litellm call id if it's not set in completion()
|
2023-10-27 18:02:14 -07:00 |
|
ishaan-jaff
|
a88ed12a9a
|
(fix) remove print statements from completion fallbacks, make them print verbose
|
2023-10-27 17:54:48 -07:00 |
|