forked from phoenix/litellm-mirror
remove trust remote code option
This commit is contained in:
parent
ccb0516e98
commit
ce62dc40b9
8 changed files with 19 additions and 3 deletions
BIN
dist/litellm-0.1.781-py3-none-any.whl
vendored
Normal file
BIN
dist/litellm-0.1.781-py3-none-any.whl
vendored
Normal file
Binary file not shown.
BIN
dist/litellm-0.1.781.tar.gz
vendored
Normal file
BIN
dist/litellm-0.1.781.tar.gz
vendored
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
|
@ -274,7 +274,7 @@ def completion(
|
||||||
top_k=top_k,
|
top_k=top_k,
|
||||||
task=task,
|
task=task,
|
||||||
remove_input=remove_input,
|
remove_input=remove_input,
|
||||||
return_full_text=return_full_text
|
return_full_text=return_full_text,
|
||||||
)
|
)
|
||||||
# For logging - save the values of the litellm-specific params passed in
|
# For logging - save the values of the litellm-specific params passed in
|
||||||
litellm_params = get_litellm_params(
|
litellm_params = get_litellm_params(
|
||||||
|
|
|
@ -200,6 +200,22 @@ def hf_test_completion():
|
||||||
|
|
||||||
# hf_test_completion()
|
# hf_test_completion()
|
||||||
|
|
||||||
|
def hf_test_completion_task_none():
|
||||||
|
try:
|
||||||
|
litellm.set_verbose=True
|
||||||
|
user_message = "My name is Merve and my favorite"
|
||||||
|
messages = [{ "content": user_message,"role": "user"}]
|
||||||
|
model = "huggingface/cerebras/btlm-3b-8k-base"
|
||||||
|
response = completion(
|
||||||
|
model=model,
|
||||||
|
messages=messages,
|
||||||
|
task=None,
|
||||||
|
)
|
||||||
|
# Add any assertions here to check the response
|
||||||
|
print(response)
|
||||||
|
except Exception as e:
|
||||||
|
pytest.fail(f"Error occurred: {e}")
|
||||||
|
hf_test_completion_task_none()
|
||||||
|
|
||||||
# this should throw an exception, to trigger https://logs.litellm.ai/
|
# this should throw an exception, to trigger https://logs.litellm.ai/
|
||||||
# def hf_test_error_logs():
|
# def hf_test_error_logs():
|
||||||
|
|
|
@ -907,7 +907,7 @@ def get_optional_params( # use the openai defaults
|
||||||
custom_llm_provider="",
|
custom_llm_provider="",
|
||||||
top_k=40,
|
top_k=40,
|
||||||
return_full_text=False,
|
return_full_text=False,
|
||||||
task=None
|
task=None,
|
||||||
):
|
):
|
||||||
optional_params = {}
|
optional_params = {}
|
||||||
if model in litellm.anthropic_models:
|
if model in litellm.anthropic_models:
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
[tool.poetry]
|
[tool.poetry]
|
||||||
name = "litellm"
|
name = "litellm"
|
||||||
version = "0.1.780"
|
version = "0.1.782"
|
||||||
description = "Library to easily interface with LLM API providers"
|
description = "Library to easily interface with LLM API providers"
|
||||||
authors = ["BerriAI"]
|
authors = ["BerriAI"]
|
||||||
license = "MIT License"
|
license = "MIT License"
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue