diff --git a/litellm/tests/test_add_function_to_prompt.py b/litellm/tests/test_add_function_to_prompt.py index 94b9de5b30..33d2ac2a97 100644 --- a/litellm/tests/test_add_function_to_prompt.py +++ b/litellm/tests/test_add_function_to_prompt.py @@ -1,75 +1,75 @@ -# #### What this tests #### -# # Allow the user to map the function to the prompt, if the model doesn't support function calling +#### What this tests #### +# Allow the user to map the function to the prompt, if the model doesn't support function calling -# import sys, os, pytest -# import traceback +import sys, os, pytest +import traceback -# sys.path.insert( -# 0, os.path.abspath("../..") -# ) # Adds the parent directory to the system path -# import litellm +sys.path.insert( + 0, os.path.abspath("../..") +) # Adds the parent directory to the system path +import litellm -# ## case 1: set_function_to_prompt not set -# def test_function_call_non_openai_model(): -# try: -# model = "claude-instant-1" -# messages=[{"role": "user", "content": "what's the weather in sf?"}] -# functions = [ -# { -# "name": "get_current_weather", -# "description": "Get the current weather in a given location", -# "parameters": { -# "type": "object", -# "properties": { -# "location": { -# "type": "string", -# "description": "The city and state, e.g. San Francisco, CA" -# }, -# "unit": { -# "type": "string", -# "enum": ["celsius", "fahrenheit"] -# } -# }, -# "required": ["location"] -# } -# } -# ] -# response = litellm.completion(model=model, messages=messages, functions=functions) -# pytest.fail(f'An error occurred') -# except Exception as e: -# pass +## case 1: set_function_to_prompt not set +def test_function_call_non_openai_model(): + try: + model = "claude-instant-1" + messages=[{"role": "user", "content": "what's the weather in sf?"}] + functions = [ + { + "name": "get_current_weather", + "description": "Get the current weather in a given location", + "parameters": { + "type": "object", + "properties": { + "location": { + "type": "string", + "description": "The city and state, e.g. San Francisco, CA" + }, + "unit": { + "type": "string", + "enum": ["celsius", "fahrenheit"] + } + }, + "required": ["location"] + } + } + ] + response = litellm.completion(model=model, messages=messages, functions=functions) + pytest.fail(f'An error occurred') + except Exception as e: + pass -# test_function_call_non_openai_model() +test_function_call_non_openai_model() -# ## case 2: add_function_to_prompt set -# def test_function_call_non_openai_model_litellm_mod_set(): -# litellm.add_function_to_prompt = True -# try: -# model = "claude-instant-1" -# messages=[{"role": "user", "content": "what's the weather in sf?"}] -# functions = [ -# { -# "name": "get_current_weather", -# "description": "Get the current weather in a given location", -# "parameters": { -# "type": "object", -# "properties": { -# "location": { -# "type": "string", -# "description": "The city and state, e.g. San Francisco, CA" -# }, -# "unit": { -# "type": "string", -# "enum": ["celsius", "fahrenheit"] -# } -# }, -# "required": ["location"] -# } -# } -# ] -# response = litellm.completion(model=model, messages=messages, functions=functions) -# print(f'response: {response}') -# except Exception as e: -# pytest.fail(f'An error occurred {e}') +## case 2: add_function_to_prompt set +def test_function_call_non_openai_model_litellm_mod_set(): + litellm.add_function_to_prompt = True + try: + model = "claude-instant-1" + messages=[{"role": "user", "content": "what's the weather in sf?"}] + functions = [ + { + "name": "get_current_weather", + "description": "Get the current weather in a given location", + "parameters": { + "type": "object", + "properties": { + "location": { + "type": "string", + "description": "The city and state, e.g. San Francisco, CA" + }, + "unit": { + "type": "string", + "enum": ["celsius", "fahrenheit"] + } + }, + "required": ["location"] + } + } + ] + response = litellm.completion(model=model, messages=messages, functions=functions) + print(f'response: {response}') + except Exception as e: + pytest.fail(f'An error occurred {e}') -# # test_function_call_non_openai_model_litellm_mod_set() \ No newline at end of file +# test_function_call_non_openai_model_litellm_mod_set() \ No newline at end of file