import os import sys, os import traceback from dotenv import load_dotenv load_dotenv() import os, io sys.path.insert( 0, os.path.abspath("../..") ) # Adds the parent directory to the system path import pytest import litellm from litellm import embedding, completion, text_completion, completion_cost from langchain.chat_models import ChatLiteLLM from langchain.prompts.chat import ( ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate, ) from langchain.schema import AIMessage, HumanMessage, SystemMessage def test_chat_gpt(): try: chat = ChatLiteLLM(model="gpt-3.5-turbo", max_tokens=10) messages = [ HumanMessage( content="what model are you" ) ] resp = chat(messages) print(resp) except Exception as e: pytest.fail(f"Error occurred: {e}") # test_chat_gpt() def test_claude(): try: chat = ChatLiteLLM(model="claude-2", max_tokens=10) messages = [ HumanMessage( content="what model are you" ) ] resp = chat(messages) print(resp) except Exception as e: pytest.fail(f"Error occurred: {e}") # test_claude() def test_palm(): try: chat = ChatLiteLLM(model="palm/chat-bison", max_tokens=10) messages = [ HumanMessage( content="what model are you" ) ] resp = chat(messages) print(resp) except Exception as e: pytest.fail(f"Error occurred: {e}") # test_palm() # def test_openai_with_params(): # try: # api_key = os.environ["OPENAI_API_KEY"] # os.environ.pop("OPENAI_API_KEY") # print("testing openai with params") # llm = ChatLiteLLM( # model="gpt-3.5-turbo", # openai_api_key=api_key, # # Prefer using None which is the default value, endpoint could be empty string # openai_api_base= None, # max_tokens=20, # temperature=0.5, # request_timeout=10, # model_kwargs={ # "frequency_penalty": 0, # "presence_penalty": 0, # }, # verbose=True, # max_retries=0, # ) # messages = [ # HumanMessage( # content="what model are you" # ) # ] # resp = llm(messages) # print(resp) # except Exception as e: # pytest.fail(f"Error occurred: {e}") # test_openai_with_params()