from litellm.main import completion import os ## Configs for Models ## # OpenAI Configs ## End Configs for Models ## messages = [{ "content": "Hello, how are you?","role": "user"}] # # openai call # response = completion(model="gpt-3.5-turbo", messages=messages) # print("\nOpenAI call") # print(response) # # azure openai call # response = completion("chatgpt-test", messages, azure=True) # print("\nAzure call") # print(response) # # text davinci openai call # response = completion("text-davinci-003", messages) # print("\nDavinci call") # print(response) # # cohere call # response = completion("command-nightly", messages) # print("\nCohere call") # print(response) # openrouter call response = completion("google/palm-2-codechat-bison", messages) print("\OpenRouter call") print(response)