# ##### THESE TESTS CAN ONLY RUN LOCALLY WITH THE OLLAMA SERVER RUNNING ###### # # https://ollama.ai/ # import sys, os # import traceback # from dotenv import load_dotenv # load_dotenv() # import os # sys.path.insert(0, os.path.abspath('../..')) # Adds the parent directory to the system path # import pytest # import litellm # from litellm import embedding, completion # import asyncio # user_message = "respond in 20 words. who are you?" # messages = [{ "content": user_message,"role": "user"}] # def test_completion_ollama(): # try: # response = completion( # model="llama2", # messages=messages, # api_base="http://localhost:11434", # custom_llm_provider="ollama" # ) # print(response) # except Exception as e: # pytest.fail(f"Error occurred: {e}") # test_completion_ollama() # def test_completion_ollama_stream(): # user_message = "what is litellm?" # messages = [{ "content": user_message,"role": "user"}] # try: # response = completion( # model="llama2", # messages=messages, # api_base="http://localhost:11434", # custom_llm_provider="ollama", # stream=True # ) # print(response) # for chunk in response: # print(chunk['choices'][0]['delta']) # except Exception as e: # pytest.fail(f"Error occurred: {e}") # # test_completion_ollama_stream()