From 0a5207a55833e809bfad8e0cf4c6f9dfe04f9683 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Sat, 9 Dec 2023 22:43:27 -0800 Subject: [PATCH] test(test_proxy_server.py): add testing for /model/new endpoint --- litellm/tests/test_proxy_server.py | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/litellm/tests/test_proxy_server.py b/litellm/tests/test_proxy_server.py index bc22c384fa..e3094237b0 100644 --- a/litellm/tests/test_proxy_server.py +++ b/litellm/tests/test_proxy_server.py @@ -122,7 +122,7 @@ def test_embedding(client): # Run the test # test_embedding() -@pytest.mark.skip(reason="hitting yaml load issues on circle-ci") +# @pytest.mark.skip(reason="hitting yaml load issues on circle-ci") def test_add_new_model(client): global headers try: @@ -142,8 +142,8 @@ def test_add_new_model(client): print(f"response: {result}") model_info = None for m in result["data"]: - if m["id"]["model_name"] == "test_openai_models": - model_info = m["id"]["model_info"] + if m["model_name"] == "test_openai_models": + model_info = m["model_info"] assert model_info["description"] == "this is a test openai model" except Exception as e: pytest.fail(f"LiteLLM Proxy test failed. Exception {str(e)}")