From b5851769fca5f6bc70a00a5582aa899c058e5e03 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Fri, 4 Apr 2025 20:26:11 -0700 Subject: [PATCH] fix: fix import --- tests/llm_translation/test_max_completion_tokens.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/tests/llm_translation/test_max_completion_tokens.py b/tests/llm_translation/test_max_completion_tokens.py index a8f3dd50a8..f1374a22a2 100644 --- a/tests/llm_translation/test_max_completion_tokens.py +++ b/tests/llm_translation/test_max_completion_tokens.py @@ -330,7 +330,7 @@ def test_all_model_configs(): drop_params=False, ) == {"max_tokens_to_sample": 10} - from litellm.llms.databricks.chat.handler import DatabricksConfig + from litellm.llms.databricks.chat.transformation import DatabricksConfig assert "max_completion_tokens" in DatabricksConfig().get_supported_openai_params()