From ff815e93a84689f269cdeef1e61c5a2d6712c19b Mon Sep 17 00:00:00 2001 From: ishaan-jaff Date: Fri, 23 Feb 2024 10:08:07 -0800 Subject: [PATCH] (fix) test_tokenizers --- litellm/tests/test_token_counter.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/litellm/tests/test_token_counter.py b/litellm/tests/test_token_counter.py index bf05232d9..af0db487e 100644 --- a/litellm/tests/test_token_counter.py +++ b/litellm/tests/test_token_counter.py @@ -53,7 +53,7 @@ def test_tokenizers(): try: ### test the openai, claude, cohere and llama2 tokenizers. ### The tokenizer value should be different for all - sample_text = "Hellö World, this is my input string!" + sample_text = "Hellö World, this is my input string! My name is ishaan CTO" # openai tokenizer openai_tokens = token_counter(model="gpt-3.5-turbo", text=sample_text)