mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 19:24:27 +00:00
fix(__init__.py): add gemini models to all model list
Fixes https://github.com/BerriAI/litellm/issues/4240
This commit is contained in:
parent
fa6ddcde3c
commit
6482e57f56
2 changed files with 32 additions and 17 deletions
|
@ -1,9 +1,9 @@
|
|||
import copy
|
||||
import sys
|
||||
from datetime import datetime
|
||||
from unittest import mock
|
||||
|
||||
from dotenv import load_dotenv
|
||||
import copy
|
||||
from datetime import datetime
|
||||
|
||||
load_dotenv()
|
||||
import os
|
||||
|
@ -12,25 +12,26 @@ sys.path.insert(
|
|||
0, os.path.abspath("../..")
|
||||
) # Adds the parent directory to the system path
|
||||
import pytest
|
||||
|
||||
import litellm
|
||||
from litellm.utils import (
|
||||
trim_messages,
|
||||
get_token_count,
|
||||
get_valid_models,
|
||||
check_valid_key,
|
||||
validate_environment,
|
||||
function_to_dict,
|
||||
token_counter,
|
||||
create_pretrained_tokenizer,
|
||||
create_tokenizer,
|
||||
get_max_tokens,
|
||||
get_supported_openai_params,
|
||||
)
|
||||
from litellm.proxy.utils import (
|
||||
_duration_in_seconds,
|
||||
_extract_from_regex,
|
||||
get_last_day_of_month,
|
||||
)
|
||||
from litellm.utils import (
|
||||
check_valid_key,
|
||||
create_pretrained_tokenizer,
|
||||
create_tokenizer,
|
||||
function_to_dict,
|
||||
get_max_tokens,
|
||||
get_supported_openai_params,
|
||||
get_token_count,
|
||||
get_valid_models,
|
||||
token_counter,
|
||||
trim_messages,
|
||||
validate_environment,
|
||||
)
|
||||
|
||||
# Assuming your trim_messages, shorten_message_to_fit_limit, and get_token_count functions are all in a module named 'message_utils'
|
||||
|
||||
|
@ -216,6 +217,16 @@ def test_get_valid_models():
|
|||
# reset replicate env key
|
||||
os.environ = old_environ
|
||||
|
||||
# GEMINI
|
||||
expected_models = litellm.gemini_models
|
||||
old_environ = os.environ
|
||||
os.environ = {"GEMINI_API_KEY": "temp"} # mock set only openai key in environ
|
||||
|
||||
valid_models = get_valid_models()
|
||||
|
||||
print(valid_models)
|
||||
assert valid_models == expected_models
|
||||
|
||||
|
||||
# test_get_valid_models()
|
||||
|
||||
|
@ -409,10 +420,10 @@ def test_redact_msgs_from_logs():
|
|||
|
||||
On the proxy some users were seeing the redaction impact client side responses
|
||||
"""
|
||||
from litellm.litellm_core_utils.litellm_logging import Logging
|
||||
from litellm.litellm_core_utils.redact_messages import (
|
||||
redact_message_input_output_from_logging,
|
||||
)
|
||||
from litellm.litellm_core_utils.litellm_logging import Logging
|
||||
|
||||
litellm.turn_off_message_logging = True
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue