mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 03:04:13 +00:00
fix - issue when loading models through ui
This commit is contained in:
parent
b773133160
commit
fc832f2196
1 changed files with 1 additions and 4 deletions
|
@ -2443,9 +2443,7 @@ class ProxyConfig:
|
|||
f"llm_router model list: {llm_router.model_list}"
|
||||
)
|
||||
else:
|
||||
new_models = await prisma_client.db.litellm_proxymodeltable.find_many(
|
||||
take=10, order={"updated_at": "desc"}
|
||||
)
|
||||
new_models = await prisma_client.db.litellm_proxymodeltable.find_many()
|
||||
# verbose_proxy_logger.debug(f"new_models: {new_models}")
|
||||
|
||||
for m in new_models:
|
||||
|
@ -2493,7 +2491,6 @@ class ProxyConfig:
|
|||
for success_callback in success_callbacks:
|
||||
if success_callback not in litellm.success_callback:
|
||||
litellm.success_callback.append(success_callback)
|
||||
_added_callback = True
|
||||
# we need to set env variables too
|
||||
environment_variables = config_data.get("environment_variables", {})
|
||||
for k, v in environment_variables.items():
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue