mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-24 18:24:20 +00:00
fix: fix linting errors
This commit is contained in:
parent
5d5ca9f7ef
commit
2f93c0155a
2 changed files with 4 additions and 4 deletions
|
@ -18,6 +18,6 @@ try:
|
|||
# print("Got openai Timeout Exception. Good job. The proxy mapped to OpenAI exceptions")
|
||||
except Exception as e:
|
||||
print("\n the proxy did not map to OpenAI exception. Instead got", e)
|
||||
print(e.type)
|
||||
print(e.message)
|
||||
print(e.code)
|
||||
print(e.type) # type: ignore
|
||||
print(e.message) # type: ignore
|
||||
print(e.code) # type: ignore
|
|
@ -52,7 +52,7 @@ class Router:
|
|||
self.set_model_list(model_list)
|
||||
self.healthy_deployments: List = self.model_list
|
||||
self.deployment_latency_map = {}
|
||||
self.cooldown_deployments = {} # {"gpt-3.5-turbo": time.time() when it failed / needed a cooldown}
|
||||
self.cooldown_deployments: dict = {} # {"gpt-3.5-turbo": time.time() when it failed / needed a cooldown}
|
||||
for m in model_list:
|
||||
self.deployment_latency_map[m["litellm_params"]["model"]] = 0
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue