mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
(test) router_get_deployments - stop using remove_model_id
This commit is contained in:
parent
0a6f6bd045
commit
fff148a835
2 changed files with 6 additions and 7 deletions
|
@ -69,7 +69,6 @@ async def _perform_health_check(model_list: list):
|
||||||
for model in model_list:
|
for model in model_list:
|
||||||
litellm_params = model["litellm_params"]
|
litellm_params = model["litellm_params"]
|
||||||
model_info = model.get("model_info", {})
|
model_info = model.get("model_info", {})
|
||||||
litellm_params["model"] = litellm.utils.remove_model_id(litellm_params["model"])
|
|
||||||
litellm_params["messages"] = _get_random_llm_message()
|
litellm_params["messages"] = _get_random_llm_message()
|
||||||
|
|
||||||
prepped_params.append(litellm_params)
|
prepped_params.append(litellm_params)
|
||||||
|
|
|
@ -49,7 +49,7 @@ def test_weighted_selection_router():
|
||||||
for _ in range(1000):
|
for _ in range(1000):
|
||||||
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
||||||
selected_model_id = selected_model["litellm_params"]["model"]
|
selected_model_id = selected_model["litellm_params"]["model"]
|
||||||
selected_model_name = litellm.utils.remove_model_id(selected_model_id)
|
selected_model_name = selected_model_id
|
||||||
selection_counts[selected_model_name] +=1
|
selection_counts[selected_model_name] +=1
|
||||||
print(selection_counts)
|
print(selection_counts)
|
||||||
|
|
||||||
|
@ -101,7 +101,7 @@ def test_weighted_selection_router_tpm():
|
||||||
for _ in range(1000):
|
for _ in range(1000):
|
||||||
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
||||||
selected_model_id = selected_model["litellm_params"]["model"]
|
selected_model_id = selected_model["litellm_params"]["model"]
|
||||||
selected_model_name = litellm.utils.remove_model_id(selected_model_id)
|
selected_model_name = selected_model_id
|
||||||
selection_counts[selected_model_name] +=1
|
selection_counts[selected_model_name] +=1
|
||||||
print(selection_counts)
|
print(selection_counts)
|
||||||
|
|
||||||
|
@ -154,7 +154,7 @@ def test_weighted_selection_router_tpm_as_router_param():
|
||||||
for _ in range(1000):
|
for _ in range(1000):
|
||||||
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
||||||
selected_model_id = selected_model["litellm_params"]["model"]
|
selected_model_id = selected_model["litellm_params"]["model"]
|
||||||
selected_model_name = litellm.utils.remove_model_id(selected_model_id)
|
selected_model_name = selected_model_id
|
||||||
selection_counts[selected_model_name] +=1
|
selection_counts[selected_model_name] +=1
|
||||||
print(selection_counts)
|
print(selection_counts)
|
||||||
|
|
||||||
|
@ -210,7 +210,7 @@ def test_weighted_selection_router_rpm_as_router_param():
|
||||||
for _ in range(1000):
|
for _ in range(1000):
|
||||||
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
||||||
selected_model_id = selected_model["litellm_params"]["model"]
|
selected_model_id = selected_model["litellm_params"]["model"]
|
||||||
selected_model_name = litellm.utils.remove_model_id(selected_model_id)
|
selected_model_name = selected_model_id
|
||||||
selection_counts[selected_model_name] +=1
|
selection_counts[selected_model_name] +=1
|
||||||
print(selection_counts)
|
print(selection_counts)
|
||||||
|
|
||||||
|
@ -270,7 +270,7 @@ def test_weighted_selection_router_no_rpm_set():
|
||||||
for _ in range(1000):
|
for _ in range(1000):
|
||||||
selected_model = router.get_available_deployment("claude-1")
|
selected_model = router.get_available_deployment("claude-1")
|
||||||
selected_model_id = selected_model["litellm_params"]["model"]
|
selected_model_id = selected_model["litellm_params"]["model"]
|
||||||
selected_model_name = litellm.utils.remove_model_id(selected_model_id)
|
selected_model_name = selected_model_id
|
||||||
selection_counts[selected_model_name] +=1
|
selection_counts[selected_model_name] +=1
|
||||||
print(selection_counts)
|
print(selection_counts)
|
||||||
|
|
||||||
|
@ -337,7 +337,7 @@ def test_model_group_aliases():
|
||||||
for _ in range(1000):
|
for _ in range(1000):
|
||||||
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
selected_model = router.get_available_deployment("gpt-3.5-turbo")
|
||||||
selected_model_id = selected_model["litellm_params"]["model"]
|
selected_model_id = selected_model["litellm_params"]["model"]
|
||||||
selected_model_name = litellm.utils.remove_model_id(selected_model_id)
|
selected_model_name = selected_model_id
|
||||||
selection_counts[selected_model_name] +=1
|
selection_counts[selected_model_name] +=1
|
||||||
print(selection_counts)
|
print(selection_counts)
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue