mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 19:24:27 +00:00
fix(router.py): fix dereferencing param order
This commit is contained in:
parent
0b794cbd8f
commit
e996e17e18
1 changed files with 4 additions and 4 deletions
|
@ -394,11 +394,11 @@ class Router:
|
||||||
|
|
||||||
response = litellm.completion(
|
response = litellm.completion(
|
||||||
**{
|
**{
|
||||||
**kwargs,
|
|
||||||
**data,
|
|
||||||
"messages": messages,
|
"messages": messages,
|
||||||
"caching": self.cache_responses,
|
"caching": self.cache_responses,
|
||||||
"client": model_client,
|
"client": model_client,
|
||||||
|
**kwargs,
|
||||||
|
**data,
|
||||||
}
|
}
|
||||||
)
|
)
|
||||||
verbose_router_logger.info(
|
verbose_router_logger.info(
|
||||||
|
@ -527,12 +527,12 @@ class Router:
|
||||||
model_client = potential_model_client
|
model_client = potential_model_client
|
||||||
self.total_calls[model_name] += 1
|
self.total_calls[model_name] += 1
|
||||||
final_data = {
|
final_data = {
|
||||||
**kwargs,
|
|
||||||
**data,
|
|
||||||
"messages": messages,
|
"messages": messages,
|
||||||
"caching": self.cache_responses,
|
"caching": self.cache_responses,
|
||||||
"client": model_client,
|
"client": model_client,
|
||||||
"timeout": self.timeout,
|
"timeout": self.timeout,
|
||||||
|
**kwargs,
|
||||||
|
**data,
|
||||||
}
|
}
|
||||||
response = await litellm.acompletion(**final_data)
|
response = await litellm.acompletion(**final_data)
|
||||||
self.success_calls[model_name] += 1
|
self.success_calls[model_name] += 1
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue