mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
fix(utils.py): move to using litellm.modify_params
to enable max output token trimming fix
This commit is contained in:
parent
bdd2004691
commit
a634424fb2
2 changed files with 2 additions and 1 deletions
|
@ -2589,7 +2589,7 @@ def client(original_function):
|
|||
if (
|
||||
kwargs.get("max_tokens", None) is not None
|
||||
and model is not None
|
||||
and litellm.drop_params
|
||||
and litellm.modify_params
|
||||
== True # user is okay with params being modified
|
||||
and (
|
||||
call_type == CallTypes.acompletion.value
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue