mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 10:44:24 +00:00
fix(router.py): fix caching for tracking cooldowns + usage
This commit is contained in:
parent
94c1d71b2c
commit
61fc76a8c4
5 changed files with 148 additions and 75 deletions
|
@ -194,7 +194,7 @@ def run_server(host, port, api_base, api_version, model, alias, add_key, headers
|
|||
"role": "user",
|
||||
"content": "this is a test request, write a short poem"
|
||||
}
|
||||
])
|
||||
], max_tokens=256)
|
||||
click.echo(f'\nLiteLLM: response from proxy {response}')
|
||||
|
||||
print("\n Making streaming request to proxy")
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue