mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 19:24:27 +00:00
fix(proxy_server.py): defaulting status to queued
This commit is contained in:
parent
a51915d96f
commit
7fb25ef2dd
1 changed files with 1 additions and 0 deletions
|
@ -599,6 +599,7 @@ async def async_chat_completions(request: Request, task_id: str):
|
||||||
job = redis_job.fetch(id=task_id, connection=redis_connection)
|
job = redis_job.fetch(id=task_id, connection=redis_connection)
|
||||||
print(f"job status: {job.get_status()}")
|
print(f"job status: {job.get_status()}")
|
||||||
result = job.result
|
result = job.result
|
||||||
|
status = "queued"
|
||||||
if result is not None:
|
if result is not None:
|
||||||
status = "finished"
|
status = "finished"
|
||||||
return {"status": status, "result": result}
|
return {"status": status, "result": result}
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue