forked from phoenix/litellm-mirror
Merge pull request #1122 from emsi/main
Fix #1119, no content when streaming.
This commit is contained in:
commit
a6e78497b5
1 changed files with 1 additions and 0 deletions
|
@ -220,6 +220,7 @@ async def ollama_async_streaming(url, data, model_response, encoding, logging_ob
|
||||||
traceback.print_exc()
|
traceback.print_exc()
|
||||||
|
|
||||||
async def ollama_acompletion(url, data, model_response, encoding, logging_obj):
|
async def ollama_acompletion(url, data, model_response, encoding, logging_obj):
|
||||||
|
data["stream"] = False
|
||||||
try:
|
try:
|
||||||
timeout = aiohttp.ClientTimeout(total=600) # 10 minutes
|
timeout = aiohttp.ClientTimeout(total=600) # 10 minutes
|
||||||
async with aiohttp.ClientSession(timeout=timeout) as session:
|
async with aiohttp.ClientSession(timeout=timeout) as session:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue