fix(gemini.py): fix async streaming + add native async completions

This commit is contained in:
Krrish Dholakia 2024-02-19 22:41:36 -08:00
parent 45326c93dc
commit 45eb4a5fcc
6 changed files with 224 additions and 17 deletions

View file

@ -263,6 +263,7 @@ async def acompletion(
or custom_llm_provider == "ollama"
or custom_llm_provider == "ollama_chat"
or custom_llm_provider == "vertex_ai"
or custom_llm_provider == "gemini"
or custom_llm_provider == "sagemaker"
or custom_llm_provider in litellm.openai_compatible_providers
): # currently implemented aiohttp calls for just azure, openai, hf, ollama, vertex ai soon all.