Added non-streaming ollama inference impl

This commit is contained in:
Hardik Shah 2024-07-30 18:11:44 -07:00
parent 5b9c05c5dd
commit 0e75e73fa7
4 changed files with 332 additions and 1 deletions

View file

@ -12,6 +12,10 @@ async def get_inference_api_instance(config: InferenceConfig):
from .inference import InferenceImpl
return InferenceImpl(config.impl_config)
elif config.impl_config.impl_type == ImplType.ollama.value:
from .inference import OllamaInference
return OllamaInference(config.impl_config)
from .client import InferenceClient