Litellm ollama refactor (#7162)

* refactor(ollama/): refactor ollama `/api/generate` to use base llm config

Addresses https://github.com/andrewyng/aisuite/issues/113#issuecomment-2512369132

* test: skip unresponsive test

* test(test_secret_manager.py): mark flaky test

* test: fix google sm test

* fix: fix init.py
This commit is contained in:
Krish Dholakia 2024-12-10 21:45:35 -08:00 committed by GitHub
parent 488913c69f
commit 405080396d
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
11 changed files with 322 additions and 235 deletions

View file

@ -3308,6 +3308,12 @@ def get_optional_params( # noqa: PLR0915
optional_params = litellm.OllamaConfig().map_openai_params(
non_default_params=non_default_params,
optional_params=optional_params,
model=model,
drop_params=(
drop_params
if drop_params is not None and isinstance(drop_params, bool)
else False
),
)
elif custom_llm_provider == "ollama_chat":
supported_params = get_supported_openai_params(
@ -6313,6 +6319,8 @@ class ProviderConfigManager:
return litellm.LMStudioChatConfig()
elif litellm.LlmProviders.GALADRIEL == provider:
return litellm.GaladrielChatConfig()
elif litellm.LlmProviders.OLLAMA == provider:
return litellm.OllamaConfig()
return litellm.OpenAIGPTConfig()