litellm-mirror/litellm/llms/llamafile/chat/transformation.py
2025-04-23 20:28:27 +01:00

46 lines
2 KiB
Python

from typing import Optional, Tuple
from litellm.secret_managers.main import get_secret_str
from ...openai.chat.gpt_transformation import OpenAIGPTConfig
class LlamafileChatConfig(OpenAIGPTConfig):
"""LlamafileChatConfig is used to provide configuration for the LlamaFile's chat API."""
@staticmethod
def _resolve_api_key(api_key: Optional[str] = None) -> str:
"""Attempt to ensure that the API key is set, preferring the user-provided key
over the secret manager key (``LLAMAFILE_API_KEY``).
If both are None, a fake API key is returned.
"""
return api_key or get_secret_str("LLAMAFILE_API_KEY") or "fake-api-key" # llamafile does not require an API key
@staticmethod
def _resolve_api_base(api_base: Optional[str] = None) -> Optional[str]:
"""Attempt to ensure that the API base is set, preferring the user-provided key
over the secret manager key (``LLAMAFILE_API_BASE``).
If both are None, a default Llamafile server URL is returned.
See: https://github.com/Mozilla-Ocho/llamafile/blob/bd1bbe9aabb1ee12dbdcafa8936db443c571eb9d/README.md#L61
"""
return api_base or get_secret_str("LLAMAFILE_API_BASE") or "http://127.0.0.1:8080/v1" # type: ignore
def _get_openai_compatible_provider_info(
self,
api_base: Optional[str],
api_key: Optional[str]
) -> Tuple[Optional[str], Optional[str]]:
"""Attempts to ensure that the API base and key are set, preferring user-provided values,
before falling back to secret manager values (``LLAMAFILE_API_BASE`` and ``LLAMAFILE_API_KEY``
respectively).
If an API key cannot be resolved via either method, a fake key is returned. Llamafile
does not require an API key, but the underlying OpenAI library may expect one anyway.
"""
api_base = LlamafileChatConfig._resolve_api_base(api_base)
dynamic_api_key = LlamafileChatConfig._resolve_api_key(api_key)
return api_base, dynamic_api_key