_get_litellm_supported_transcription_kwargs

This commit is contained in:
Ishaan Jaff 2025-03-21 19:20:13 -07:00
parent 478d209a0c
commit c63217e4d4
2 changed files with 45 additions and 8 deletions

View file

@ -1,9 +1,5 @@
from typing import Set
from openai.types.audio.transcription_create_params import (
TranscriptionCreateParamsNonStreaming,
TranscriptionCreateParamsStreaming,
)
from openai.types.chat.completion_create_params import (
CompletionCreateParamsNonStreaming,
CompletionCreateParamsStreaming,
@ -16,6 +12,7 @@ from openai.types.completion_create_params import (
)
from openai.types.embedding_create_params import EmbeddingCreateParams
from litellm._logging import verbose_logger
from litellm.types.rerank import RerankRequest
@ -126,10 +123,19 @@ class ModelParamHelper:
This follows the OpenAI API Spec
"""
try:
from openai.types.audio.transcription_create_params import (
TranscriptionCreateParamsNonStreaming,
TranscriptionCreateParamsStreaming,
)
all_transcription_kwargs = set(
TranscriptionCreateParamsNonStreaming.__annotations__.keys()
).union(set(TranscriptionCreateParamsStreaming.__annotations__.keys()))
return all_transcription_kwargs
except Exception as e:
verbose_logger.warning("Error getting transcription kwargs %s", str(e))
return set()
@staticmethod
def _get_exclude_kwargs() -> Set[str]:

View file

@ -1426,6 +1426,25 @@
"supports_vision": false,
"supports_prompt_caching": true
},
"azure/gpt-4.5-preview": {
"max_tokens": 16384,
"max_input_tokens": 128000,
"max_output_tokens": 16384,
"input_cost_per_token": 0.000075,
"output_cost_per_token": 0.00015,
"input_cost_per_token_batches": 0.0000375,
"output_cost_per_token_batches": 0.000075,
"cache_read_input_token_cost": 0.0000375,
"litellm_provider": "openai",
"mode": "chat",
"supports_function_calling": true,
"supports_parallel_function_calling": true,
"supports_response_schema": true,
"supports_vision": true,
"supports_prompt_caching": true,
"supports_system_messages": true,
"supports_tool_choice": true
},
"azure/gpt-4o": {
"max_tokens": 16384,
"max_input_tokens": 128000,
@ -2091,6 +2110,18 @@
"mode": "chat",
"supports_tool_choice": true
},
"azure_ai/mistral-small-2503": {
"max_tokens": 128000,
"max_input_tokens": 128000,
"max_output_tokens": 128000,
"input_cost_per_token": 0.000001,
"output_cost_per_token": 0.000003,
"litellm_provider": "azure_ai",
"mode": "chat",
"supports_function_calling": true,
"supports_vision": true,
"supports_tool_choice": true
},
"azure_ai/mistral-large-2407": {
"max_tokens": 4096,
"max_input_tokens": 128000,