mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
_get_litellm_supported_transcription_kwargs
This commit is contained in:
parent
478d209a0c
commit
c63217e4d4
2 changed files with 45 additions and 8 deletions
|
@ -1,9 +1,5 @@
|
||||||
from typing import Set
|
from typing import Set
|
||||||
|
|
||||||
from openai.types.audio.transcription_create_params import (
|
|
||||||
TranscriptionCreateParamsNonStreaming,
|
|
||||||
TranscriptionCreateParamsStreaming,
|
|
||||||
)
|
|
||||||
from openai.types.chat.completion_create_params import (
|
from openai.types.chat.completion_create_params import (
|
||||||
CompletionCreateParamsNonStreaming,
|
CompletionCreateParamsNonStreaming,
|
||||||
CompletionCreateParamsStreaming,
|
CompletionCreateParamsStreaming,
|
||||||
|
@ -16,6 +12,7 @@ from openai.types.completion_create_params import (
|
||||||
)
|
)
|
||||||
from openai.types.embedding_create_params import EmbeddingCreateParams
|
from openai.types.embedding_create_params import EmbeddingCreateParams
|
||||||
|
|
||||||
|
from litellm._logging import verbose_logger
|
||||||
from litellm.types.rerank import RerankRequest
|
from litellm.types.rerank import RerankRequest
|
||||||
|
|
||||||
|
|
||||||
|
@ -126,10 +123,19 @@ class ModelParamHelper:
|
||||||
|
|
||||||
This follows the OpenAI API Spec
|
This follows the OpenAI API Spec
|
||||||
"""
|
"""
|
||||||
|
try:
|
||||||
|
from openai.types.audio.transcription_create_params import (
|
||||||
|
TranscriptionCreateParamsNonStreaming,
|
||||||
|
TranscriptionCreateParamsStreaming,
|
||||||
|
)
|
||||||
|
|
||||||
all_transcription_kwargs = set(
|
all_transcription_kwargs = set(
|
||||||
TranscriptionCreateParamsNonStreaming.__annotations__.keys()
|
TranscriptionCreateParamsNonStreaming.__annotations__.keys()
|
||||||
).union(set(TranscriptionCreateParamsStreaming.__annotations__.keys()))
|
).union(set(TranscriptionCreateParamsStreaming.__annotations__.keys()))
|
||||||
return all_transcription_kwargs
|
return all_transcription_kwargs
|
||||||
|
except Exception as e:
|
||||||
|
verbose_logger.warning("Error getting transcription kwargs %s", str(e))
|
||||||
|
return set()
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _get_exclude_kwargs() -> Set[str]:
|
def _get_exclude_kwargs() -> Set[str]:
|
||||||
|
|
|
@ -1426,6 +1426,25 @@
|
||||||
"supports_vision": false,
|
"supports_vision": false,
|
||||||
"supports_prompt_caching": true
|
"supports_prompt_caching": true
|
||||||
},
|
},
|
||||||
|
"azure/gpt-4.5-preview": {
|
||||||
|
"max_tokens": 16384,
|
||||||
|
"max_input_tokens": 128000,
|
||||||
|
"max_output_tokens": 16384,
|
||||||
|
"input_cost_per_token": 0.000075,
|
||||||
|
"output_cost_per_token": 0.00015,
|
||||||
|
"input_cost_per_token_batches": 0.0000375,
|
||||||
|
"output_cost_per_token_batches": 0.000075,
|
||||||
|
"cache_read_input_token_cost": 0.0000375,
|
||||||
|
"litellm_provider": "openai",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_parallel_function_calling": true,
|
||||||
|
"supports_response_schema": true,
|
||||||
|
"supports_vision": true,
|
||||||
|
"supports_prompt_caching": true,
|
||||||
|
"supports_system_messages": true,
|
||||||
|
"supports_tool_choice": true
|
||||||
|
},
|
||||||
"azure/gpt-4o": {
|
"azure/gpt-4o": {
|
||||||
"max_tokens": 16384,
|
"max_tokens": 16384,
|
||||||
"max_input_tokens": 128000,
|
"max_input_tokens": 128000,
|
||||||
|
@ -2091,6 +2110,18 @@
|
||||||
"mode": "chat",
|
"mode": "chat",
|
||||||
"supports_tool_choice": true
|
"supports_tool_choice": true
|
||||||
},
|
},
|
||||||
|
"azure_ai/mistral-small-2503": {
|
||||||
|
"max_tokens": 128000,
|
||||||
|
"max_input_tokens": 128000,
|
||||||
|
"max_output_tokens": 128000,
|
||||||
|
"input_cost_per_token": 0.000001,
|
||||||
|
"output_cost_per_token": 0.000003,
|
||||||
|
"litellm_provider": "azure_ai",
|
||||||
|
"mode": "chat",
|
||||||
|
"supports_function_calling": true,
|
||||||
|
"supports_vision": true,
|
||||||
|
"supports_tool_choice": true
|
||||||
|
},
|
||||||
"azure_ai/mistral-large-2407": {
|
"azure_ai/mistral-large-2407": {
|
||||||
"max_tokens": 4096,
|
"max_tokens": 4096,
|
||||||
"max_input_tokens": 128000,
|
"max_input_tokens": 128000,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue