Krish Dholakia
c0845fec1f
Add OpenAI gpt-4o-transcribe support ( #9517 )
...
* refactor: introduce new transformation config for gpt-4o-transcribe models
* refactor: expose new transformation configs for audio transcription
* ci: fix config yml
* feat(openai/transcriptions): support provider config transformation on openai audio transcriptions
allows gpt-4o and whisper audio transformation to work as expected
* refactor: migrate fireworks ai + deepgram to new transform request pattern
* feat(openai/): working support for gpt-4o-audio-transcribe
* build(model_prices_and_context_window.json): add gpt-4o-transcribe to model cost map
* build(model_prices_and_context_window.json): specify what endpoints are supported for `/audio/transcriptions`
* fix(get_supported_openai_params.py): fix return
* refactor(deepgram/): migrate unit test to deepgram handler
* refactor: cleanup unused imports
* fix(get_supported_openai_params.py): fix linting error
* test: update test
2025-03-26 23:10:25 -07:00
Krish Dholakia
4351c77253
Support Gemini audio token cost tracking + fix openai audio input token cost tracking ( #9535 )
...
* fix(vertex_and_google_ai_studio_gemini.py): log gemini audio tokens in usage object
enables accurate cost tracking
* refactor(vertex_ai/cost_calculator.py): refactor 128k+ token cost calculation to only run if model info has it
Google has moved away from this for gemini-2.0 models
* refactor(vertex_ai/cost_calculator.py): migrate to usage object for more flexible data passthrough
* fix(llm_cost_calc/utils.py): support audio token cost tracking in generic cost per token
enables vertex ai cost tracking to work with audio tokens
* fix(llm_cost_calc/utils.py): default to total prompt tokens if text tokens field not set
* refactor(llm_cost_calc/utils.py): move openai cost tracking to generic cost per token
more consistent behaviour across providers
* test: add unit test for gemini audio token cost calculation
* ci: bump ci config
* test: fix test
2025-03-26 17:26:25 -07:00
Krrish Dholakia
6a0cf3db50
fix(litellm_logging.py): always log the api base
...
Fixes issue where api base missing from spend logs due to refactor
2025-03-24 13:45:39 -07:00
Ishaan Jaff
edfcff47b9
fix bug when unable to look up model info
2025-03-22 18:24:47 -07:00
Ishaan Jaff
1b9f1837cd
_get_file_search_tool_call
2025-03-22 18:04:33 -07:00
Ishaan Jaff
ded612563d
_get_file_search_tool_call
2025-03-22 18:02:22 -07:00
Ishaan Jaff
8766b43a4b
fix StandardBuiltInToolsParams
2025-03-22 17:53:06 -07:00
Ishaan Jaff
0f36b8afc2
chat_completion_response_includes_annotations
2025-03-22 17:36:22 -07:00
Ishaan Jaff
7f807cd6ba
fix code quality check
2025-03-22 17:31:47 -07:00
Ishaan Jaff
f74640cb99
test_openai_responses_api_web_search_cost_tracking
2025-03-22 17:24:29 -07:00
Ishaan Jaff
81033dc8c9
test_openai_responses_api_web_search_cost_tracking
2025-03-22 17:21:41 -07:00
Ishaan Jaff
904795443d
get_cost_for_built_in_tools
2025-03-22 17:06:34 -07:00
Ishaan Jaff
475dfaa156
fixes for web search cost tracking
2025-03-22 16:56:32 -07:00
Ishaan Jaff
bfe3132bb6
StandardBuiltInToolCostTracking
2025-03-22 16:06:47 -07:00
Ishaan Jaff
10da287225
initialize_standard_built_in_tools_params
2025-03-22 16:03:53 -07:00
Ishaan Jaff
792a2d6115
test_is_chunk_non_empty_with_annotations
2025-03-22 11:41:53 -07:00
Ishaan Jaff
dc5bf10009
feat - add openai web search
2025-03-22 10:43:35 -07:00
Ishaan Jaff
9bc5aaee64
fix ModelParamHelper
2025-03-21 21:35:49 -07:00
Ishaan Jaff
c63217e4d4
_get_litellm_supported_transcription_kwargs
2025-03-21 19:20:13 -07:00
Hammad Saeed
732ab55cc8
fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs()
to use proper annotations from TranscriptionCreateParamsNonStreaming
& `TranscriptionCreateParamsStreaming
2025-03-21 16:55:35 -07:00
Hammad Saeed
3b9dd9be36
fix(model_param_helper.py): change TranscriptionCreateParams.__annotations__ to TranscriptionCreateParams.__dict__ to clean logging error // handle typeddict
2025-03-21 16:30:46 -07:00
Ishaan Jaff
dae5c24b55
get_chat_completion_prompt
2025-03-19 20:50:15 -07:00
Ishaan Jaff
a49bd460e0
get_custom_logger_for_prompt_management
2025-03-19 17:38:24 -07:00
Ishaan Jaff
805542005a
CustomPromptManagement
2025-03-19 16:38:59 -07:00
Ishaan Jaff
730f61f476
get_custom_loggers_for_type
2025-03-19 16:36:41 -07:00
Ishaan Jaff
f5ef0c3cb7
fix code quality checks
2025-03-18 22:34:43 -07:00
Ishaan Jaff
0f2e095b6b
_arize_otel_logger
2025-03-18 22:19:51 -07:00
Ishaan Jaff
57e5c94360
Merge branch 'main' into litellm_arize_dynamic_logging
2025-03-18 22:13:35 -07:00
Krrish Dholakia
a34cc2031d
fix(response_metadata.py): log the litellm_model_name
...
make it easier to track the model sent to the provider
2025-03-18 17:46:33 -07:00
Krrish Dholakia
453729b6c9
fix(litellm_logging.py): fix update
2025-03-18 17:18:58 -07:00
Ishaan Jaff
f4b7b49e9a
fix arize logging
2025-03-18 15:51:58 -07:00
Krrish Dholakia
2c094c2356
fix(get_llm_provider.py): Fixes https://github.com/BerriAI/litellm/issues/9291
2025-03-18 10:48:28 -07:00
Krish Dholakia
cd5024f3b1
Merge pull request #9333 from BerriAI/litellm_dev_03_17_2025_p2
...
fix(ollama/completions/transformation.py): pass prompt, untemplated o…
2025-03-17 21:48:30 -07:00
Krrish Dholakia
22faf7d232
fix(ollama/completions/transformation.py): pass prompt, untemplated on /completions
request
...
Fixes https://github.com/BerriAI/litellm/issues/6900
2025-03-17 18:35:44 -07:00
Krrish Dholakia
c4b2e0ae3d
fix(streaming_handler.py): support logging complete streaming response on cache hit
2025-03-17 18:10:39 -07:00
Krrish Dholakia
dd9e79adbd
fix(streaming_handler.py): emit deep copy of completed chunk
2025-03-17 17:26:21 -07:00
Krrish Dholakia
a5b497667c
fix(logging_utils.py): revert change
2025-03-16 21:04:41 -07:00
Krrish Dholakia
a99251a4ab
fix(streaming_handler.py): raise stop iteration post-finish reason
2025-03-16 20:40:41 -07:00
Krrish Dholakia
bde9ae8a95
fix(litellm_logging.py): remove unused import
2025-03-16 20:24:27 -07:00
Krrish Dholakia
c0a76427d2
fix(streaming_handler.py): pass complete streaming response on completion
2025-03-16 20:22:12 -07:00
Krrish Dholakia
08b297230e
fix(streaming_handler.py): return model response on finished chunk
2025-03-16 13:05:46 -07:00
Krrish Dholakia
612d5a284d
refactor(litellm_logging.py): delegate returning a complete response to the streaming_handler
...
Removes incorrect logic for calculating complete streaming response from litellm logging
2025-03-15 09:55:33 -07:00
Krrish Dholakia
dd2c980d5b
fix(utils.py): Prevents final chunk w/ usage from being ignored
...
Fixes https://github.com/BerriAI/litellm/issues/7112
2025-03-15 09:12:14 -07:00
Krrish Dholakia
a9dceacc1b
fix(factory.py): reduce ollama pt LOC < 50
2025-03-14 21:10:05 -07:00
Krish Dholakia
59fd58643b
Merge pull request #9261 from briandevvn/fix_ollama_pt
...
Fix "system" role has become unacceptable in ollama
2025-03-14 20:13:28 -07:00
Krrish Dholakia
f089b1e23f
feat(endpoints.py): support adding credentials by model id
...
Allows user to reuse existing model credentials
2025-03-14 12:32:32 -07:00
Krrish Dholakia
605a4d1121
feat(endpoints.py): enable retrieving existing credentials by model name
...
Enables reusing existing credentials
2025-03-14 12:02:50 -07:00
Brian Dev
12db28b0af
Support 'system' role ollama
2025-03-15 00:55:18 +07:00
Ishaan Jaff
276a7089df
Merge pull request #9220 from BerriAI/litellm_qa_responses_api
...
[Fixes] Responses API - allow /responses and subpaths as LLM API route + Add exception mapping for responses API
2025-03-13 21:36:59 -07:00
Ishaan Jaff
7827c275ba
exception_type
2025-03-13 20:09:32 -07:00