Ishaan Jaff
|
edfcff47b9
|
fix bug when unable to look up model info
|
2025-03-22 18:24:47 -07:00 |
|
Ishaan Jaff
|
1b9f1837cd
|
_get_file_search_tool_call
|
2025-03-22 18:04:33 -07:00 |
|
Ishaan Jaff
|
ded612563d
|
_get_file_search_tool_call
|
2025-03-22 18:02:22 -07:00 |
|
Ishaan Jaff
|
8766b43a4b
|
fix StandardBuiltInToolsParams
|
2025-03-22 17:53:06 -07:00 |
|
Ishaan Jaff
|
0f36b8afc2
|
chat_completion_response_includes_annotations
|
2025-03-22 17:36:22 -07:00 |
|
Ishaan Jaff
|
7f807cd6ba
|
fix code quality check
|
2025-03-22 17:31:47 -07:00 |
|
Ishaan Jaff
|
f74640cb99
|
test_openai_responses_api_web_search_cost_tracking
|
2025-03-22 17:24:29 -07:00 |
|
Ishaan Jaff
|
81033dc8c9
|
test_openai_responses_api_web_search_cost_tracking
|
2025-03-22 17:21:41 -07:00 |
|
Ishaan Jaff
|
904795443d
|
get_cost_for_built_in_tools
|
2025-03-22 17:06:34 -07:00 |
|
Ishaan Jaff
|
475dfaa156
|
fixes for web search cost tracking
|
2025-03-22 16:56:32 -07:00 |
|
Ishaan Jaff
|
bfe3132bb6
|
StandardBuiltInToolCostTracking
|
2025-03-22 16:06:47 -07:00 |
|
Ishaan Jaff
|
10da287225
|
initialize_standard_built_in_tools_params
|
2025-03-22 16:03:53 -07:00 |
|
Ishaan Jaff
|
792a2d6115
|
test_is_chunk_non_empty_with_annotations
|
2025-03-22 11:41:53 -07:00 |
|
Ishaan Jaff
|
dc5bf10009
|
feat - add openai web search
|
2025-03-22 10:43:35 -07:00 |
|
Ishaan Jaff
|
9bc5aaee64
|
fix ModelParamHelper
|
2025-03-21 21:35:49 -07:00 |
|
Ishaan Jaff
|
c63217e4d4
|
_get_litellm_supported_transcription_kwargs
|
2025-03-21 19:20:13 -07:00 |
|
Hammad Saeed
|
732ab55cc8
|
fix(model_param_helper.py): update _get_litellm_supported_transcription_kwargs() to use proper annotations from TranscriptionCreateParamsNonStreaming & `TranscriptionCreateParamsStreaming
|
2025-03-21 16:55:35 -07:00 |
|
Hammad Saeed
|
3b9dd9be36
|
fix(model_param_helper.py): change TranscriptionCreateParams.__annotations__ to TranscriptionCreateParams.__dict__ to clean logging error // handle typeddict
|
2025-03-21 16:30:46 -07:00 |
|
Ishaan Jaff
|
dae5c24b55
|
get_chat_completion_prompt
|
2025-03-19 20:50:15 -07:00 |
|
Ishaan Jaff
|
a49bd460e0
|
get_custom_logger_for_prompt_management
|
2025-03-19 17:38:24 -07:00 |
|
Ishaan Jaff
|
805542005a
|
CustomPromptManagement
|
2025-03-19 16:38:59 -07:00 |
|
Ishaan Jaff
|
730f61f476
|
get_custom_loggers_for_type
|
2025-03-19 16:36:41 -07:00 |
|
Ishaan Jaff
|
f5ef0c3cb7
|
fix code quality checks
|
2025-03-18 22:34:43 -07:00 |
|
Ishaan Jaff
|
0f2e095b6b
|
_arize_otel_logger
|
2025-03-18 22:19:51 -07:00 |
|
Ishaan Jaff
|
57e5c94360
|
Merge branch 'main' into litellm_arize_dynamic_logging
|
2025-03-18 22:13:35 -07:00 |
|
Krrish Dholakia
|
a34cc2031d
|
fix(response_metadata.py): log the litellm_model_name
make it easier to track the model sent to the provider
|
2025-03-18 17:46:33 -07:00 |
|
Krrish Dholakia
|
453729b6c9
|
fix(litellm_logging.py): fix update
|
2025-03-18 17:18:58 -07:00 |
|
Ishaan Jaff
|
f4b7b49e9a
|
fix arize logging
|
2025-03-18 15:51:58 -07:00 |
|
Krrish Dholakia
|
2c094c2356
|
fix(get_llm_provider.py): Fixes https://github.com/BerriAI/litellm/issues/9291
|
2025-03-18 10:48:28 -07:00 |
|
Krish Dholakia
|
cd5024f3b1
|
Merge pull request #9333 from BerriAI/litellm_dev_03_17_2025_p2
fix(ollama/completions/transformation.py): pass prompt, untemplated o…
|
2025-03-17 21:48:30 -07:00 |
|
Krrish Dholakia
|
22faf7d232
|
fix(ollama/completions/transformation.py): pass prompt, untemplated on /completions request
Fixes https://github.com/BerriAI/litellm/issues/6900
|
2025-03-17 18:35:44 -07:00 |
|
Krrish Dholakia
|
c4b2e0ae3d
|
fix(streaming_handler.py): support logging complete streaming response on cache hit
|
2025-03-17 18:10:39 -07:00 |
|
Krrish Dholakia
|
dd9e79adbd
|
fix(streaming_handler.py): emit deep copy of completed chunk
|
2025-03-17 17:26:21 -07:00 |
|
Krrish Dholakia
|
a5b497667c
|
fix(logging_utils.py): revert change
|
2025-03-16 21:04:41 -07:00 |
|
Krrish Dholakia
|
a99251a4ab
|
fix(streaming_handler.py): raise stop iteration post-finish reason
|
2025-03-16 20:40:41 -07:00 |
|
Krrish Dholakia
|
bde9ae8a95
|
fix(litellm_logging.py): remove unused import
|
2025-03-16 20:24:27 -07:00 |
|
Krrish Dholakia
|
c0a76427d2
|
fix(streaming_handler.py): pass complete streaming response on completion
|
2025-03-16 20:22:12 -07:00 |
|
Krrish Dholakia
|
08b297230e
|
fix(streaming_handler.py): return model response on finished chunk
|
2025-03-16 13:05:46 -07:00 |
|
Krrish Dholakia
|
612d5a284d
|
refactor(litellm_logging.py): delegate returning a complete response to the streaming_handler
Removes incorrect logic for calculating complete streaming response from litellm logging
|
2025-03-15 09:55:33 -07:00 |
|
Krrish Dholakia
|
dd2c980d5b
|
fix(utils.py): Prevents final chunk w/ usage from being ignored
Fixes https://github.com/BerriAI/litellm/issues/7112
|
2025-03-15 09:12:14 -07:00 |
|
Krrish Dholakia
|
a9dceacc1b
|
fix(factory.py): reduce ollama pt LOC < 50
|
2025-03-14 21:10:05 -07:00 |
|
Krish Dholakia
|
59fd58643b
|
Merge pull request #9261 from briandevvn/fix_ollama_pt
Fix "system" role has become unacceptable in ollama
|
2025-03-14 20:13:28 -07:00 |
|
Krrish Dholakia
|
f089b1e23f
|
feat(endpoints.py): support adding credentials by model id
Allows user to reuse existing model credentials
|
2025-03-14 12:32:32 -07:00 |
|
Krrish Dholakia
|
605a4d1121
|
feat(endpoints.py): enable retrieving existing credentials by model name
Enables reusing existing credentials
|
2025-03-14 12:02:50 -07:00 |
|
Brian Dev
|
12db28b0af
|
Support 'system' role ollama
|
2025-03-15 00:55:18 +07:00 |
|
Ishaan Jaff
|
276a7089df
|
Merge pull request #9220 from BerriAI/litellm_qa_responses_api
[Fixes] Responses API - allow /responses and subpaths as LLM API route + Add exception mapping for responses API
|
2025-03-13 21:36:59 -07:00 |
|
Ishaan Jaff
|
7827c275ba
|
exception_type
|
2025-03-13 20:09:32 -07:00 |
|
Sunny Wan
|
f9a5109203
|
Merge branch 'BerriAI:main' into main
|
2025-03-13 19:37:22 -04:00 |
|
Ishaan Jaff
|
15d618f5b1
|
Add exception mapping for responses API
|
2025-03-13 15:57:58 -07:00 |
|
Ishaan Jaff
|
1ee6b7852f
|
fix exception_type
|
2025-03-13 15:33:17 -07:00 |
|