mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
* build(model_prices_and_context_window.json): add vertex ai gemini-2.5-flash pricing * build(model_prices_and_context_window.json): add gemini reasoning token pricing * fix(vertex_and_google_ai_studio_gemini.py): support counting thinking tokens for gemini allows accurate cost calc * fix(utils.py): add reasoning token cost calc to generic cost calc ensures gemini-2.5-flash cost calculation is accurate * build(model_prices_and_context_window.json): mark gemini-2.5-flash as 'supports_reasoning' * feat(gemini/): support 'thinking' + 'reasoning_effort' params + new unit tests allow controlling thinking effort for gemini-2.5-flash models * test: update unit testing * feat(vertex_and_google_ai_studio_gemini.py): return reasoning content if given in gemini response * test: update model name * fix: fix ruff check * test(test_spend_management_endpoints.py): update tests to be less sensitive to new keys / updates to usage object * fix(vertex_and_google_ai_studio_gemini.py): fix translation |
||
---|---|---|
.. | ||
auth | ||
common_utils | ||
db | ||
experimental/mcp_server | ||
health_endpoints | ||
hooks | ||
management_endpoints | ||
middleware | ||
openai_files_endpoint | ||
pass_through_endpoints | ||
spend_tracking | ||
types_utils | ||
ui_crud_endpoints | ||
test_caching_routes.py | ||
test_common_request_processing.py | ||
test_litellm_pre_call_utils.py | ||
test_proxy_cli.py | ||
test_proxy_server.py | ||
test_route_llm_request.py |