forked from phoenix/litellm-mirror
* fix(caching_handler.py): handle positional arguments in add cache logic Fixes https://github.com/BerriAI/litellm/issues/6264 * feat(litellm_pre_call_utils.py): allow forwarding openai org id to backend client https://github.com/BerriAI/litellm/issues/6237 * docs(configs.md): add 'forward_openai_org_id' to docs * fix(proxy_server.py): return model info if user_model is set Fixes https://github.com/BerriAI/litellm/issues/6233 * fix(hosted_vllm/chat/transformation.py): don't set tools unless non-none * fix(openai.py): improve debug log for openai 'str' error Addresses https://github.com/BerriAI/litellm/issues/6272 * fix(proxy_server.py): fix linting error * fix(proxy_server.py): fix linting errors * test: skip WIP test * docs(openai.md): add docs on passing openai org id from client to openai |
||
---|---|---|
.. | ||
conftest.py | ||
Readme.md | ||
test_anthropic_completion.py | ||
test_databricks.py | ||
test_fireworks_ai_translation.py | ||
test_max_completion_tokens.py | ||
test_nvidia_nim.py | ||
test_openai_o1.py | ||
test_optional_params.py | ||
test_prompt_caching.py | ||
test_supports_vision.py | ||
test_vertex.py |
More tests under litellm/litellm/tests/*
.