working responses_api

This commit is contained in:
Ishaan Jaff 2025-03-12 17:26:12 -07:00
parent 7546dfde41
commit 98718f6f09
4 changed files with 16 additions and 11 deletions

View file

@ -62,10 +62,16 @@ def _get_metadata_variable_name(request: Request) -> str:
"""
if RouteChecks._is_assistants_api_request(request):
return "litellm_metadata"
if "batches" in request.url.path:
return "litellm_metadata"
if "/v1/messages" in request.url.path:
# anthropic API has a field called metadata
LITELLM_METADATA_ROUTES = [
"batches" "/v1/messages",
"responses",
]
if any(
[
litellm_metadata_route in request.url.path
for litellm_metadata_route in LITELLM_METADATA_ROUTES
]
):
return "litellm_metadata"
else:
return "metadata"

View file

@ -1,10 +1,6 @@
model_list:
- model_name: thinking-us.anthropic.claude-3-7-sonnet-20250219-v1:0
- model_name: gpt-4o
litellm_params:
model: bedrock/us.anthropic.claude-3-7-sonnet-20250219-v1:0
thinking: {"type": "enabled", "budget_tokens": 1024}
max_tokens: 1080
merge_reasoning_content_in_choices: true
model: gpt-4o

View file

@ -233,6 +233,7 @@ from litellm.proxy.pass_through_endpoints.pass_through_endpoints import (
router as pass_through_router,
)
from litellm.proxy.rerank_endpoints.endpoints import router as rerank_router
from litellm.proxy.response_api_endpoints.endpoints import router as response_router
from litellm.proxy.route_llm_request import route_request
from litellm.proxy.spend_tracking.spend_management_endpoints import (
router as spend_management_router,
@ -8390,6 +8391,7 @@ async def get_routes():
app.include_router(router)
app.include_router(response_router)
app.include_router(batches_router)
app.include_router(rerank_router)
app.include_router(fine_tuning_router)

View file

@ -5,7 +5,6 @@ from litellm._logging import verbose_proxy_logger
from litellm.proxy._types import *
from litellm.proxy.auth.user_api_key_auth import UserAPIKeyAuth, user_api_key_auth
from litellm.proxy.common_request_processing import ProxyBaseLLMRequestProcessing
from litellm.proxy.proxy_server import _read_request_body, select_data_generator
router = APIRouter()
@ -44,10 +43,12 @@ async def responses_api(
```
"""
from litellm.proxy.proxy_server import (
_read_request_body,
general_settings,
llm_router,
proxy_config,
proxy_logging_obj,
select_data_generator,
user_api_base,
user_max_tokens,
user_model,