mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
working responses_api
This commit is contained in:
parent
7546dfde41
commit
98718f6f09
4 changed files with 16 additions and 11 deletions
|
@ -62,10 +62,16 @@ def _get_metadata_variable_name(request: Request) -> str:
|
||||||
"""
|
"""
|
||||||
if RouteChecks._is_assistants_api_request(request):
|
if RouteChecks._is_assistants_api_request(request):
|
||||||
return "litellm_metadata"
|
return "litellm_metadata"
|
||||||
if "batches" in request.url.path:
|
LITELLM_METADATA_ROUTES = [
|
||||||
return "litellm_metadata"
|
"batches" "/v1/messages",
|
||||||
if "/v1/messages" in request.url.path:
|
"responses",
|
||||||
# anthropic API has a field called metadata
|
]
|
||||||
|
if any(
|
||||||
|
[
|
||||||
|
litellm_metadata_route in request.url.path
|
||||||
|
for litellm_metadata_route in LITELLM_METADATA_ROUTES
|
||||||
|
]
|
||||||
|
):
|
||||||
return "litellm_metadata"
|
return "litellm_metadata"
|
||||||
else:
|
else:
|
||||||
return "metadata"
|
return "metadata"
|
||||||
|
|
|
@ -1,10 +1,6 @@
|
||||||
model_list:
|
model_list:
|
||||||
- model_name: thinking-us.anthropic.claude-3-7-sonnet-20250219-v1:0
|
- model_name: gpt-4o
|
||||||
litellm_params:
|
litellm_params:
|
||||||
model: bedrock/us.anthropic.claude-3-7-sonnet-20250219-v1:0
|
model: gpt-4o
|
||||||
thinking: {"type": "enabled", "budget_tokens": 1024}
|
|
||||||
max_tokens: 1080
|
|
||||||
merge_reasoning_content_in_choices: true
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -233,6 +233,7 @@ from litellm.proxy.pass_through_endpoints.pass_through_endpoints import (
|
||||||
router as pass_through_router,
|
router as pass_through_router,
|
||||||
)
|
)
|
||||||
from litellm.proxy.rerank_endpoints.endpoints import router as rerank_router
|
from litellm.proxy.rerank_endpoints.endpoints import router as rerank_router
|
||||||
|
from litellm.proxy.response_api_endpoints.endpoints import router as response_router
|
||||||
from litellm.proxy.route_llm_request import route_request
|
from litellm.proxy.route_llm_request import route_request
|
||||||
from litellm.proxy.spend_tracking.spend_management_endpoints import (
|
from litellm.proxy.spend_tracking.spend_management_endpoints import (
|
||||||
router as spend_management_router,
|
router as spend_management_router,
|
||||||
|
@ -8390,6 +8391,7 @@ async def get_routes():
|
||||||
|
|
||||||
|
|
||||||
app.include_router(router)
|
app.include_router(router)
|
||||||
|
app.include_router(response_router)
|
||||||
app.include_router(batches_router)
|
app.include_router(batches_router)
|
||||||
app.include_router(rerank_router)
|
app.include_router(rerank_router)
|
||||||
app.include_router(fine_tuning_router)
|
app.include_router(fine_tuning_router)
|
||||||
|
|
|
@ -5,7 +5,6 @@ from litellm._logging import verbose_proxy_logger
|
||||||
from litellm.proxy._types import *
|
from litellm.proxy._types import *
|
||||||
from litellm.proxy.auth.user_api_key_auth import UserAPIKeyAuth, user_api_key_auth
|
from litellm.proxy.auth.user_api_key_auth import UserAPIKeyAuth, user_api_key_auth
|
||||||
from litellm.proxy.common_request_processing import ProxyBaseLLMRequestProcessing
|
from litellm.proxy.common_request_processing import ProxyBaseLLMRequestProcessing
|
||||||
from litellm.proxy.proxy_server import _read_request_body, select_data_generator
|
|
||||||
|
|
||||||
router = APIRouter()
|
router = APIRouter()
|
||||||
|
|
||||||
|
@ -44,10 +43,12 @@ async def responses_api(
|
||||||
```
|
```
|
||||||
"""
|
"""
|
||||||
from litellm.proxy.proxy_server import (
|
from litellm.proxy.proxy_server import (
|
||||||
|
_read_request_body,
|
||||||
general_settings,
|
general_settings,
|
||||||
llm_router,
|
llm_router,
|
||||||
proxy_config,
|
proxy_config,
|
||||||
proxy_logging_obj,
|
proxy_logging_obj,
|
||||||
|
select_data_generator,
|
||||||
user_api_base,
|
user_api_base,
|
||||||
user_max_tokens,
|
user_max_tokens,
|
||||||
user_model,
|
user_model,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue