fix: OAI compat endpoint for meta reference inference provider

This commit is contained in:
Eric Huang 2025-04-17 11:10:09 -07:00 committed by Eric Huang
parent 8bd6665775
commit c171fc6062
8 changed files with 1184 additions and 44 deletions

View file

@ -12,7 +12,9 @@ from typing import Any
import pytest
from pydantic import BaseModel
from tests.verifications.openai_api.fixtures.fixtures import _load_all_verification_configs
from tests.verifications.openai_api.fixtures.fixtures import (
_load_all_verification_configs,
)
from tests.verifications.openai_api.fixtures.load import load_test_cases
chat_completion_test_cases = load_test_cases("chat_completion")
@ -272,7 +274,6 @@ def test_chat_non_streaming_tool_choice_required(request, openai_client, model,
tool_choice="required", # Force tool call
stream=False,
)
print(response)
assert response.choices[0].message.role == "assistant"
assert len(response.choices[0].message.tool_calls) > 0, "Expected tool call when tool_choice='required'"