llama-stack-mirror/tests/integration/inference/recordings/d3cdda901f5ae18d0d636419bd15fd4c5c45dfd6b0f35b0a708f10953874f047.json
2025-10-31 00:41:27 +00:00

62 lines
2.6 KiB
JSON
Generated

{
"test_id": "tests/integration/inference/test_openai_completion.py::test_openai_chat_completion_with_tool_choice_none[txt=ollama/llama3.2:3b-instruct-fp16-inference:chat_completion:tool_calling]",
"request": {
"method": "POST",
"url": "http://0.0.0.0:11434/v1/v1/chat/completions",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"messages": [
{
"role": "system",
"content": "Pretend you are a weather assistant."
},
{
"role": "user",
"content": "What's the weather like in San Francisco, CA?"
}
],
"stream": false
},
"endpoint": "/v1/chat/completions",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": {
"__type__": "openai.types.chat.chat_completion.ChatCompletion",
"__data__": {
"id": "rec-d3cdda901f5a",
"choices": [
{
"finish_reason": "stop",
"index": 0,
"logprobs": null,
"message": {
"content": "I'd be happy to provide you with the latest weather information for San Francisco, CA.\n\nCurrently, it's a beautiful day in San Francisco with mostly sunny skies and mild temperatures. The current temperature is around 58\u00b0F (14\u00b0C) making it a perfect day for outdoor activities.\n\nHowever, I must note that San Francisco is known for its unpredictable weather due to its coastal location and proximity to the Pacific Ocean. There's a possibility of scattered clouds and light drizzle later in the day, but overall, it should remain quite pleasant.\n\nHere are the current conditions:\n\n- Temperature: 58\u00b0F (14\u00b0C)\n- Humidity: 60%\n- Wind: Gusty NW winds at 10 mph\n- Precipitation chance: 20% (light drizzle late afternoon)\n\nPlease keep in mind that weather forecasts can change rapidly, and I recommend checking the latest updates for the most accurate information.",
"refusal": null,
"role": "assistant",
"annotations": null,
"audio": null,
"function_call": null,
"tool_calls": null
}
}
],
"created": 0,
"model": "llama3.2:3b-instruct-fp16",
"object": "chat.completion",
"service_tier": null,
"system_fingerprint": "fp_ollama",
"usage": {
"completion_tokens": 184,
"prompt_tokens": 45,
"total_tokens": 229,
"completion_tokens_details": null,
"prompt_tokens_details": null
}
}
},
"is_streaming": false
},
"id_normalization_mapping": {}
}