{ "test_id": "tests/integration/inference/test_openai_completion.py::test_openai_chat_completion_with_tool_choice_none[txt=ollama/llama3.2:3b-instruct-fp16-inference:chat_completion:tool_calling]", "request": { "method": "POST", "url": "http://0.0.0.0:11434/v1/v1/chat/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", "messages": [ { "role": "system", "content": "Pretend you are a weather assistant." }, { "role": "user", "content": "What's the weather like in San Francisco, CA?" } ], "stream": false }, "endpoint": "/v1/chat/completions", "model": "llama3.2:3b-instruct-fp16" }, "response": { "body": { "__type__": "openai.types.chat.chat_completion.ChatCompletion", "__data__": { "id": "rec-d3cdda901f5a", "choices": [ { "finish_reason": "stop", "index": 0, "logprobs": null, "message": { "content": "As your weather assistant today, I've got the latest forecast for you.\n\nSan Francisco, California currently has a partially cloudy sky with a mix of sun and clouds. The temperature is mild, around 68 degrees Fahrenheit (20 degrees Celsius) with a gentle breeze blowing at about 5 miles per hour from the northwest.\n\nHowever, please note that San Francisco's weather is notoriously unpredictable due to its proximity to the Pacific Ocean, so conditions can change rapidly throughout the day.\n\nHere's what you might expect later today:\n\n* High pressure will prevail in the Bay Area, leading to mostly sunny skies with a few puffy clouds by mid-afternoon.\n* Overnight lows are expected to dip to around 55 degrees Fahrenheit (13 degrees Celsius), with a slight chance of scattered showers returning tomorrow morning.\n\nIf you're planning on visiting San Francisco or have outdoor plans today, I recommend checking the forecast again before heading out and packing accordingly!\n\nWould you like more detailed forecasts for specific times, days, or areas in San Francisco?", "refusal": null, "role": "assistant", "annotations": null, "audio": null, "function_call": null, "tool_calls": null } } ], "created": 0, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion", "service_tier": null, "system_fingerprint": "fp_ollama", "usage": { "completion_tokens": 204, "prompt_tokens": 45, "total_tokens": 249, "completion_tokens_details": null, "prompt_tokens_details": null } } }, "is_streaming": false }, "id_normalization_mapping": {} }