{ "test_id": "tests/integration/inference/test_openai_completion.py::test_openai_chat_completion_non_streaming[client_with_models-txt=ollama/llama3.2:3b-instruct-fp16-inference:chat_completion:non_streaming_02]", "request": { "method": "POST", "url": "http://localhost:11434/api/ps", "headers": {}, "body": {}, "endpoint": "/api/ps", "model": "" }, "response": { "body": { "__type__": "ollama._types.ProcessResponse", "__data__": { "models": [ { "model": "llama3.2:3b", "name": "llama3.2:3b", "digest": "a80c4f17acd55265feec403c7aef86be0c25983ab279d83f3bcd3abbcb5b8b72", "expires_at": "2025-10-04T12:20:09.202126-07:00", "size": 3367856128, "size_vram": 3367856128, "details": { "parent_model": "", "format": "gguf", "family": "llama", "families": [ "llama" ], "parameter_size": "3.2B", "quantization_level": "Q4_K_M" }, "context_length": 4096 }, { "model": "llama3.2:3b-instruct-fp16", "name": "llama3.2:3b-instruct-fp16", "digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d", "expires_at": "2025-10-04T11:27:25.408562-07:00", "size": 7919570944, "size_vram": 7919570944, "details": { "parent_model": "", "format": "gguf", "family": "llama", "families": [ "llama" ], "parameter_size": "3.2B", "quantization_level": "F16" }, "context_length": 4096 } ] } }, "is_streaming": false } }