{ "test_id": "tests/integration/inference/test_openai_completion.py::test_inference_store[openai_client-txt=ollama/llama3.2:3b-instruct-fp16-False]", "request": { "method": "POST", "url": "http://localhost:11434/api/ps", "headers": {}, "body": {}, "endpoint": "/api/ps", "model": "" }, "response": { "body": { "__type__": "ollama._types.ProcessResponse", "__data__": { "models": [ { "model": "llama3.2:3b", "name": "llama3.2:3b", "digest": "a80c4f17acd55265feec403c7aef86be0c25983ab279d83f3bcd3abbcb5b8b72", "expires_at": "2025-10-08T16:14:05.423042-07:00", "size": 3367856128, "size_vram": 3367856128, "details": { "parent_model": "", "format": "gguf", "family": "llama", "families": [ "llama" ], "parameter_size": "3.2B", "quantization_level": "Q4_K_M" } }, { "model": "llama3.2:3b-instruct-fp16", "name": "llama3.2:3b-instruct-fp16", "digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d", "expires_at": "2025-10-08T11:37:37.135763-07:00", "size": 7919570944, "size_vram": 7919570944, "details": { "parent_model": "", "format": "gguf", "family": "llama", "families": [ "llama" ], "parameter_size": "3.2B", "quantization_level": "F16" } }, { "model": "all-minilm:l6-v2", "name": "all-minilm:l6-v2", "digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef", "expires_at": "2025-10-08T11:35:16.599153-07:00", "size": 585846784, "size_vram": 585846784, "details": { "parent_model": "", "format": "gguf", "family": "bert", "families": [ "bert" ], "parameter_size": "23M", "quantization_level": "F16" } } ] } }, "is_streaming": false } }