{ "test_id": "tests/integration/inference/test_openai_embeddings.py::test_openai_embeddings_single_string[llama_stack_client-emb=ollama/all-minilm:l6-v2]", "request": { "method": "POST", "url": "http://localhost:11434/api/ps", "headers": {}, "body": {}, "endpoint": "/api/ps", "model": "" }, "response": { "body": { "__type__": "ollama._types.ProcessResponse", "__data__": { "models": [ { "model": "llama3.2:3b-instruct-fp16", "name": "llama3.2:3b-instruct-fp16", "digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d", "expires_at": "2025-10-03T20:06:54.261283-07:00", "size": 7919570944, "size_vram": 7919570944, "details": { "parent_model": "", "format": "gguf", "family": "llama", "families": [ "llama" ], "parameter_size": "3.2B", "quantization_level": "F16" }, "context_length": 4096 }, { "model": "llama-guard3:1b", "name": "llama-guard3:1b", "digest": "494147e06bf99e10dbe67b63a07ac81c162f18ef3341aa3390007ac828571b3b", "expires_at": "2025-10-03T20:06:23.356289-07:00", "size": 2350966784, "size_vram": 2350966784, "details": { "parent_model": "", "format": "gguf", "family": "llama", "families": [ "llama" ], "parameter_size": "1.5B", "quantization_level": "Q8_0" }, "context_length": 4096 } ] } }, "is_streaming": false } }