llama-stack-mirror/tests/integration/recordings/responses/d0ac68cbde69.json

19 lines
335 B
JSON

{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/ps",
"headers": {},
"body": {},
"endpoint": "/api/ps",
"model": ""
},
"response": {
"body": {
"__type__": "ollama._types.ProcessResponse",
"__data__": {
"models": []
}
},
"is_streaming": false
}
}