{ "test_id": "tests/integration/batches/test_batches.py::TestBatchesIntegration::test_batch_e2e_completions[txt=ollama/llama3.2:3b-instruct-fp16]", "request": { "method": "POST", "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", "prompt": "Say completions", "max_tokens": 20 }, "endpoint": "/v1/completions", "model": "llama3.2:3b-instruct-fp16" }, "response": { "body": { "__type__": "openai.types.completion.Completion", "__data__": { "id": "rec-92d49675c903", "choices": [ { "finish_reason": "length", "index": 0, "logprobs": null, "text": "What would you like me to say completion about? Would you like me to complete a thought, finish" } ], "created": 0, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", "usage": { "completion_tokens": 20, "prompt_tokens": 28, "total_tokens": 48, "completion_tokens_details": null, "prompt_tokens_details": null } } }, "is_streaming": false }, "id_normalization_mapping": {} }