{ "request": { "method": "POST", "url": "https://api.together.xyz/v1/v1/chat/completions", "headers": {}, "body": { "model": "meta-llama/Llama-3.3-70B-Instruct-Turbo-Free", "messages": [ { "role": "user", "content": "Test trace openai 2" } ], "stream": false }, "endpoint": "/v1/chat/completions", "model": "meta-llama/Llama-3.3-70B-Instruct-Turbo-Free" }, "response": { "body": { "__type__": "openai.types.chat.chat_completion.ChatCompletion", "__data__": { "id": "oCfxBri-4Yz4kd-984c2b177fb74ce3", "choices": [ { "finish_reason": "stop", "index": 0, "logprobs": null, "message": { "content": "This conversation has just begun. What would you like to talk about? I can summarize our conversation at the end, if you like.", "refusal": null, "role": "assistant", "annotations": null, "audio": null, "function_call": null, "tool_calls": [] }, "seed": 7149743687991911000 } ], "created": 1758820576, "model": "meta-llama/Llama-3.3-70B-Instruct-Turbo-Free", "object": "chat.completion", "service_tier": null, "system_fingerprint": null, "usage": { "completion_tokens": 28, "prompt_tokens": 41, "total_tokens": 69, "completion_tokens_details": null, "prompt_tokens_details": null, "cached_tokens": 0 }, "prompt": [] } }, "is_streaming": false } }