litellm/litellm/tests/langfuse.log

173 lines
49 KiB
Text

`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
Creating trace id='8a5b5991-4a10-4dfe-ae1d-52a9195fcde1' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 41, 996862, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
Creating trace id='dba8aaeb-d435-4c82-94f9-9615ace72313' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 41, 998554, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
Creating generation trace_id='dba8aaeb-d435-4c82-94f9-9615ace72313' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 993400) metadata={'litellm_response_cost': None, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-41-993400_chatcmpl-296c9e6a-4142-431c-86ba-c68b550c8005' end_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 995938) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 995938) model='chatgpt-v-2' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=None) prompt_name=None prompt_version=None...
item size 502
item size 930
Creating generation trace_id='8a5b5991-4a10-4dfe-ae1d-52a9195fcde1' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 992764) metadata={'litellm_response_cost': None, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-41-992764_chatcmpl-d449c7f4-47c5-4e46-a9f1-4cd2c1fcf257' end_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 995293) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 995293) model='chatgpt-v-2' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=None) prompt_name=None prompt_version=None...
item size 502
item size 930
Creating trace id='d8d919d4-0c8d-47c5-b1a8-38d884379c74' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 42, 766, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
Creating generation trace_id='d8d919d4-0c8d-47c5-b1a8-38d884379c74' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 994499) metadata={'litellm_response_cost': None, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-41-994499_chatcmpl-5b902172-1250-46e4-bcdd-fd071b951f8c' end_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 996988) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 996988) model='chatgpt-v-2' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=None) prompt_name=None prompt_version=None...
Creating trace id='fac644ff-5f4c-4b45-a55d-a5c4f93ceeec' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 42, 1164, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
item size 502
item size 930
Creating generation trace_id='fac644ff-5f4c-4b45-a55d-a5c4f93ceeec' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 993972) metadata={'litellm_response_cost': None, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-41-993972_chatcmpl-ecc3628b-922b-430d-85b5-068bc01b9cb9' end_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 996217) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 996217) model='chatgpt-v-2' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=None) prompt_name=None prompt_version=None...
item size 502
item size 930
Creating trace id='979fb198-f54d-4e11-ba5c-dec4b5906322' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 42, 2797, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
Creating generation trace_id='979fb198-f54d-4e11-ba5c-dec4b5906322' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 41, 995013) metadata={'litellm_response_cost': None, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-41-995013_chatcmpl-41cfe580-e870-4af9-830d-53472ba99028' end_time=datetime.datetime(2024, 7, 11, 7, 57, 42, 2338) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 42, 2338) model='chatgpt-v-2' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=None) prompt_name=None prompt_version=None...
item size 502
item size 930
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
~0 items in the Langfuse queue
uploading batch of 10 items
uploading data: {'batch': [{'id': 'b4eb919c-efde-48b5-9113-b089b46746c3', 'type': 'trace-create', 'body': {'id': 'dba8aaeb-d435-4c82-94f9-9615ace72313', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 41, 998554, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 41, 998725, tzinfo=datetime.timezone.utc)}, {'id': 'ef605d6e-effe-43eb-894b-57cdc6e28ae4', 'type': 'generation-create', 'body': {'traceId': 'dba8aaeb-d435-4c82-94f9-9615ace72313', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 993400), 'metadata': {'litellm_response_cost': None, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-41-993400_chatcmpl-296c9e6a-4142-431c-86ba-c68b550c8005', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 995938), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 995938), 'model': 'chatgpt-v-2', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 41, 999054, tzinfo=datetime.timezone.utc)}, {'id': '34ab3341-9b8f-45ef-ba32-31d6678c12e1', 'type': 'trace-create', 'body': {'id': '8a5b5991-4a10-4dfe-ae1d-52a9195fcde1', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 41, 996862, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 41, 999535, tzinfo=datetime.timezone.utc)}, {'id': '17e0038d-eb35-47f0-8501-2e9298e25ab3', 'type': 'generation-create', 'body': {'traceId': '8a5b5991-4a10-4dfe-ae1d-52a9195fcde1', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 992764), 'metadata': {'litellm_response_cost': None, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-41-992764_chatcmpl-d449c7f4-47c5-4e46-a9f1-4cd2c1fcf257', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 995293), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 995293), 'model': 'chatgpt-v-2', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 41, 999833, tzinfo=datetime.timezone.utc)}, {'id': '282dc858-0b89-44b2-9946-1b4580a3b99d', 'type': 'trace-create', 'body': {'id': 'd8d919d4-0c8d-47c5-b1a8-38d884379c74', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 766, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 890, tzinfo=datetime.timezone.utc)}, {'id': '8de8e06e-69a5-4636-a1be-7ce1179b7607', 'type': 'generation-create', 'body': {'traceId': 'd8d919d4-0c8d-47c5-b1a8-38d884379c74', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 994499), 'metadata': {'litellm_response_cost': None, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-41-994499_chatcmpl-5b902172-1250-46e4-bcdd-fd071b951f8c', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 996988), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 996988), 'model': 'chatgpt-v-2', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 1121, tzinfo=datetime.timezone.utc)}, {'id': '576825dd-27ff-49b2-8af0-24a6ccd5f853', 'type': 'trace-create', 'body': {'id': 'fac644ff-5f4c-4b45-a55d-a5c4f93ceeec', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 1164, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 1564, tzinfo=datetime.timezone.utc)}, {'id': '4de8b9d4-96c2-492c-b273-f1d114a710da', 'type': 'generation-create', 'body': {'traceId': 'fac644ff-5f4c-4b45-a55d-a5c4f93ceeec', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 993972), 'metadata': {'litellm_response_cost': None, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-41-993972_chatcmpl-ecc3628b-922b-430d-85b5-068bc01b9cb9', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 996217), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 996217), 'model': 'chatgpt-v-2', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 1799, tzinfo=datetime.timezone.utc)}, {'id': 'bdbc23a6-844e-415b-b33f-ede756955c74', 'type': 'trace-create', 'body': {'id': '979fb198-f54d-4e11-ba5c-dec4b5906322', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 2797, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 2921, tzinfo=datetime.timezone.utc)}, {'id': '8b74d44d-cd6d-469f-9077-efea4e82d81b', 'type': 'generation-create', 'body': {'traceId': '979fb198-f54d-4e11-ba5c-dec4b5906322', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 41, 995013), 'metadata': {'litellm_response_cost': None, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': "It's simple to use and easy to get started", 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-41-995013_chatcmpl-41cfe580-e870-4af9-830d-53472ba99028', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 42, 2338), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 42, 2338), 'model': 'chatgpt-v-2', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 42, 3176, tzinfo=datetime.timezone.utc)}], 'metadata': {'batch_size': 10, 'sdk_integration': 'default', 'sdk_name': 'python', 'sdk_version': '2.39.1', 'public_key': 'pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003'}}
making request: {"batch": [{"id": "b4eb919c-efde-48b5-9113-b089b46746c3", "type": "trace-create", "body": {"id": "dba8aaeb-d435-4c82-94f9-9615ace72313", "timestamp": "2024-07-11T14:57:41.998554Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:41.998725Z"}, {"id": "ef605d6e-effe-43eb-894b-57cdc6e28ae4", "type": "generation-create", "body": {"traceId": "dba8aaeb-d435-4c82-94f9-9615ace72313", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:41.993400-07:00", "metadata": {"litellm_response_cost": null, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-41-993400_chatcmpl-296c9e6a-4142-431c-86ba-c68b550c8005", "endTime": "2024-07-11T07:57:41.995938-07:00", "completionStartTime": "2024-07-11T07:57:41.995938-07:00", "model": "chatgpt-v-2", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS"}}, "timestamp": "2024-07-11T14:57:41.999054Z"}, {"id": "34ab3341-9b8f-45ef-ba32-31d6678c12e1", "type": "trace-create", "body": {"id": "8a5b5991-4a10-4dfe-ae1d-52a9195fcde1", "timestamp": "2024-07-11T14:57:41.996862Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:41.999535Z"}, {"id": "17e0038d-eb35-47f0-8501-2e9298e25ab3", "type": "generation-create", "body": {"traceId": "8a5b5991-4a10-4dfe-ae1d-52a9195fcde1", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:41.992764-07:00", "metadata": {"litellm_response_cost": null, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-41-992764_chatcmpl-d449c7f4-47c5-4e46-a9f1-4cd2c1fcf257", "endTime": "2024-07-11T07:57:41.995293-07:00", "completionStartTime": "2024-07-11T07:57:41.995293-07:00", "model": "chatgpt-v-2", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS"}}, "timestamp": "2024-07-11T14:57:41.999833Z"}, {"id": "282dc858-0b89-44b2-9946-1b4580a3b99d", "type": "trace-create", "body": {"id": "d8d919d4-0c8d-47c5-b1a8-38d884379c74", "timestamp": "2024-07-11T14:57:42.000766Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:42.000890Z"}, {"id": "8de8e06e-69a5-4636-a1be-7ce1179b7607", "type": "generation-create", "body": {"traceId": "d8d919d4-0c8d-47c5-b1a8-38d884379c74", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:41.994499-07:00", "metadata": {"litellm_response_cost": null, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-41-994499_chatcmpl-5b902172-1250-46e4-bcdd-fd071b951f8c", "endTime": "2024-07-11T07:57:41.996988-07:00", "completionStartTime": "2024-07-11T07:57:41.996988-07:00", "model": "chatgpt-v-2", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS"}}, "timestamp": "2024-07-11T14:57:42.001121Z"}, {"id": "576825dd-27ff-49b2-8af0-24a6ccd5f853", "type": "trace-create", "body": {"id": "fac644ff-5f4c-4b45-a55d-a5c4f93ceeec", "timestamp": "2024-07-11T14:57:42.001164Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:42.001564Z"}, {"id": "4de8b9d4-96c2-492c-b273-f1d114a710da", "type": "generation-create", "body": {"traceId": "fac644ff-5f4c-4b45-a55d-a5c4f93ceeec", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:41.993972-07:00", "metadata": {"litellm_response_cost": null, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-41-993972_chatcmpl-ecc3628b-922b-430d-85b5-068bc01b9cb9", "endTime": "2024-07-11T07:57:41.996217-07:00", "completionStartTime": "2024-07-11T07:57:41.996217-07:00", "model": "chatgpt-v-2", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS"}}, "timestamp": "2024-07-11T14:57:42.001799Z"}, {"id": "bdbc23a6-844e-415b-b33f-ede756955c74", "type": "trace-create", "body": {"id": "979fb198-f54d-4e11-ba5c-dec4b5906322", "timestamp": "2024-07-11T14:57:42.002797Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:42.002921Z"}, {"id": "8b74d44d-cd6d-469f-9077-efea4e82d81b", "type": "generation-create", "body": {"traceId": "979fb198-f54d-4e11-ba5c-dec4b5906322", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:41.995013-07:00", "metadata": {"litellm_response_cost": null, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "It's simple to use and easy to get started", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-41-995013_chatcmpl-41cfe580-e870-4af9-830d-53472ba99028", "endTime": "2024-07-11T07:57:42.002338-07:00", "completionStartTime": "2024-07-11T07:57:42.002338-07:00", "model": "chatgpt-v-2", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS"}}, "timestamp": "2024-07-11T14:57:42.003176Z"}], "metadata": {"batch_size": 10, "sdk_integration": "default", "sdk_name": "python", "sdk_version": "2.39.1", "public_key": "pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003"}} to https://us.cloud.langfuse.com/api/public/ingestion
received response: {"errors":[],"successes":[{"id":"b4eb919c-efde-48b5-9113-b089b46746c3","status":201},{"id":"ef605d6e-effe-43eb-894b-57cdc6e28ae4","status":201},{"id":"34ab3341-9b8f-45ef-ba32-31d6678c12e1","status":201},{"id":"17e0038d-eb35-47f0-8501-2e9298e25ab3","status":201},{"id":"282dc858-0b89-44b2-9946-1b4580a3b99d","status":201},{"id":"8de8e06e-69a5-4636-a1be-7ce1179b7607","status":201},{"id":"576825dd-27ff-49b2-8af0-24a6ccd5f853","status":201},{"id":"4de8b9d4-96c2-492c-b273-f1d114a710da","status":201},{"id":"bdbc23a6-844e-415b-b33f-ede756955c74","status":201},{"id":"8b74d44d-cd6d-469f-9077-efea4e82d81b","status":201}]}
successfully uploaded batch of 10 items
~0 items in the Langfuse queue
consumer is running...
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
Creating trace id='litellm-test-41e731d8-eec7-484d-a123-3f5369227667' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 43, 673224, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]} output={'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
Creating generation trace_id='litellm-test-41e731d8-eec7-484d-a123-3f5369227667' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 43, 671900) metadata={'litellm_response_cost': 5.4999999999999995e-05, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]} output={'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-43-671900_chatcmpl-84840e13-5f25-494b-a771-43d6dc9f4825' end_time=datetime.datetime(2024, 7, 11, 7, 57, 43, 672638) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 43, 672638) model='gpt-3.5-turbo' model_parameters={'temperature': '0.7', 'stream': False, 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=5.4999999999999995e-05) prompt_name=None prompt_version=None...
item size 497
item size 999
flushing queue
successfully flushed about 0 items.
~0 items in the Langfuse queue
uploading batch of 2 items
uploading data: {'batch': [{'id': 'b37f0713-1089-44ec-99d2-12bd69d6b4ed', 'type': 'trace-create', 'body': {'id': 'litellm-test-41e731d8-eec7-484d-a123-3f5369227667', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 43, 673224, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]}, 'output': {'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 43, 673428, tzinfo=datetime.timezone.utc)}, {'id': '9e0c04ab-30c0-47ec-8507-9877c9bff1c7', 'type': 'generation-create', 'body': {'traceId': 'litellm-test-41e731d8-eec7-484d-a123-3f5369227667', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 43, 671900), 'metadata': {'litellm_response_cost': 5.4999999999999995e-05, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]}, 'output': {'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-43-671900_chatcmpl-84840e13-5f25-494b-a771-43d6dc9f4825', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 43, 672638), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 43, 672638), 'model': 'gpt-3.5-turbo', 'modelParameters': {'temperature': '0.7', 'stream': False, 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>, 'totalCost': 5.4999999999999995e-05}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 43, 673780, tzinfo=datetime.timezone.utc)}], 'metadata': {'batch_size': 2, 'sdk_integration': 'default', 'sdk_name': 'python', 'sdk_version': '2.39.1', 'public_key': 'pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003'}}
making request: {"batch": [{"id": "b37f0713-1089-44ec-99d2-12bd69d6b4ed", "type": "trace-create", "body": {"id": "litellm-test-41e731d8-eec7-484d-a123-3f5369227667", "timestamp": "2024-07-11T14:57:43.673224Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "redacted-by-litellm"}]}, "output": {"content": "redacted-by-litellm", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:43.673428Z"}, {"id": "9e0c04ab-30c0-47ec-8507-9877c9bff1c7", "type": "generation-create", "body": {"traceId": "litellm-test-41e731d8-eec7-484d-a123-3f5369227667", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:43.671900-07:00", "metadata": {"litellm_response_cost": 5.4999999999999995e-05, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "redacted-by-litellm"}]}, "output": {"content": "redacted-by-litellm", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-43-671900_chatcmpl-84840e13-5f25-494b-a771-43d6dc9f4825", "endTime": "2024-07-11T07:57:43.672638-07:00", "completionStartTime": "2024-07-11T07:57:43.672638-07:00", "model": "gpt-3.5-turbo", "modelParameters": {"temperature": "0.7", "stream": false, "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS", "totalCost": 5.4999999999999995e-05}}, "timestamp": "2024-07-11T14:57:43.673780Z"}], "metadata": {"batch_size": 2, "sdk_integration": "default", "sdk_name": "python", "sdk_version": "2.39.1", "public_key": "pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003"}} to https://us.cloud.langfuse.com/api/public/ingestion
~0 items in the Langfuse queue
received response: {"errors":[],"successes":[{"id":"b37f0713-1089-44ec-99d2-12bd69d6b4ed","status":201},{"id":"9e0c04ab-30c0-47ec-8507-9877c9bff1c7","status":201}]}
successfully uploaded batch of 2 items
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
Getting observations... None, None, None, None, litellm-test-41e731d8-eec7-484d-a123-3f5369227667, None, None, None, GENERATION
~0 items in the Langfuse queue
~0 items in the Langfuse queue
consumer is running...
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
flushing queue
successfully flushed about 0 items.
~0 items in the Langfuse queue
Creating trace id='litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 46, 266724, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]} output={'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
item size 497
Creating generation trace_id='litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 46, 11092) metadata={'litellm_response_cost': 4.1e-05, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]} output={'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-46-011092_chatcmpl-bbd13417-0ddd-46d0-a4fc-ded148ff2ae9' end_time=datetime.datetime(2024, 7, 11, 7, 57, 46, 22279) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 46, 13610) model='gpt-3.5-turbo' model_parameters={'temperature': '0.7', 'stream': True, 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=14, output=10, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=4.1e-05) prompt_name=None prompt_version=None...
item size 968
~0 items in the Langfuse queue
uploading batch of 2 items
uploading data: {'batch': [{'id': 'b05c61c3-d1c8-49fa-8bbe-7c9d1a9aba87', 'type': 'trace-create', 'body': {'id': 'litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 46, 266724, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]}, 'output': {'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 46, 267376, tzinfo=datetime.timezone.utc)}, {'id': 'f1e3d54a-800e-40ba-8214-50fa8ba1a40e', 'type': 'generation-create', 'body': {'traceId': 'litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 46, 11092), 'metadata': {'litellm_response_cost': 4.1e-05, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'redacted-by-litellm'}]}, 'output': {'content': 'redacted-by-litellm', 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-46-011092_chatcmpl-bbd13417-0ddd-46d0-a4fc-ded148ff2ae9', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 46, 22279), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 46, 13610), 'model': 'gpt-3.5-turbo', 'modelParameters': {'temperature': '0.7', 'stream': True, 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 14, 'output': 10, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>, 'totalCost': 4.1e-05}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 46, 268553, tzinfo=datetime.timezone.utc)}], 'metadata': {'batch_size': 2, 'sdk_integration': 'default', 'sdk_name': 'python', 'sdk_version': '2.39.1', 'public_key': 'pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003'}}
making request: {"batch": [{"id": "b05c61c3-d1c8-49fa-8bbe-7c9d1a9aba87", "type": "trace-create", "body": {"id": "litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21", "timestamp": "2024-07-11T14:57:46.266724Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "redacted-by-litellm"}]}, "output": {"content": "redacted-by-litellm", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:46.267376Z"}, {"id": "f1e3d54a-800e-40ba-8214-50fa8ba1a40e", "type": "generation-create", "body": {"traceId": "litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:46.011092-07:00", "metadata": {"litellm_response_cost": 4.1e-05, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "redacted-by-litellm"}]}, "output": {"content": "redacted-by-litellm", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-46-011092_chatcmpl-bbd13417-0ddd-46d0-a4fc-ded148ff2ae9", "endTime": "2024-07-11T07:57:46.022279-07:00", "completionStartTime": "2024-07-11T07:57:46.013610-07:00", "model": "gpt-3.5-turbo", "modelParameters": {"temperature": "0.7", "stream": true, "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 14, "output": 10, "unit": "TOKENS", "totalCost": 4.1e-05}}, "timestamp": "2024-07-11T14:57:46.268553Z"}], "metadata": {"batch_size": 2, "sdk_integration": "default", "sdk_name": "python", "sdk_version": "2.39.1", "public_key": "pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003"}} to https://us.cloud.langfuse.com/api/public/ingestion
received response: {"errors":[],"successes":[{"id":"b05c61c3-d1c8-49fa-8bbe-7c9d1a9aba87","status":201},{"id":"f1e3d54a-800e-40ba-8214-50fa8ba1a40e","status":201}]}
successfully uploaded batch of 2 items
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
Getting observations... None, None, None, None, litellm-test-e079602a-d96c-4178-a423-f45d4e8c2d21, None, None, None, GENERATION
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
consumer is running...
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
Creating trace id='litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 48, 311181, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input='redacted-by-litellm' output='redacted-by-litellm' session_id=None release=None version=None metadata=None tags=[] public=None
Creating generation trace_id='litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 48, 309974) metadata={'litellm_response_cost': 5.4999999999999995e-05, 'cache_hit': False} input='redacted-by-litellm' output='redacted-by-litellm' level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-48-309974_chatcmpl-9feff8c8-bddc-45fa-b3e7-3e83768589b4' end_time=datetime.datetime(2024, 7, 11, 7, 57, 48, 310635) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 48, 310635) model='gpt-3.5-turbo' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=5.4999999999999995e-05) prompt_name=None prompt_version=None...
item size 375
item size 860
flushing queue
successfully flushed about 0 items.
~0 items in the Langfuse queue
uploading batch of 2 items
uploading data: {'batch': [{'id': '414f3e3a-35b4-4e24-a64a-e9303de94d0d', 'type': 'trace-create', 'body': {'id': 'litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 48, 311181, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': 'redacted-by-litellm', 'output': 'redacted-by-litellm', 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 48, 311330, tzinfo=datetime.timezone.utc)}, {'id': '890d05df-f7e0-4c1a-ad49-6b1704c36136', 'type': 'generation-create', 'body': {'traceId': 'litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 48, 309974), 'metadata': {'litellm_response_cost': 5.4999999999999995e-05, 'cache_hit': False}, 'input': 'redacted-by-litellm', 'output': 'redacted-by-litellm', 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-48-309974_chatcmpl-9feff8c8-bddc-45fa-b3e7-3e83768589b4', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 48, 310635), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 48, 310635), 'model': 'gpt-3.5-turbo', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>, 'totalCost': 5.4999999999999995e-05}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 48, 311640, tzinfo=datetime.timezone.utc)}], 'metadata': {'batch_size': 2, 'sdk_integration': 'default', 'sdk_name': 'python', 'sdk_version': '2.39.1', 'public_key': 'pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003'}}
~0 items in the Langfuse queue
making request: {"batch": [{"id": "414f3e3a-35b4-4e24-a64a-e9303de94d0d", "type": "trace-create", "body": {"id": "litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4", "timestamp": "2024-07-11T14:57:48.311181Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": "redacted-by-litellm", "output": "redacted-by-litellm", "tags": []}, "timestamp": "2024-07-11T14:57:48.311330Z"}, {"id": "890d05df-f7e0-4c1a-ad49-6b1704c36136", "type": "generation-create", "body": {"traceId": "litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:48.309974-07:00", "metadata": {"litellm_response_cost": 5.4999999999999995e-05, "cache_hit": false}, "input": "redacted-by-litellm", "output": "redacted-by-litellm", "level": "DEFAULT", "id": "time-07-57-48-309974_chatcmpl-9feff8c8-bddc-45fa-b3e7-3e83768589b4", "endTime": "2024-07-11T07:57:48.310635-07:00", "completionStartTime": "2024-07-11T07:57:48.310635-07:00", "model": "gpt-3.5-turbo", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS", "totalCost": 5.4999999999999995e-05}}, "timestamp": "2024-07-11T14:57:48.311640Z"}], "metadata": {"batch_size": 2, "sdk_integration": "default", "sdk_name": "python", "sdk_version": "2.39.1", "public_key": "pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003"}} to https://us.cloud.langfuse.com/api/public/ingestion
~0 items in the Langfuse queue
~0 items in the Langfuse queue
received response: {"errors":[],"successes":[{"id":"414f3e3a-35b4-4e24-a64a-e9303de94d0d","status":201},{"id":"890d05df-f7e0-4c1a-ad49-6b1704c36136","status":201}]}
successfully uploaded batch of 2 items
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
Getting trace litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4
~0 items in the Langfuse queue
~0 items in the Langfuse queue
Getting observations... None, None, None, None, litellm-test-624f61d8-2604-406a-bd85-f7eb9f9878e4, None, None, None, GENERATION
~0 items in the Langfuse queue
`litellm.set_verbose` is deprecated. Please set `os.environ['LITELLM_LOG'] = 'DEBUG'` for debug logs.
Creating trace id='litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be' timestamp=datetime.datetime(2024, 7, 11, 14, 57, 50, 588244, tzinfo=datetime.timezone.utc) name='litellm-acompletion' user_id='langfuse_latency_test_user' input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': 'This is a test response', 'role': 'assistant', 'tool_calls': None, 'function_call': None} session_id=None release=None version=None metadata=None tags=[] public=None
Creating generation trace_id='litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be' name='litellm-acompletion' start_time=datetime.datetime(2024, 7, 11, 7, 57, 50, 585058) metadata={'litellm_response_cost': 5.4999999999999995e-05, 'cache_hit': False} input={'messages': [{'role': 'user', 'content': 'This is a test'}]} output={'content': 'This is a test response', 'role': 'assistant', 'tool_calls': None, 'function_call': None} level=<ObservationLevel.DEFAULT: 'DEFAULT'> status_message=None parent_observation_id=None version=None id='time-07-57-50-585058_chatcmpl-05db6b0b-4847-4b87-a8ef-7d90bef7f81c' end_time=datetime.datetime(2024, 7, 11, 7, 57, 50, 587160) completion_start_time=datetime.datetime(2024, 7, 11, 7, 57, 50, 587160) model='gpt-3.5-turbo' model_parameters={'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'} usage=Usage(input=10, output=20, total=None, unit=<ModelUsageUnit.TOKENS: 'TOKENS'>, input_cost=None, output_cost=None, total_cost=5.4999999999999995e-05) prompt_name=None prompt_version=None...
item size 496
item size 981
flushing queue
successfully flushed about 0 items.
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
uploading batch of 2 items
uploading data: {'batch': [{'id': '27ef4895-21f0-4485-b9e6-083e1d64e5fa', 'type': 'trace-create', 'body': {'id': 'litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be', 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 50, 588244, tzinfo=datetime.timezone.utc), 'name': 'litellm-acompletion', 'userId': 'langfuse_latency_test_user', 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': 'This is a test response', 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'tags': []}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 50, 589416, tzinfo=datetime.timezone.utc)}, {'id': '8f77e83d-d4b4-4890-b37a-bf04611f2ed6', 'type': 'generation-create', 'body': {'traceId': 'litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be', 'name': 'litellm-acompletion', 'startTime': datetime.datetime(2024, 7, 11, 7, 57, 50, 585058), 'metadata': {'litellm_response_cost': 5.4999999999999995e-05, 'cache_hit': False}, 'input': {'messages': [{'role': 'user', 'content': 'This is a test'}]}, 'output': {'content': 'This is a test response', 'role': 'assistant', 'tool_calls': None, 'function_call': None}, 'level': <ObservationLevel.DEFAULT: 'DEFAULT'>, 'id': 'time-07-57-50-585058_chatcmpl-05db6b0b-4847-4b87-a8ef-7d90bef7f81c', 'endTime': datetime.datetime(2024, 7, 11, 7, 57, 50, 587160), 'completionStartTime': datetime.datetime(2024, 7, 11, 7, 57, 50, 587160), 'model': 'gpt-3.5-turbo', 'modelParameters': {'temperature': '0.7', 'max_tokens': 5, 'user': 'langfuse_latency_test_user', 'extra_body': '{}'}, 'usage': {'input': 10, 'output': 20, 'unit': <ModelUsageUnit.TOKENS: 'TOKENS'>, 'totalCost': 5.4999999999999995e-05}}, 'timestamp': datetime.datetime(2024, 7, 11, 14, 57, 50, 590731, tzinfo=datetime.timezone.utc)}], 'metadata': {'batch_size': 2, 'sdk_integration': 'default', 'sdk_name': 'python', 'sdk_version': '2.39.1', 'public_key': 'pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003'}}
making request: {"batch": [{"id": "27ef4895-21f0-4485-b9e6-083e1d64e5fa", "type": "trace-create", "body": {"id": "litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be", "timestamp": "2024-07-11T14:57:50.588244Z", "name": "litellm-acompletion", "userId": "langfuse_latency_test_user", "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "This is a test response", "role": "assistant", "tool_calls": null, "function_call": null}, "tags": []}, "timestamp": "2024-07-11T14:57:50.589416Z"}, {"id": "8f77e83d-d4b4-4890-b37a-bf04611f2ed6", "type": "generation-create", "body": {"traceId": "litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be", "name": "litellm-acompletion", "startTime": "2024-07-11T07:57:50.585058-07:00", "metadata": {"litellm_response_cost": 5.4999999999999995e-05, "cache_hit": false}, "input": {"messages": [{"role": "user", "content": "This is a test"}]}, "output": {"content": "This is a test response", "role": "assistant", "tool_calls": null, "function_call": null}, "level": "DEFAULT", "id": "time-07-57-50-585058_chatcmpl-05db6b0b-4847-4b87-a8ef-7d90bef7f81c", "endTime": "2024-07-11T07:57:50.587160-07:00", "completionStartTime": "2024-07-11T07:57:50.587160-07:00", "model": "gpt-3.5-turbo", "modelParameters": {"temperature": "0.7", "max_tokens": 5, "user": "langfuse_latency_test_user", "extra_body": "{}"}, "usage": {"input": 10, "output": 20, "unit": "TOKENS", "totalCost": 5.4999999999999995e-05}}, "timestamp": "2024-07-11T14:57:50.590731Z"}], "metadata": {"batch_size": 2, "sdk_integration": "default", "sdk_name": "python", "sdk_version": "2.39.1", "public_key": "pk-lf-b3db7e8e-c2f6-4fc7-825c-a541a8fbe003"}} to https://us.cloud.langfuse.com/api/public/ingestion
~0 items in the Langfuse queue
received response: {"errors":[],"successes":[{"id":"27ef4895-21f0-4485-b9e6-083e1d64e5fa","status":201},{"id":"8f77e83d-d4b4-4890-b37a-bf04611f2ed6","status":201}]}
successfully uploaded batch of 2 items
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
Getting trace litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be
~0 items in the Langfuse queue
Getting observations... None, None, None, None, litellm-test-0020e225-3dc4-43d5-b75f-468a042e51be, None, None, None, GENERATION
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
joining 1 consumer threads
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
consumer thread 0 joined
joining 1 consumer threads
~0 items in the Langfuse queue
~0 items in the Langfuse queue
~0 items in the Langfuse queue
consumer thread 0 joined
joining 1 consumer threads
~0 items in the Langfuse queue
~0 items in the Langfuse queue
consumer thread 0 joined
joining 1 consumer threads
~0 items in the Langfuse queue
consumer thread 0 joined