chore: fix: integration tests failures marked as successful (#3039)

This commit is contained in:
ehhuang 2025-08-04 17:06:28 -07:00 committed by GitHub
parent eac1e0c7d4
commit bb6b6041d6
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
20 changed files with 8449 additions and 871 deletions

View file

@ -22,15 +22,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.522189Z",
"created_at": "2025-08-04T22:55:05.685988Z",
"done": true,
"done_reason": "stop",
"total_duration": 2392200042,
"load_duration": 150986167,
"total_duration": 14128980625,
"load_duration": 7220159208,
"prompt_eval_count": 18,
"prompt_eval_duration": 90000000,
"prompt_eval_duration": 4658000000,
"eval_count": 43,
"eval_duration": 2150000000,
"eval_duration": 2224000000,
"response": " _______.\n\nThe best answer is blue. The traditional nursery rhyme goes like this:\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you! (Or something similar.)",
"thinking": null,
"context": null

View file

@ -20,15 +20,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.076962Z",
"created_at": "2025-08-04T22:55:11.15982Z",
"done": true,
"done_reason": "stop",
"total_duration": 321102500,
"load_duration": 20223000,
"total_duration": 498612042,
"load_duration": 71411834,
"prompt_eval_count": 23,
"prompt_eval_duration": 72000000,
"prompt_eval_duration": 102000000,
"eval_count": 6,
"eval_duration": 227000000,
"eval_duration": 323000000,
"response": "Humans live on Earth.",
"thinking": null,
"context": null

View file

@ -21,7 +21,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.287573Z",
"created_at": "2025-08-04T22:55:11.338232Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -39,7 +39,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.341788Z",
"created_at": "2025-08-04T22:55:11.39419Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -57,7 +57,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.388655Z",
"created_at": "2025-08-04T22:55:11.445346Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -75,7 +75,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.43929Z",
"created_at": "2025-08-04T22:55:11.496701Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -93,7 +93,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.485696Z",
"created_at": "2025-08-04T22:55:11.546804Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -111,7 +111,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.533286Z",
"created_at": "2025-08-04T22:55:11.601009Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -129,7 +129,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.578007Z",
"created_at": "2025-08-04T22:55:11.652788Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -147,7 +147,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.623059Z",
"created_at": "2025-08-04T22:55:11.703325Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -165,7 +165,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.669478Z",
"created_at": "2025-08-04T22:55:11.754033Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -183,7 +183,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.715842Z",
"created_at": "2025-08-04T22:55:11.804654Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -201,15 +201,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.760345Z",
"created_at": "2025-08-04T22:55:11.854841Z",
"done": true,
"done_reason": "stop",
"total_duration": 624960709,
"load_duration": 59848000,
"total_duration": 652371000,
"load_duration": 42086042,
"prompt_eval_count": 26,
"prompt_eval_duration": 87000000,
"prompt_eval_duration": 78000000,
"eval_count": 11,
"eval_duration": 476000000,
"eval_duration": 531000000,
"response": "",
"thinking": null,
"context": null

View file

@ -67,15 +67,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:11.058922Z",
"created_at": "2025-08-04T22:55:40.583477Z",
"done": true,
"done_reason": "stop",
"total_duration": 3810519000,
"load_duration": 148409834,
"total_duration": 3928481500,
"load_duration": 151903250,
"prompt_eval_count": 259,
"prompt_eval_duration": 635000000,
"prompt_eval_duration": 468000000,
"eval_count": 60,
"eval_duration": 3025000000,
"eval_duration": 3306000000,
"response": "{\n \"first_name\": \"Michael\",\n \"last_name\": \"Jordan\",\n \"year_of_birth\": 1963,\n \"nba_stats\": {\n \"year_for_draft\": 1984,\n \"num_seasons_in_nba\": 15\n }\n}",
"thinking": null,
"context": null

View file

@ -14,7 +14,7 @@
"models": [
{
"model": "nomic-embed-text:latest",
"modified_at": "2025-08-01T16:13:37.801531-07:00",
"modified_at": "2025-08-04T15:54:50.584357-07:00",
"digest": "0a109f422b47e3a30ba2b10eca18548e944e8a23073ee3f3e947efcf3c45e59f",
"size": 274302450,
"details": {

View file

@ -0,0 +1,221 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/generate",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"raw": true,
"prompt": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You have access to functions, but you should only use them if they are required.\nYou are an expert in composing functions. You are given a question and a set of possible functions.\nBased on the question, you may or may not need to make one function/tool call to achieve the purpose.\n\nIf you decide to invoke any of the function(s), you MUST put it in the format of [func_name1(params_name1=params_value1, params_name2=params_value2...), func_name2(params)]\nIf you decide to invoke a function, you SHOULD NOT include any other text in the response. besides the function call in the above format.\nFor a boolean parameter, be sure to use `True` or `False` (capitalized) for the value.\n\n\nHere is a list of functions in JSON format that you can invoke.\n\n[\n {\n \"name\": \"get_weather\",\n \"description\": \"Get the current weather\",\n \"parameters\": {\n \"type\": \"dict\",\n \"required\": [\"location\"],\n \"properties\": {\n \"location\": {\n \"type\": \"string\",\n \"description\": \"The city and state (both required), e.g. San Francisco, CA.\"\n }\n }\n }\n }\n]\n\nYou can answer general questions or invoke tools when necessary.\nIn addition to tool calls, you should also augment your responses by using the tool outputs.\nPretend you are a weather assistant.\nYou MUST use one of the provided functions/tools to answer the user query.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\nWhat's the weather like in San Francisco, CA?<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"options": {
"temperature": 0.0
},
"stream": true
},
"endpoint": "/api/generate",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": [
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.141947Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "[",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.194979Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "get",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.248312Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "_weather",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.301911Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "(location",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.354437Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "=\"",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.406821Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "San",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.457633Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " Francisco",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.507857Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": ",",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.558847Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " CA",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.609969Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "\")]",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.660997Z",
"done": true,
"done_reason": "stop",
"total_duration": 715356542,
"load_duration": 59747500,
"prompt_eval_count": 341,
"prompt_eval_duration": 128000000,
"eval_count": 11,
"eval_duration": 526000000,
"response": "",
"thinking": null,
"context": null
}
}
],
"is_streaming": true
}
}

View file

@ -0,0 +1,39 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/generate",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"raw": true,
"prompt": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You have access to functions, but you should only use them if they are required.\nYou are an expert in composing functions. You are given a question and a set of possible functions.\nBased on the question, you may or may not need to make one function/tool call to achieve the purpose.\n\nIf you decide to invoke any of the function(s), you MUST put it in the format of [func_name1(params_name1=params_value1, params_name2=params_value2...), func_name2(params)]\nIf you decide to invoke a function, you SHOULD NOT include any other text in the response. besides the function call in the above format.\nFor a boolean parameter, be sure to use `True` or `False` (capitalized) for the value.\n\n\nHere is a list of functions in JSON format that you can invoke.\n\n[\n {\n \"name\": \"get_weather\",\n \"description\": \"Get the current weather\",\n \"parameters\": {\n \"type\": \"dict\",\n \"required\": [\"location\"],\n \"properties\": {\n \"location\": {\n \"type\": \"string\",\n \"description\": \"The city and state (both required), e.g. San Francisco, CA.\"\n }\n }\n }\n }\n]\n\nYou can answer general questions or invoke tools when necessary.\nIn addition to tool calls, you should also augment your responses by using the tool outputs.\nPretend you are a weather assistant.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\nWhat's the weather like in San Francisco, CA?<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"options": {
"temperature": 0.0
},
"stream": false
},
"endpoint": "/api/generate",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": {
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.25248Z",
"done": true,
"done_reason": "stop",
"total_duration": 1344654917,
"load_duration": 200585375,
"prompt_eval_count": 326,
"prompt_eval_duration": 564000000,
"eval_count": 11,
"eval_duration": 578000000,
"response": "[get_weather(location=\"San Francisco, CA\")]",
"thinking": null,
"context": null
}
},
"is_streaming": false
}
}

View file

@ -0,0 +1,221 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/generate",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"raw": true,
"prompt": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You have access to functions, but you should only use them if they are required.\nYou are an expert in composing functions. You are given a question and a set of possible functions.\nBased on the question, you may or may not need to make one function/tool call to achieve the purpose.\n\nIf you decide to invoke any of the function(s), you MUST put it in the format of [func_name1(params_name1=params_value1, params_name2=params_value2...), func_name2(params)]\nIf you decide to invoke a function, you SHOULD NOT include any other text in the response. besides the function call in the above format.\nFor a boolean parameter, be sure to use `True` or `False` (capitalized) for the value.\n\n\nHere is a list of functions in JSON format that you can invoke.\n\n[\n {\n \"name\": \"get_weather\",\n \"description\": \"Get the current weather\",\n \"parameters\": {\n \"type\": \"dict\",\n \"required\": [\"location\"],\n \"properties\": {\n \"location\": {\n \"type\": \"string\",\n \"description\": \"The city and state (both required), e.g. San Francisco, CA.\"\n }\n }\n }\n }\n]\n\nYou can answer general questions or invoke tools when necessary.\nIn addition to tool calls, you should also augment your responses by using the tool outputs.\nPretend you are a weather assistant.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\nWhat's the weather like in San Francisco, CA?<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"options": {
"temperature": 0.0
},
"stream": true
},
"endpoint": "/api/generate",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": [
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.354888Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "[",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.427569Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "get",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.486244Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "_weather",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.540455Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "(location",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.594439Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "=\"",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.649837Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "San",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.703358Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " Francisco",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.7553Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": ",",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.807251Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " CA",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.857952Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "\")]",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.918522Z",
"done": true,
"done_reason": "stop",
"total_duration": 647785042,
"load_duration": 26355584,
"prompt_eval_count": 326,
"prompt_eval_duration": 55000000,
"eval_count": 11,
"eval_duration": 557000000,
"response": "",
"thinking": null,
"context": null
}
}
],
"is_streaming": true
}
}

View file

@ -21,7 +21,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.7568Z",
"created_at": "2025-08-04T22:55:55.9885Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -39,7 +39,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.808114Z",
"created_at": "2025-08-04T22:55:56.054143Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -57,7 +57,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.862297Z",
"created_at": "2025-08-04T22:55:56.117658Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -75,7 +75,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.910323Z",
"created_at": "2025-08-04T22:55:56.179422Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -93,7 +93,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.95935Z",
"created_at": "2025-08-04T22:55:56.240328Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -111,7 +111,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.013354Z",
"created_at": "2025-08-04T22:55:56.295992Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -129,7 +129,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.07322Z",
"created_at": "2025-08-04T22:55:56.355683Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -147,7 +147,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.131785Z",
"created_at": "2025-08-04T22:55:56.412176Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -165,7 +165,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.180306Z",
"created_at": "2025-08-04T22:55:56.466952Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -183,7 +183,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.230096Z",
"created_at": "2025-08-04T22:55:56.517222Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -201,7 +201,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.281864Z",
"created_at": "2025-08-04T22:55:56.570491Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -219,7 +219,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.335153Z",
"created_at": "2025-08-04T22:55:56.623189Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -237,7 +237,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.388394Z",
"created_at": "2025-08-04T22:55:56.679221Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -255,7 +255,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.441268Z",
"created_at": "2025-08-04T22:55:56.731373Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -273,7 +273,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.493573Z",
"created_at": "2025-08-04T22:55:56.781364Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -291,7 +291,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.543884Z",
"created_at": "2025-08-04T22:55:56.831951Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -309,7 +309,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.592363Z",
"created_at": "2025-08-04T22:55:56.888381Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -327,7 +327,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.640596Z",
"created_at": "2025-08-04T22:55:56.943539Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -345,7 +345,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.687018Z",
"created_at": "2025-08-04T22:55:56.997422Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -363,15 +363,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.73497Z",
"created_at": "2025-08-04T22:55:57.056259Z",
"done": true,
"done_reason": "stop",
"total_duration": 1139022875,
"load_duration": 78012000,
"total_duration": 1289815458,
"load_duration": 119745583,
"prompt_eval_count": 26,
"prompt_eval_duration": 75000000,
"prompt_eval_duration": 98000000,
"eval_count": 20,
"eval_duration": 984000000,
"eval_duration": 1071000000,
"response": "",
"thinking": null,
"context": null

View file

@ -20,15 +20,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.567617Z",
"created_at": "2025-08-04T22:55:55.720345Z",
"done": true,
"done_reason": "stop",
"total_duration": 3565468958,
"load_duration": 31098250,
"total_duration": 3865701084,
"load_duration": 52435459,
"prompt_eval_count": 30,
"prompt_eval_duration": 77000000,
"prompt_eval_duration": 99000000,
"eval_count": 70,
"eval_duration": 3456000000,
"eval_duration": 3712000000,
"response": "The answer is Saturn! Saturn's ring system is one of the most iconic and well-known in our solar system. The rings are made up of ice particles, rock debris, and dust that orbit around the planet due to its gravitational pull.\n\nWould you like to know more about Saturn's rings or is there something else I can help you with?",
"thinking": null,
"context": null

View file

@ -45,15 +45,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:48.732625Z",
"created_at": "2025-08-04T22:55:10.58267Z",
"done": true,
"done_reason": "stop",
"total_duration": 1710455750,
"load_duration": 47968625,
"total_duration": 1981967792,
"load_duration": 63184458,
"prompt_eval_count": 119,
"prompt_eval_duration": 223000000,
"prompt_eval_duration": 259000000,
"eval_count": 29,
"eval_duration": 1437000000,
"eval_duration": 1582000000,
"response": "{ \"name\": \"Michael Jordan\", \"year_born\": \"1963\", \"year_retired\": \"2003\"}\n ",
"thinking": null,
"context": null

File diff suppressed because it is too large Load diff

View file

@ -1,7 +1,7 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/v1/v1/completions",
"url": "http://0.0.0.0:11434/v1/v1/completions",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
@ -15,23 +15,23 @@
"body": {
"__type__": "openai.types.completion.Completion",
"__data__": {
"id": "cmpl-963",
"id": "cmpl-43",
"choices": [
{
"finish_reason": "stop",
"index": 0,
"logprobs": null,
"text": "Blue.\n\nMy answer is \"blue\" because it's a common rhyme that completes the famous nursery rhyme: \"Roses are red, violets are blue.\" This rhyme has been passed down for generations and is often recited as a greeting or a way to express appreciation. The phrase typically rhymes with the word \"blue\", which is why I chose this response."
"text": "Blue.\n\nMy response is based on the traditional English rhyme that pairs the colors of roses (red) with violets in a poetic and somewhat whimsical way. This specific version of the rhyme goes like this:\n\n\"Roses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you.\"\n\nIn modern times, variations of this rhyme can deviate from the original \"blue\" for violets, but in my complete sentence as requested, sticking with a widely recognized completion adds an air of timelessness and familiarity to the phrase."
}
],
"created": 1754081785,
"created": 1754348148,
"model": "llama3.2:3b-instruct-fp16",
"object": "text_completion",
"system_fingerprint": "fp_ollama",
"usage": {
"completion_tokens": 76,
"completion_tokens": 113,
"prompt_tokens": 50,
"total_tokens": 126,
"total_tokens": 163,
"completion_tokens_details": null,
"prompt_tokens_details": null
}

View file

@ -23,7 +23,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.766028Z",
"created_at": "2025-08-04T22:55:05.992185Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -41,7 +41,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.810433Z",
"created_at": "2025-08-04T22:55:06.047726Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -59,7 +59,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.869051Z",
"created_at": "2025-08-04T22:55:06.123375Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -77,7 +77,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.928272Z",
"created_at": "2025-08-04T22:55:06.182233Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -95,7 +95,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.987671Z",
"created_at": "2025-08-04T22:55:06.244655Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -113,7 +113,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.051505Z",
"created_at": "2025-08-04T22:55:06.304777Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -131,7 +131,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.183306Z",
"created_at": "2025-08-04T22:55:06.361584Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -149,7 +149,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.240354Z",
"created_at": "2025-08-04T22:55:06.419647Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -167,7 +167,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.31441Z",
"created_at": "2025-08-04T22:55:06.477037Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -185,7 +185,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.370332Z",
"created_at": "2025-08-04T22:55:06.534717Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -203,7 +203,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.431975Z",
"created_at": "2025-08-04T22:55:06.600289Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -221,7 +221,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.481015Z",
"created_at": "2025-08-04T22:55:06.658769Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -239,7 +239,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.530441Z",
"created_at": "2025-08-04T22:55:06.71323Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -257,7 +257,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.591576Z",
"created_at": "2025-08-04T22:55:06.764206Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -275,7 +275,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.656562Z",
"created_at": "2025-08-04T22:55:06.815428Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -293,7 +293,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.707469Z",
"created_at": "2025-08-04T22:55:06.86906Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -311,7 +311,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.793146Z",
"created_at": "2025-08-04T22:55:06.92191Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -329,7 +329,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.84461Z",
"created_at": "2025-08-04T22:55:06.97464Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -347,7 +347,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.892168Z",
"created_at": "2025-08-04T22:55:07.026686Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -365,7 +365,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.9392Z",
"created_at": "2025-08-04T22:55:07.078382Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -383,7 +383,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.986145Z",
"created_at": "2025-08-04T22:55:07.131717Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -401,7 +401,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.032843Z",
"created_at": "2025-08-04T22:55:07.188206Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -419,7 +419,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.079678Z",
"created_at": "2025-08-04T22:55:07.243218Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -437,7 +437,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.125344Z",
"created_at": "2025-08-04T22:55:07.298542Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -455,7 +455,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.171365Z",
"created_at": "2025-08-04T22:55:07.355167Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -473,7 +473,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.21944Z",
"created_at": "2025-08-04T22:55:07.41078Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -491,7 +491,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.266493Z",
"created_at": "2025-08-04T22:55:07.463639Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -509,7 +509,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.31207Z",
"created_at": "2025-08-04T22:55:07.515619Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -527,7 +527,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.359314Z",
"created_at": "2025-08-04T22:55:07.572461Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -545,7 +545,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.406254Z",
"created_at": "2025-08-04T22:55:07.626345Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -563,7 +563,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.452757Z",
"created_at": "2025-08-04T22:55:07.680673Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -581,7 +581,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.499632Z",
"created_at": "2025-08-04T22:55:07.736803Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -599,7 +599,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.545593Z",
"created_at": "2025-08-04T22:55:07.789556Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -617,7 +617,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.595114Z",
"created_at": "2025-08-04T22:55:07.841142Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -635,7 +635,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.641522Z",
"created_at": "2025-08-04T22:55:07.896607Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -653,7 +653,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.688054Z",
"created_at": "2025-08-04T22:55:07.953628Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -671,7 +671,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.737359Z",
"created_at": "2025-08-04T22:55:08.007575Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -689,7 +689,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.786187Z",
"created_at": "2025-08-04T22:55:08.061895Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -707,7 +707,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.834145Z",
"created_at": "2025-08-04T22:55:08.121698Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -725,7 +725,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.882153Z",
"created_at": "2025-08-04T22:55:08.175866Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -743,7 +743,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.928181Z",
"created_at": "2025-08-04T22:55:08.231661Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -761,7 +761,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.974825Z",
"created_at": "2025-08-04T22:55:08.285188Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -779,15 +779,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:46.022862Z",
"created_at": "2025-08-04T22:55:08.334914Z",
"done": true,
"done_reason": "stop",
"total_duration": 2444514250,
"load_duration": 107849833,
"total_duration": 2543128958,
"load_duration": 133497375,
"prompt_eval_count": 18,
"prompt_eval_duration": 65000000,
"prompt_eval_duration": 62000000,
"eval_count": 43,
"eval_duration": 2270000000,
"eval_duration": 2346000000,
"response": "",
"thinking": null,
"context": null

View file

@ -20,15 +20,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:22.827462Z",
"created_at": "2025-08-04T22:55:57.955211Z",
"done": true,
"done_reason": "stop",
"total_duration": 731625417,
"load_duration": 58932417,
"total_duration": 842946458,
"load_duration": 91343000,
"prompt_eval_count": 386,
"prompt_eval_duration": 609000000,
"prompt_eval_duration": 685000000,
"eval_count": 2,
"eval_duration": 61000000,
"eval_duration": 64000000,
"response": "[]",
"thinking": null,
"context": null

File diff suppressed because it is too large Load diff

View file

@ -16,7 +16,7 @@
"model": "llama3.2:3b-instruct-fp16",
"name": "llama3.2:3b-instruct-fp16",
"digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d",
"expires_at": "2025-08-01T16:18:11.656155-07:00",
"expires_at": "2025-08-04T16:00:57.955349-07:00",
"size": 8581748736,
"size_vram": 8581748736,
"details": {
@ -29,24 +29,6 @@
"parameter_size": "3.2B",
"quantization_level": "F16"
}
},
{
"model": "llama-guard3:1b",
"name": "llama-guard3:1b",
"digest": "494147e06bf99e10dbe67b63a07ac81c162f18ef3341aa3390007ac828571b3b",
"expires_at": "2025-08-01T16:17:54.635270-07:00",
"size": 2770397184,
"size_vram": 2770397184,
"details": {
"parent_model": "",
"format": "gguf",
"family": "llama",
"families": [
"llama"
],
"parameter_size": "1.5B",
"quantization_level": "Q8_0"
}
}
]
}

View file

@ -21,7 +21,7 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:11.920796Z",
"created_at": "2025-08-04T22:55:41.559883Z",
"done": false,
"done_reason": null,
"total_duration": null,
@ -39,15 +39,15 @@
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:11.963674Z",
"created_at": "2025-08-04T22:55:41.619829Z",
"done": true,
"done_reason": "stop",
"total_duration": 837987583,
"load_duration": 158475167,
"total_duration": 915493834,
"load_duration": 167838417,
"prompt_eval_count": 386,
"prompt_eval_duration": 617000000,
"prompt_eval_duration": 683000000,
"eval_count": 2,
"eval_duration": 60000000,
"eval_duration": 63000000,
"response": "",
"thinking": null,
"context": null