chore: fix: integration tests failures marked as successful (#3039)

This commit is contained in:
ehhuang 2025-08-04 17:06:28 -07:00 committed by GitHub
parent eac1e0c7d4
commit bb6b6041d6
No known key found for this signature in database
GPG key ID: B5690EEEBB952194
20 changed files with 8449 additions and 871 deletions

View file

@ -5,7 +5,7 @@
# This source code is licensed under the terms described in the LICENSE file in # This source code is licensed under the terms described in the LICENSE file in
# the root directory of this source tree. # the root directory of this source tree.
set -e set -euo pipefail
# Integration test runner script for Llama Stack # Integration test runner script for Llama Stack
# This script extracts the integration test logic from GitHub Actions # This script extracts the integration test logic from GitHub Actions

View file

@ -22,15 +22,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.522189Z", "created_at": "2025-08-04T22:55:05.685988Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 2392200042, "total_duration": 14128980625,
"load_duration": 150986167, "load_duration": 7220159208,
"prompt_eval_count": 18, "prompt_eval_count": 18,
"prompt_eval_duration": 90000000, "prompt_eval_duration": 4658000000,
"eval_count": 43, "eval_count": 43,
"eval_duration": 2150000000, "eval_duration": 2224000000,
"response": " _______.\n\nThe best answer is blue. The traditional nursery rhyme goes like this:\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you! (Or something similar.)", "response": " _______.\n\nThe best answer is blue. The traditional nursery rhyme goes like this:\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you! (Or something similar.)",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -20,15 +20,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.076962Z", "created_at": "2025-08-04T22:55:11.15982Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 321102500, "total_duration": 498612042,
"load_duration": 20223000, "load_duration": 71411834,
"prompt_eval_count": 23, "prompt_eval_count": 23,
"prompt_eval_duration": 72000000, "prompt_eval_duration": 102000000,
"eval_count": 6, "eval_count": 6,
"eval_duration": 227000000, "eval_duration": 323000000,
"response": "Humans live on Earth.", "response": "Humans live on Earth.",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -21,7 +21,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.287573Z", "created_at": "2025-08-04T22:55:11.338232Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -39,7 +39,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.341788Z", "created_at": "2025-08-04T22:55:11.39419Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -57,7 +57,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.388655Z", "created_at": "2025-08-04T22:55:11.445346Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -75,7 +75,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.43929Z", "created_at": "2025-08-04T22:55:11.496701Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -93,7 +93,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.485696Z", "created_at": "2025-08-04T22:55:11.546804Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -111,7 +111,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.533286Z", "created_at": "2025-08-04T22:55:11.601009Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -129,7 +129,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.578007Z", "created_at": "2025-08-04T22:55:11.652788Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -147,7 +147,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.623059Z", "created_at": "2025-08-04T22:55:11.703325Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -165,7 +165,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.669478Z", "created_at": "2025-08-04T22:55:11.754033Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -183,7 +183,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.715842Z", "created_at": "2025-08-04T22:55:11.804654Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -201,15 +201,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:49.760345Z", "created_at": "2025-08-04T22:55:11.854841Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 624960709, "total_duration": 652371000,
"load_duration": 59848000, "load_duration": 42086042,
"prompt_eval_count": 26, "prompt_eval_count": 26,
"prompt_eval_duration": 87000000, "prompt_eval_duration": 78000000,
"eval_count": 11, "eval_count": 11,
"eval_duration": 476000000, "eval_duration": 531000000,
"response": "", "response": "",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -67,15 +67,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:11.058922Z", "created_at": "2025-08-04T22:55:40.583477Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 3810519000, "total_duration": 3928481500,
"load_duration": 148409834, "load_duration": 151903250,
"prompt_eval_count": 259, "prompt_eval_count": 259,
"prompt_eval_duration": 635000000, "prompt_eval_duration": 468000000,
"eval_count": 60, "eval_count": 60,
"eval_duration": 3025000000, "eval_duration": 3306000000,
"response": "{\n \"first_name\": \"Michael\",\n \"last_name\": \"Jordan\",\n \"year_of_birth\": 1963,\n \"nba_stats\": {\n \"year_for_draft\": 1984,\n \"num_seasons_in_nba\": 15\n }\n}", "response": "{\n \"first_name\": \"Michael\",\n \"last_name\": \"Jordan\",\n \"year_of_birth\": 1963,\n \"nba_stats\": {\n \"year_for_draft\": 1984,\n \"num_seasons_in_nba\": 15\n }\n}",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -14,7 +14,7 @@
"models": [ "models": [
{ {
"model": "nomic-embed-text:latest", "model": "nomic-embed-text:latest",
"modified_at": "2025-08-01T16:13:37.801531-07:00", "modified_at": "2025-08-04T15:54:50.584357-07:00",
"digest": "0a109f422b47e3a30ba2b10eca18548e944e8a23073ee3f3e947efcf3c45e59f", "digest": "0a109f422b47e3a30ba2b10eca18548e944e8a23073ee3f3e947efcf3c45e59f",
"size": 274302450, "size": 274302450,
"details": { "details": {

View file

@ -0,0 +1,221 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/generate",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"raw": true,
"prompt": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You have access to functions, but you should only use them if they are required.\nYou are an expert in composing functions. You are given a question and a set of possible functions.\nBased on the question, you may or may not need to make one function/tool call to achieve the purpose.\n\nIf you decide to invoke any of the function(s), you MUST put it in the format of [func_name1(params_name1=params_value1, params_name2=params_value2...), func_name2(params)]\nIf you decide to invoke a function, you SHOULD NOT include any other text in the response. besides the function call in the above format.\nFor a boolean parameter, be sure to use `True` or `False` (capitalized) for the value.\n\n\nHere is a list of functions in JSON format that you can invoke.\n\n[\n {\n \"name\": \"get_weather\",\n \"description\": \"Get the current weather\",\n \"parameters\": {\n \"type\": \"dict\",\n \"required\": [\"location\"],\n \"properties\": {\n \"location\": {\n \"type\": \"string\",\n \"description\": \"The city and state (both required), e.g. San Francisco, CA.\"\n }\n }\n }\n }\n]\n\nYou can answer general questions or invoke tools when necessary.\nIn addition to tool calls, you should also augment your responses by using the tool outputs.\nPretend you are a weather assistant.\nYou MUST use one of the provided functions/tools to answer the user query.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\nWhat's the weather like in San Francisco, CA?<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"options": {
"temperature": 0.0
},
"stream": true
},
"endpoint": "/api/generate",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": [
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.141947Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "[",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.194979Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "get",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.248312Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "_weather",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.301911Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "(location",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.354437Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "=\"",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.406821Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "San",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.457633Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " Francisco",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.507857Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": ",",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.558847Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " CA",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.609969Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "\")]",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:14.660997Z",
"done": true,
"done_reason": "stop",
"total_duration": 715356542,
"load_duration": 59747500,
"prompt_eval_count": 341,
"prompt_eval_duration": 128000000,
"eval_count": 11,
"eval_duration": 526000000,
"response": "",
"thinking": null,
"context": null
}
}
],
"is_streaming": true
}
}

View file

@ -0,0 +1,39 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/generate",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"raw": true,
"prompt": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You have access to functions, but you should only use them if they are required.\nYou are an expert in composing functions. You are given a question and a set of possible functions.\nBased on the question, you may or may not need to make one function/tool call to achieve the purpose.\n\nIf you decide to invoke any of the function(s), you MUST put it in the format of [func_name1(params_name1=params_value1, params_name2=params_value2...), func_name2(params)]\nIf you decide to invoke a function, you SHOULD NOT include any other text in the response. besides the function call in the above format.\nFor a boolean parameter, be sure to use `True` or `False` (capitalized) for the value.\n\n\nHere is a list of functions in JSON format that you can invoke.\n\n[\n {\n \"name\": \"get_weather\",\n \"description\": \"Get the current weather\",\n \"parameters\": {\n \"type\": \"dict\",\n \"required\": [\"location\"],\n \"properties\": {\n \"location\": {\n \"type\": \"string\",\n \"description\": \"The city and state (both required), e.g. San Francisco, CA.\"\n }\n }\n }\n }\n]\n\nYou can answer general questions or invoke tools when necessary.\nIn addition to tool calls, you should also augment your responses by using the tool outputs.\nPretend you are a weather assistant.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\nWhat's the weather like in San Francisco, CA?<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"options": {
"temperature": 0.0
},
"stream": false
},
"endpoint": "/api/generate",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": {
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.25248Z",
"done": true,
"done_reason": "stop",
"total_duration": 1344654917,
"load_duration": 200585375,
"prompt_eval_count": 326,
"prompt_eval_duration": 564000000,
"eval_count": 11,
"eval_duration": 578000000,
"response": "[get_weather(location=\"San Francisco, CA\")]",
"thinking": null,
"context": null
}
},
"is_streaming": false
}
}

View file

@ -0,0 +1,221 @@
{
"request": {
"method": "POST",
"url": "http://localhost:11434/api/generate",
"headers": {},
"body": {
"model": "llama3.2:3b-instruct-fp16",
"raw": true,
"prompt": "<|begin_of_text|><|start_header_id|>system<|end_header_id|>\n\nYou are a helpful assistant. You have access to functions, but you should only use them if they are required.\nYou are an expert in composing functions. You are given a question and a set of possible functions.\nBased on the question, you may or may not need to make one function/tool call to achieve the purpose.\n\nIf you decide to invoke any of the function(s), you MUST put it in the format of [func_name1(params_name1=params_value1, params_name2=params_value2...), func_name2(params)]\nIf you decide to invoke a function, you SHOULD NOT include any other text in the response. besides the function call in the above format.\nFor a boolean parameter, be sure to use `True` or `False` (capitalized) for the value.\n\n\nHere is a list of functions in JSON format that you can invoke.\n\n[\n {\n \"name\": \"get_weather\",\n \"description\": \"Get the current weather\",\n \"parameters\": {\n \"type\": \"dict\",\n \"required\": [\"location\"],\n \"properties\": {\n \"location\": {\n \"type\": \"string\",\n \"description\": \"The city and state (both required), e.g. San Francisco, CA.\"\n }\n }\n }\n }\n]\n\nYou can answer general questions or invoke tools when necessary.\nIn addition to tool calls, you should also augment your responses by using the tool outputs.\nPretend you are a weather assistant.<|eot_id|><|start_header_id|>user<|end_header_id|>\n\nWhat's the weather like in San Francisco, CA?<|eot_id|><|start_header_id|>assistant<|end_header_id|>\n\n",
"options": {
"temperature": 0.0
},
"stream": true
},
"endpoint": "/api/generate",
"model": "llama3.2:3b-instruct-fp16"
},
"response": {
"body": [
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.354888Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "[",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.427569Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "get",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.486244Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "_weather",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.540455Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "(location",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.594439Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "=\"",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.649837Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "San",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.703358Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " Francisco",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.7553Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": ",",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.807251Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": " CA",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.857952Z",
"done": false,
"done_reason": null,
"total_duration": null,
"load_duration": null,
"prompt_eval_count": null,
"prompt_eval_duration": null,
"eval_count": null,
"eval_duration": null,
"response": "\")]",
"thinking": null,
"context": null
}
},
{
"__type__": "ollama._types.GenerateResponse",
"__data__": {
"model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-04T22:55:13.918522Z",
"done": true,
"done_reason": "stop",
"total_duration": 647785042,
"load_duration": 26355584,
"prompt_eval_count": 326,
"prompt_eval_duration": 55000000,
"eval_count": 11,
"eval_duration": 557000000,
"response": "",
"thinking": null,
"context": null
}
}
],
"is_streaming": true
}
}

View file

@ -21,7 +21,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.7568Z", "created_at": "2025-08-04T22:55:55.9885Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -39,7 +39,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.808114Z", "created_at": "2025-08-04T22:55:56.054143Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -57,7 +57,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.862297Z", "created_at": "2025-08-04T22:55:56.117658Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -75,7 +75,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.910323Z", "created_at": "2025-08-04T22:55:56.179422Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -93,7 +93,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.95935Z", "created_at": "2025-08-04T22:55:56.240328Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -111,7 +111,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.013354Z", "created_at": "2025-08-04T22:55:56.295992Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -129,7 +129,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.07322Z", "created_at": "2025-08-04T22:55:56.355683Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -147,7 +147,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.131785Z", "created_at": "2025-08-04T22:55:56.412176Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -165,7 +165,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.180306Z", "created_at": "2025-08-04T22:55:56.466952Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -183,7 +183,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.230096Z", "created_at": "2025-08-04T22:55:56.517222Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -201,7 +201,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.281864Z", "created_at": "2025-08-04T22:55:56.570491Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -219,7 +219,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.335153Z", "created_at": "2025-08-04T22:55:56.623189Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -237,7 +237,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.388394Z", "created_at": "2025-08-04T22:55:56.679221Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -255,7 +255,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.441268Z", "created_at": "2025-08-04T22:55:56.731373Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -273,7 +273,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.493573Z", "created_at": "2025-08-04T22:55:56.781364Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -291,7 +291,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.543884Z", "created_at": "2025-08-04T22:55:56.831951Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -309,7 +309,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.592363Z", "created_at": "2025-08-04T22:55:56.888381Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -327,7 +327,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.640596Z", "created_at": "2025-08-04T22:55:56.943539Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -345,7 +345,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.687018Z", "created_at": "2025-08-04T22:55:56.997422Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -363,15 +363,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:21.73497Z", "created_at": "2025-08-04T22:55:57.056259Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 1139022875, "total_duration": 1289815458,
"load_duration": 78012000, "load_duration": 119745583,
"prompt_eval_count": 26, "prompt_eval_count": 26,
"prompt_eval_duration": 75000000, "prompt_eval_duration": 98000000,
"eval_count": 20, "eval_count": 20,
"eval_duration": 984000000, "eval_duration": 1071000000,
"response": "", "response": "",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -20,15 +20,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:20.567617Z", "created_at": "2025-08-04T22:55:55.720345Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 3565468958, "total_duration": 3865701084,
"load_duration": 31098250, "load_duration": 52435459,
"prompt_eval_count": 30, "prompt_eval_count": 30,
"prompt_eval_duration": 77000000, "prompt_eval_duration": 99000000,
"eval_count": 70, "eval_count": 70,
"eval_duration": 3456000000, "eval_duration": 3712000000,
"response": "The answer is Saturn! Saturn's ring system is one of the most iconic and well-known in our solar system. The rings are made up of ice particles, rock debris, and dust that orbit around the planet due to its gravitational pull.\n\nWould you like to know more about Saturn's rings or is there something else I can help you with?", "response": "The answer is Saturn! Saturn's ring system is one of the most iconic and well-known in our solar system. The rings are made up of ice particles, rock debris, and dust that orbit around the planet due to its gravitational pull.\n\nWould you like to know more about Saturn's rings or is there something else I can help you with?",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -45,15 +45,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:48.732625Z", "created_at": "2025-08-04T22:55:10.58267Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 1710455750, "total_duration": 1981967792,
"load_duration": 47968625, "load_duration": 63184458,
"prompt_eval_count": 119, "prompt_eval_count": 119,
"prompt_eval_duration": 223000000, "prompt_eval_duration": 259000000,
"eval_count": 29, "eval_count": 29,
"eval_duration": 1437000000, "eval_duration": 1582000000,
"response": "{ \"name\": \"Michael Jordan\", \"year_born\": \"1963\", \"year_retired\": \"2003\"}\n ", "response": "{ \"name\": \"Michael Jordan\", \"year_born\": \"1963\", \"year_retired\": \"2003\"}\n ",
"thinking": null, "thinking": null,
"context": null "context": null

File diff suppressed because it is too large Load diff

View file

@ -1,7 +1,7 @@
{ {
"request": { "request": {
"method": "POST", "method": "POST",
"url": "http://localhost:11434/v1/v1/completions", "url": "http://0.0.0.0:11434/v1/v1/completions",
"headers": {}, "headers": {},
"body": { "body": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
@ -15,23 +15,23 @@
"body": { "body": {
"__type__": "openai.types.completion.Completion", "__type__": "openai.types.completion.Completion",
"__data__": { "__data__": {
"id": "cmpl-963", "id": "cmpl-43",
"choices": [ "choices": [
{ {
"finish_reason": "stop", "finish_reason": "stop",
"index": 0, "index": 0,
"logprobs": null, "logprobs": null,
"text": "Blue.\n\nMy answer is \"blue\" because it's a common rhyme that completes the famous nursery rhyme: \"Roses are red, violets are blue.\" This rhyme has been passed down for generations and is often recited as a greeting or a way to express appreciation. The phrase typically rhymes with the word \"blue\", which is why I chose this response." "text": "Blue.\n\nMy response is based on the traditional English rhyme that pairs the colors of roses (red) with violets in a poetic and somewhat whimsical way. This specific version of the rhyme goes like this:\n\n\"Roses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you.\"\n\nIn modern times, variations of this rhyme can deviate from the original \"blue\" for violets, but in my complete sentence as requested, sticking with a widely recognized completion adds an air of timelessness and familiarity to the phrase."
} }
], ],
"created": 1754081785, "created": 1754348148,
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"object": "text_completion", "object": "text_completion",
"system_fingerprint": "fp_ollama", "system_fingerprint": "fp_ollama",
"usage": { "usage": {
"completion_tokens": 76, "completion_tokens": 113,
"prompt_tokens": 50, "prompt_tokens": 50,
"total_tokens": 126, "total_tokens": 163,
"completion_tokens_details": null, "completion_tokens_details": null,
"prompt_tokens_details": null "prompt_tokens_details": null
} }

View file

@ -23,7 +23,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.766028Z", "created_at": "2025-08-04T22:55:05.992185Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -41,7 +41,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.810433Z", "created_at": "2025-08-04T22:55:06.047726Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -59,7 +59,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.869051Z", "created_at": "2025-08-04T22:55:06.123375Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -77,7 +77,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.928272Z", "created_at": "2025-08-04T22:55:06.182233Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -95,7 +95,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:43.987671Z", "created_at": "2025-08-04T22:55:06.244655Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -113,7 +113,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.051505Z", "created_at": "2025-08-04T22:55:06.304777Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -131,7 +131,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.183306Z", "created_at": "2025-08-04T22:55:06.361584Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -149,7 +149,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.240354Z", "created_at": "2025-08-04T22:55:06.419647Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -167,7 +167,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.31441Z", "created_at": "2025-08-04T22:55:06.477037Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -185,7 +185,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.370332Z", "created_at": "2025-08-04T22:55:06.534717Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -203,7 +203,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.431975Z", "created_at": "2025-08-04T22:55:06.600289Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -221,7 +221,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.481015Z", "created_at": "2025-08-04T22:55:06.658769Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -239,7 +239,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.530441Z", "created_at": "2025-08-04T22:55:06.71323Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -257,7 +257,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.591576Z", "created_at": "2025-08-04T22:55:06.764206Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -275,7 +275,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.656562Z", "created_at": "2025-08-04T22:55:06.815428Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -293,7 +293,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.707469Z", "created_at": "2025-08-04T22:55:06.86906Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -311,7 +311,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.793146Z", "created_at": "2025-08-04T22:55:06.92191Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -329,7 +329,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.84461Z", "created_at": "2025-08-04T22:55:06.97464Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -347,7 +347,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.892168Z", "created_at": "2025-08-04T22:55:07.026686Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -365,7 +365,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.9392Z", "created_at": "2025-08-04T22:55:07.078382Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -383,7 +383,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:44.986145Z", "created_at": "2025-08-04T22:55:07.131717Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -401,7 +401,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.032843Z", "created_at": "2025-08-04T22:55:07.188206Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -419,7 +419,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.079678Z", "created_at": "2025-08-04T22:55:07.243218Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -437,7 +437,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.125344Z", "created_at": "2025-08-04T22:55:07.298542Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -455,7 +455,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.171365Z", "created_at": "2025-08-04T22:55:07.355167Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -473,7 +473,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.21944Z", "created_at": "2025-08-04T22:55:07.41078Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -491,7 +491,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.266493Z", "created_at": "2025-08-04T22:55:07.463639Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -509,7 +509,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.31207Z", "created_at": "2025-08-04T22:55:07.515619Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -527,7 +527,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.359314Z", "created_at": "2025-08-04T22:55:07.572461Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -545,7 +545,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.406254Z", "created_at": "2025-08-04T22:55:07.626345Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -563,7 +563,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.452757Z", "created_at": "2025-08-04T22:55:07.680673Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -581,7 +581,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.499632Z", "created_at": "2025-08-04T22:55:07.736803Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -599,7 +599,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.545593Z", "created_at": "2025-08-04T22:55:07.789556Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -617,7 +617,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.595114Z", "created_at": "2025-08-04T22:55:07.841142Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -635,7 +635,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.641522Z", "created_at": "2025-08-04T22:55:07.896607Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -653,7 +653,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.688054Z", "created_at": "2025-08-04T22:55:07.953628Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -671,7 +671,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.737359Z", "created_at": "2025-08-04T22:55:08.007575Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -689,7 +689,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.786187Z", "created_at": "2025-08-04T22:55:08.061895Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -707,7 +707,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.834145Z", "created_at": "2025-08-04T22:55:08.121698Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -725,7 +725,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.882153Z", "created_at": "2025-08-04T22:55:08.175866Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -743,7 +743,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.928181Z", "created_at": "2025-08-04T22:55:08.231661Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -761,7 +761,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:45.974825Z", "created_at": "2025-08-04T22:55:08.285188Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -779,15 +779,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:56:46.022862Z", "created_at": "2025-08-04T22:55:08.334914Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 2444514250, "total_duration": 2543128958,
"load_duration": 107849833, "load_duration": 133497375,
"prompt_eval_count": 18, "prompt_eval_count": 18,
"prompt_eval_duration": 65000000, "prompt_eval_duration": 62000000,
"eval_count": 43, "eval_count": 43,
"eval_duration": 2270000000, "eval_duration": 2346000000,
"response": "", "response": "",
"thinking": null, "thinking": null,
"context": null "context": null

View file

@ -20,15 +20,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:22.827462Z", "created_at": "2025-08-04T22:55:57.955211Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 731625417, "total_duration": 842946458,
"load_duration": 58932417, "load_duration": 91343000,
"prompt_eval_count": 386, "prompt_eval_count": 386,
"prompt_eval_duration": 609000000, "prompt_eval_duration": 685000000,
"eval_count": 2, "eval_count": 2,
"eval_duration": 61000000, "eval_duration": 64000000,
"response": "[]", "response": "[]",
"thinking": null, "thinking": null,
"context": null "context": null

File diff suppressed because it is too large Load diff

View file

@ -16,7 +16,7 @@
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"name": "llama3.2:3b-instruct-fp16", "name": "llama3.2:3b-instruct-fp16",
"digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d", "digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d",
"expires_at": "2025-08-01T16:18:11.656155-07:00", "expires_at": "2025-08-04T16:00:57.955349-07:00",
"size": 8581748736, "size": 8581748736,
"size_vram": 8581748736, "size_vram": 8581748736,
"details": { "details": {
@ -29,24 +29,6 @@
"parameter_size": "3.2B", "parameter_size": "3.2B",
"quantization_level": "F16" "quantization_level": "F16"
} }
},
{
"model": "llama-guard3:1b",
"name": "llama-guard3:1b",
"digest": "494147e06bf99e10dbe67b63a07ac81c162f18ef3341aa3390007ac828571b3b",
"expires_at": "2025-08-01T16:17:54.635270-07:00",
"size": 2770397184,
"size_vram": 2770397184,
"details": {
"parent_model": "",
"format": "gguf",
"family": "llama",
"families": [
"llama"
],
"parameter_size": "1.5B",
"quantization_level": "Q8_0"
}
} }
] ]
} }

View file

@ -21,7 +21,7 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:11.920796Z", "created_at": "2025-08-04T22:55:41.559883Z",
"done": false, "done": false,
"done_reason": null, "done_reason": null,
"total_duration": null, "total_duration": null,
@ -39,15 +39,15 @@
"__type__": "ollama._types.GenerateResponse", "__type__": "ollama._types.GenerateResponse",
"__data__": { "__data__": {
"model": "llama3.2:3b-instruct-fp16", "model": "llama3.2:3b-instruct-fp16",
"created_at": "2025-08-01T20:57:11.963674Z", "created_at": "2025-08-04T22:55:41.619829Z",
"done": true, "done": true,
"done_reason": "stop", "done_reason": "stop",
"total_duration": 837987583, "total_duration": 915493834,
"load_duration": 158475167, "load_duration": 167838417,
"prompt_eval_count": 386, "prompt_eval_count": 386,
"prompt_eval_duration": 617000000, "prompt_eval_duration": 683000000,
"eval_count": 2, "eval_count": 2,
"eval_duration": 60000000, "eval_duration": 63000000,
"response": "", "response": "",
"thinking": null, "thinking": null,
"context": null "context": null