mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-15 22:18:00 +00:00
chore: fix: integration tests failures marked as successful (#3039)
This commit is contained in:
parent
eac1e0c7d4
commit
bb6b6041d6
20 changed files with 8449 additions and 871 deletions
|
@ -21,7 +21,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.287573Z",
|
||||
"created_at": "2025-08-04T22:55:11.338232Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -39,7 +39,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.341788Z",
|
||||
"created_at": "2025-08-04T22:55:11.39419Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -57,7 +57,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.388655Z",
|
||||
"created_at": "2025-08-04T22:55:11.445346Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -75,7 +75,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.43929Z",
|
||||
"created_at": "2025-08-04T22:55:11.496701Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -93,7 +93,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.485696Z",
|
||||
"created_at": "2025-08-04T22:55:11.546804Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -111,7 +111,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.533286Z",
|
||||
"created_at": "2025-08-04T22:55:11.601009Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -129,7 +129,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.578007Z",
|
||||
"created_at": "2025-08-04T22:55:11.652788Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -147,7 +147,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.623059Z",
|
||||
"created_at": "2025-08-04T22:55:11.703325Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -165,7 +165,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.669478Z",
|
||||
"created_at": "2025-08-04T22:55:11.754033Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -183,7 +183,7 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.715842Z",
|
||||
"created_at": "2025-08-04T22:55:11.804654Z",
|
||||
"done": false,
|
||||
"done_reason": null,
|
||||
"total_duration": null,
|
||||
|
@ -201,15 +201,15 @@
|
|||
"__type__": "ollama._types.GenerateResponse",
|
||||
"__data__": {
|
||||
"model": "llama3.2:3b-instruct-fp16",
|
||||
"created_at": "2025-08-01T20:56:49.760345Z",
|
||||
"created_at": "2025-08-04T22:55:11.854841Z",
|
||||
"done": true,
|
||||
"done_reason": "stop",
|
||||
"total_duration": 624960709,
|
||||
"load_duration": 59848000,
|
||||
"total_duration": 652371000,
|
||||
"load_duration": 42086042,
|
||||
"prompt_eval_count": 26,
|
||||
"prompt_eval_duration": 87000000,
|
||||
"prompt_eval_duration": 78000000,
|
||||
"eval_count": 11,
|
||||
"eval_duration": 476000000,
|
||||
"eval_duration": 531000000,
|
||||
"response": "",
|
||||
"thinking": null,
|
||||
"context": null
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue