From 2665f0010299b5877fc9e268f755d268aa5d0758 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Wed, 30 Jul 2025 23:30:53 -0700 Subject: [PATCH 01/19] chore(rename): move llama_stack.distribution to llama_stack.core (#2975) We would like to rename the term `template` to `distribution`. To prepare for that, this is a precursor. cc @leseb --- .github/workflows/providers-build.yml | 12 ++++---- .../test-external-provider-module.yml | 5 +++- .github/workflows/test-external.yml | 4 +-- MANIFEST.in | 2 +- docs/getting_started.ipynb | 2 +- docs/getting_started_llama4.ipynb | 2 +- docs/getting_started_llama_api.ipynb | 2 +- .../Alpha_Llama_Stack_Post_Training.ipynb | 28 +++++++++---------- .../Llama_Stack_Agent_Workflows.ipynb | 2 +- .../Llama_Stack_Benchmark_Evals.ipynb | 2 +- .../notebooks/Llama_Stack_RAG_Lifecycle.ipynb | 2 +- .../Llama_Stack_NVIDIA_E2E_Flow.ipynb | 2 +- .../2_finetuning_and_inference.ipynb | 2 +- .../tool_calling/3_model_evaluation.ipynb | 2 +- .../4_adding_safety_guardrails.ipynb | 2 +- docs/openapi_generator/README.md | 2 +- docs/openapi_generator/generate.py | 2 +- docs/openapi_generator/pyopenapi/utility.py | 2 +- docs/original_rfc.md | 2 +- docs/quick_start.ipynb | 2 +- docs/source/apis/external.md | 2 +- .../building_applications/playground/index.md | 2 +- .../source/deploying/kubernetes_deployment.md | 2 +- docs/source/distributions/building_distro.md | 2 +- .../distributions/importing_as_library.md | 2 +- .../distributions/k8s/stack-k8s.yaml.template | 2 +- docs/zero_to_hero_guide/06_Safety101.ipynb | 2 +- llama_stack/__init__.py | 2 +- llama_stack/cli/download.py | 6 ++-- llama_stack/cli/model/list.py | 2 +- llama_stack/cli/model/remove.py | 2 +- llama_stack/cli/stack/_build.py | 22 +++++++-------- llama_stack/cli/stack/build.py | 2 +- llama_stack/cli/stack/list_apis.py | 2 +- llama_stack/cli/stack/list_providers.py | 4 +-- llama_stack/cli/stack/run.py | 8 +++--- llama_stack/cli/verify_download.py | 2 +- .../{distribution => core}/__init__.py | 0 .../access_control/__init__.py | 0 .../access_control/access_control.py | 2 +- .../access_control/conditions.py | 0 .../access_control/datatypes.py | 0 llama_stack/{distribution => core}/build.py | 16 +++++------ .../{distribution => core}/build_conda_env.sh | 0 .../{distribution => core}/build_container.sh | 4 +-- .../{distribution => core}/build_venv.sh | 0 llama_stack/{distribution => core}/client.py | 0 llama_stack/{distribution => core}/common.sh | 0 .../{distribution => core}/configure.py | 12 ++++---- .../{distribution => core}/datatypes.py | 2 +- .../{distribution => core}/distribution.py | 4 +-- .../{distribution => core}/external.py | 2 +- llama_stack/{distribution => core}/inspect.py | 6 ++-- .../{distribution => core}/library_client.py | 20 ++++++------- .../{distribution => core}/providers.py | 0 .../{distribution => core}/request_headers.py | 2 +- .../{distribution => core}/resolver.py | 16 +++++------ .../routers/__init__.py | 6 ++-- .../routers/datasets.py | 0 .../routers/eval_scoring.py | 0 .../routers/inference.py | 0 .../{distribution => core}/routers/safety.py | 0 .../routers/tool_runtime.py | 0 .../routers/vector_io.py | 0 .../routing_tables/__init__.py | 0 .../routing_tables/benchmarks.py | 2 +- .../routing_tables/common.py | 10 +++---- .../routing_tables/datasets.py | 2 +- .../routing_tables/models.py | 2 +- .../routing_tables/scoring_functions.py | 2 +- .../routing_tables/shields.py | 2 +- .../routing_tables/toolgroups.py | 2 +- .../routing_tables/vector_dbs.py | 2 +- .../{distribution => core}/server/__init__.py | 0 .../{distribution => core}/server/auth.py | 8 +++--- .../server/auth_providers.py | 2 +- .../{distribution => core}/server/quota.py | 0 .../{distribution => core}/server/routes.py | 2 +- .../{distribution => core}/server/server.py | 22 +++++++-------- llama_stack/{distribution => core}/stack.py | 16 +++++------ .../{distribution => core}/start_stack.sh | 2 +- .../{distribution => core}/store/__init__.py | 0 .../{distribution => core}/store/registry.py | 4 +-- .../{distribution => core}/ui/Containerfile | 0 .../{distribution => core}/ui/README.md | 2 +- .../{distribution => core}/ui/__init__.py | 0 llama_stack/{distribution => core}/ui/app.py | 0 .../ui/modules/__init__.py | 0 .../{distribution => core}/ui/modules/api.py | 0 .../ui/modules/utils.py | 0 .../ui/page/__init__.py | 0 .../ui/page/distribution/__init__.py | 0 .../ui/page/distribution/datasets.py | 2 +- .../ui/page/distribution/eval_tasks.py | 2 +- .../ui/page/distribution/models.py | 2 +- .../ui/page/distribution/providers.py | 2 +- .../ui/page/distribution/resources.py | 12 ++++---- .../ui/page/distribution/scoring_functions.py | 2 +- .../ui/page/distribution/shields.py | 2 +- .../ui/page/distribution/vector_dbs.py | 2 +- .../ui/page/evaluations/__init__.py | 0 .../ui/page/evaluations/app_eval.py | 4 +-- .../ui/page/evaluations/native_eval.py | 2 +- .../ui/page/playground/__init__.py | 0 .../ui/page/playground/chat.py | 2 +- .../ui/page/playground/rag.py | 4 +-- .../ui/page/playground/tools.py | 2 +- .../ui/requirements.txt | 0 .../{distribution => core}/utils/__init__.py | 0 .../{distribution => core}/utils/config.py | 0 .../utils/config_dirs.py | 0 .../utils/config_resolution.py | 4 +-- .../{distribution => core}/utils/context.py | 0 .../{distribution => core}/utils/dynamic.py | 0 .../{distribution => core}/utils/exec.py | 4 +-- .../utils/image_types.py | 0 .../utils/model_utils.py | 0 .../utils/prompt_for_config.py | 0 .../{distribution => core}/utils/serialize.py | 0 llama_stack/log.py | 2 +- .../inline/agents/meta_reference/__init__.py | 2 +- .../agents/meta_reference/agent_instance.py | 2 +- .../inline/agents/meta_reference/agents.py | 2 +- .../agents/meta_reference/persistence.py | 8 +++--- .../inline/eval/meta_reference/__init__.py | 2 +- .../inline/files/localfs/__init__.py | 2 +- .../providers/inline/files/localfs/files.py | 2 +- .../inline/inference/meta_reference/common.py | 2 +- .../post_training/huggingface/__init__.py | 2 +- .../post_training/torchtune/__init__.py | 2 +- .../recipes/lora_finetuning_single_device.py | 4 +-- .../inline/safety/llama_guard/llama_guard.py | 2 +- .../safety/prompt_guard/prompt_guard.py | 2 +- .../inline/scoring/basic/__init__.py | 2 +- .../providers/inline/scoring/basic/scoring.py | 2 +- .../inline/scoring/braintrust/__init__.py | 2 +- .../inline/scoring/braintrust/braintrust.py | 4 +-- .../inline/scoring/llm_as_judge/__init__.py | 2 +- .../inline/scoring/llm_as_judge/scoring.py | 2 +- .../telemetry/meta_reference/__init__.py | 2 +- .../inline/telemetry/meta_reference/config.py | 2 +- .../telemetry/meta_reference/telemetry.py | 2 +- .../remote/datasetio/nvidia/README.md | 2 +- .../providers/remote/eval/nvidia/__init__.py | 2 +- .../remote/inference/fireworks/fireworks.py | 2 +- .../remote/inference/nvidia/NVIDIA.md | 2 +- .../inference/passthrough/passthrough.py | 2 +- .../remote/inference/together/together.py | 2 +- .../remote/post_training/nvidia/README.md | 2 +- .../providers/remote/safety/nvidia/README.md | 2 +- .../remote/safety/sambanova/sambanova.py | 2 +- .../tool_runtime/bing_search/bing_search.py | 2 +- .../tool_runtime/brave_search/brave_search.py | 2 +- .../model_context_protocol.py | 2 +- .../tavily_search/tavily_search.py | 2 +- .../wolfram_alpha/wolfram_alpha.py | 2 +- .../remote/vector_io/weaviate/weaviate.py | 2 +- .../utils/common/data_schema_validator.py | 2 +- .../utils/inference/inference_store.py | 4 +-- .../utils/inference/litellm_openai_mixin.py | 2 +- llama_stack/providers/utils/kvstore/config.py | 2 +- .../utils/responses/responses_store.py | 4 +-- .../utils/sqlstore/authorized_sqlstore.py | 10 +++---- .../providers/utils/sqlstore/sqlstore.py | 2 +- llama_stack/providers/utils/tools/mcp.py | 2 +- llama_stack/templates/dell/dell.py | 2 +- .../meta-reference-gpu/meta_reference.py | 2 +- llama_stack/templates/nvidia/nvidia.py | 2 +- .../open-benchmark/open_benchmark.py | 2 +- .../templates/postgres-demo/postgres_demo.py | 2 +- llama_stack/templates/starter/starter.py | 4 +-- llama_stack/templates/template.py | 6 ++-- llama_stack/templates/watsonx/watsonx.py | 2 +- pyproject.toml | 22 +++++++-------- scripts/provider_codegen.py | 2 +- .../agents/test_openai_responses.py | 2 +- tests/integration/files/test_files.py | 4 +-- tests/integration/fixtures/common.py | 2 +- .../inference/test_openai_completion.py | 2 +- .../inference/test_openai_embeddings.py | 2 +- .../sqlstore/test_authorized_sqlstore.py | 6 ++-- tests/integration/tool_runtime/test_mcp.py | 2 +- .../vector_io/test_openai_vector_stores.py | 2 +- tests/unit/cli/test_stack_config.py | 2 +- .../routers/test_routing_tables.py | 16 +++++------ .../routing_tables/test_vector_dbs.py | 8 +++--- tests/unit/distribution/test_build_path.py | 4 +-- tests/unit/distribution/test_context.py | 2 +- tests/unit/distribution/test_distribution.py | 12 ++++---- .../test_library_client_initialization.py | 2 +- tests/unit/files/test_files.py | 2 +- tests/unit/fixtures.py | 2 +- .../meta_reference/test_openai_responses.py | 2 +- .../agents/test_persistence_access_control.py | 2 +- .../test_inference_client_caching.py | 2 +- .../inference/test_openai_base_url_config.py | 2 +- .../unit/providers/nvidia/test_parameters.py | 2 +- .../nvidia/test_supervised_fine_tuning.py | 2 +- tests/unit/providers/test_configs.py | 4 +-- tests/unit/registry/test_registry.py | 2 +- tests/unit/registry/test_registry_acl.py | 4 +-- tests/unit/server/test_access_control.py | 20 ++++++------- tests/unit/server/test_auth.py | 14 +++++----- tests/unit/server/test_auth_github.py | 10 +++---- tests/unit/server/test_quota.py | 4 +-- tests/unit/server/test_replace_env_vars.py | 2 +- tests/unit/server/test_resolver.py | 8 +++--- tests/unit/server/test_server.py | 8 +++--- tests/unit/server/test_sse.py | 2 +- tests/unit/utils/test_authorized_sqlstore.py | 6 ++-- .../openai_api/test_responses.py | 2 +- 211 files changed, 351 insertions(+), 348 deletions(-) rename llama_stack/{distribution => core}/__init__.py (100%) rename llama_stack/{distribution => core}/access_control/__init__.py (100%) rename llama_stack/{distribution => core}/access_control/access_control.py (98%) rename llama_stack/{distribution => core}/access_control/conditions.py (100%) rename llama_stack/{distribution => core}/access_control/datatypes.py (100%) rename llama_stack/{distribution => core}/build.py (93%) rename llama_stack/{distribution => core}/build_conda_env.sh (100%) rename llama_stack/{distribution => core}/build_container.sh (98%) rename llama_stack/{distribution => core}/build_venv.sh (100%) rename llama_stack/{distribution => core}/client.py (100%) rename llama_stack/{distribution => core}/common.sh (100%) rename llama_stack/{distribution => core}/configure.py (93%) rename llama_stack/{distribution => core}/datatypes.py (99%) rename llama_stack/{distribution => core}/distribution.py (98%) rename llama_stack/{distribution => core}/external.py (96%) rename llama_stack/{distribution => core}/inspect.py (93%) rename llama_stack/{distribution => core}/library_client.py (96%) rename llama_stack/{distribution => core}/providers.py (100%) rename llama_stack/{distribution => core}/request_headers.py (98%) rename llama_stack/{distribution => core}/resolver.py (97%) rename llama_stack/{distribution => core}/routers/__init__.py (94%) rename llama_stack/{distribution => core}/routers/datasets.py (100%) rename llama_stack/{distribution => core}/routers/eval_scoring.py (100%) rename llama_stack/{distribution => core}/routers/inference.py (100%) rename llama_stack/{distribution => core}/routers/safety.py (100%) rename llama_stack/{distribution => core}/routers/tool_runtime.py (100%) rename llama_stack/{distribution => core}/routers/vector_io.py (100%) rename llama_stack/{distribution => core}/routing_tables/__init__.py (100%) rename llama_stack/{distribution => core}/routing_tables/benchmarks.py (97%) rename llama_stack/{distribution => core}/routing_tables/common.py (96%) rename llama_stack/{distribution => core}/routing_tables/datasets.py (98%) rename llama_stack/{distribution => core}/routing_tables/models.py (99%) rename llama_stack/{distribution => core}/routing_tables/scoring_functions.py (97%) rename llama_stack/{distribution => core}/routing_tables/shields.py (97%) rename llama_stack/{distribution => core}/routing_tables/toolgroups.py (98%) rename llama_stack/{distribution => core}/routing_tables/vector_dbs.py (99%) rename llama_stack/{distribution => core}/server/__init__.py (100%) rename llama_stack/{distribution => core}/server/auth.py (95%) rename llama_stack/{distribution => core}/server/auth_providers.py (99%) rename llama_stack/{distribution => core}/server/quota.py (100%) rename llama_stack/{distribution => core}/server/routes.py (98%) rename llama_stack/{distribution => core}/server/server.py (96%) rename llama_stack/{distribution => core}/stack.py (96%) rename llama_stack/{distribution => core}/start_stack.sh (98%) rename llama_stack/{distribution => core}/store/__init__.py (100%) rename llama_stack/{distribution => core}/store/registry.py (98%) rename llama_stack/{distribution => core}/ui/Containerfile (100%) rename llama_stack/{distribution => core}/ui/README.md (96%) rename llama_stack/{distribution => core}/ui/__init__.py (100%) rename llama_stack/{distribution => core}/ui/app.py (100%) rename llama_stack/{distribution => core}/ui/modules/__init__.py (100%) rename llama_stack/{distribution => core}/ui/modules/api.py (100%) rename llama_stack/{distribution => core}/ui/modules/utils.py (100%) rename llama_stack/{distribution => core}/ui/page/__init__.py (100%) rename llama_stack/{distribution => core}/ui/page/distribution/__init__.py (100%) rename llama_stack/{distribution => core}/ui/page/distribution/datasets.py (88%) rename llama_stack/{distribution => core}/ui/page/distribution/eval_tasks.py (90%) rename llama_stack/{distribution => core}/ui/page/distribution/models.py (87%) rename llama_stack/{distribution => core}/ui/page/distribution/providers.py (91%) rename llama_stack/{distribution => core}/ui/page/distribution/resources.py (70%) rename llama_stack/{distribution => core}/ui/page/distribution/scoring_functions.py (89%) rename llama_stack/{distribution => core}/ui/page/distribution/shields.py (88%) rename llama_stack/{distribution => core}/ui/page/distribution/vector_dbs.py (90%) rename llama_stack/{distribution => core}/ui/page/evaluations/__init__.py (100%) rename llama_stack/{distribution => core}/ui/page/evaluations/app_eval.py (97%) rename llama_stack/{distribution => core}/ui/page/evaluations/native_eval.py (99%) rename llama_stack/{distribution => core}/ui/page/playground/__init__.py (100%) rename llama_stack/{distribution => core}/ui/page/playground/chat.py (98%) rename llama_stack/{distribution => core}/ui/page/playground/rag.py (98%) rename llama_stack/{distribution => core}/ui/page/playground/tools.py (99%) rename llama_stack/{distribution => core}/ui/requirements.txt (100%) rename llama_stack/{distribution => core}/utils/__init__.py (100%) rename llama_stack/{distribution => core}/utils/config.py (100%) rename llama_stack/{distribution => core}/utils/config_dirs.py (100%) rename llama_stack/{distribution => core}/utils/config_resolution.py (96%) rename llama_stack/{distribution => core}/utils/context.py (100%) rename llama_stack/{distribution => core}/utils/dynamic.py (100%) rename llama_stack/{distribution => core}/utils/exec.py (96%) rename llama_stack/{distribution => core}/utils/image_types.py (100%) rename llama_stack/{distribution => core}/utils/model_utils.py (100%) rename llama_stack/{distribution => core}/utils/prompt_for_config.py (100%) rename llama_stack/{distribution => core}/utils/serialize.py (100%) diff --git a/.github/workflows/providers-build.yml b/.github/workflows/providers-build.yml index 284076d50..e0edb5486 100644 --- a/.github/workflows/providers-build.yml +++ b/.github/workflows/providers-build.yml @@ -9,8 +9,8 @@ on: paths: - 'llama_stack/cli/stack/build.py' - 'llama_stack/cli/stack/_build.py' - - 'llama_stack/distribution/build.*' - - 'llama_stack/distribution/*.sh' + - 'llama_stack/core/build.*' + - 'llama_stack/core/*.sh' - '.github/workflows/providers-build.yml' - 'llama_stack/templates/**' - 'pyproject.toml' @@ -19,8 +19,8 @@ on: paths: - 'llama_stack/cli/stack/build.py' - 'llama_stack/cli/stack/_build.py' - - 'llama_stack/distribution/build.*' - - 'llama_stack/distribution/*.sh' + - 'llama_stack/core/build.*' + - 'llama_stack/core/*.sh' - '.github/workflows/providers-build.yml' - 'llama_stack/templates/**' - 'pyproject.toml' @@ -108,7 +108,7 @@ jobs: IMAGE_ID=$(docker images --format "{{.Repository}}:{{.Tag}}" | head -n 1) entrypoint=$(docker inspect --format '{{ .Config.Entrypoint }}' $IMAGE_ID) echo "Entrypoint: $entrypoint" - if [ "$entrypoint" != "[python -m llama_stack.distribution.server.server --config /app/run.yaml]" ]; then + if [ "$entrypoint" != "[python -m llama_stack.core.server.server --config /app/run.yaml]" ]; then echo "Entrypoint is not correct" exit 1 fi @@ -142,7 +142,7 @@ jobs: IMAGE_ID=$(docker images --format "{{.Repository}}:{{.Tag}}" | head -n 1) entrypoint=$(docker inspect --format '{{ .Config.Entrypoint }}' $IMAGE_ID) echo "Entrypoint: $entrypoint" - if [ "$entrypoint" != "[python -m llama_stack.distribution.server.server --config /app/run.yaml]" ]; then + if [ "$entrypoint" != "[python -m llama_stack.core.server.server --config /app/run.yaml]" ]; then echo "Entrypoint is not correct" exit 1 fi diff --git a/.github/workflows/test-external-provider-module.yml b/.github/workflows/test-external-provider-module.yml index 8567a9446..c3d11f28a 100644 --- a/.github/workflows/test-external-provider-module.yml +++ b/.github/workflows/test-external-provider-module.yml @@ -16,6 +16,9 @@ on: - 'tests/external/*' - '.github/workflows/test-external-provider-module.yml' # This workflow +# This workflow is disabled. See https://github.com/meta-llama/llama-stack/pull/2975#issuecomment-3138702984 for details +if: false + jobs: test-external-providers-from-module: runs-on: ubuntu-latest @@ -47,7 +50,7 @@ jobs: - name: Build distro from config file run: | - USE_COPY_NOT_MOUNT=true LLAMA_STACK_DIR=. llama stack build --config tests/external/ramalama-stack/build.yaml + USE_COPY_NOT_MOUNT=true LLAMA_STACK_DIR=. uv run llama stack build --config tests/external/ramalama-stack/build.yaml - name: Start Llama Stack server in background if: ${{ matrix.image-type }} == 'venv' diff --git a/.github/workflows/test-external.yml b/.github/workflows/test-external.yml index 053b38fab..27181a236 100644 --- a/.github/workflows/test-external.yml +++ b/.github/workflows/test-external.yml @@ -43,11 +43,11 @@ jobs: - name: Print distro dependencies run: | - USE_COPY_NOT_MOUNT=true LLAMA_STACK_DIR=. llama stack build --config tests/external/build.yaml --print-deps-only + USE_COPY_NOT_MOUNT=true LLAMA_STACK_DIR=. uv run llama stack build --config tests/external/build.yaml --print-deps-only - name: Build distro from config file run: | - USE_COPY_NOT_MOUNT=true LLAMA_STACK_DIR=. llama stack build --config tests/external/build.yaml + USE_COPY_NOT_MOUNT=true LLAMA_STACK_DIR=. uv run llama stack build --config tests/external/build.yaml - name: Start Llama Stack server in background if: ${{ matrix.image-type }} == 'venv' diff --git a/MANIFEST.in b/MANIFEST.in index 88bd11767..fde033d96 100644 --- a/MANIFEST.in +++ b/MANIFEST.in @@ -1,7 +1,7 @@ include pyproject.toml include llama_stack/models/llama/llama3/tokenizer.model include llama_stack/models/llama/llama4/tokenizer.model -include llama_stack/distribution/*.sh +include llama_stack.core/*.sh include llama_stack/cli/scripts/*.sh include llama_stack/templates/*/*.yaml include llama_stack/providers/tests/test_cases/inference/*.json diff --git a/docs/getting_started.ipynb b/docs/getting_started.ipynb index 88878c9be..c87ce4152 100644 --- a/docs/getting_started.ipynb +++ b/docs/getting_started.ipynb @@ -165,7 +165,7 @@ "# use this helper if needed to kill the server \n", "def kill_llama_stack_server():\n", " # Kill any existing llama stack server processes\n", - " os.system(\"ps aux | grep -v grep | grep llama_stack.distribution.server.server | awk '{print $2}' | xargs kill -9\")\n" + " os.system(\"ps aux | grep -v grep | grep llama_stack.core.server.server | awk '{print $2}' | xargs kill -9\")\n" ] }, { diff --git a/docs/getting_started_llama4.ipynb b/docs/getting_started_llama4.ipynb index 82aef6039..7c274de8c 100644 --- a/docs/getting_started_llama4.ipynb +++ b/docs/getting_started_llama4.ipynb @@ -275,7 +275,7 @@ "# use this helper if needed to kill the server \n", "def kill_llama_stack_server():\n", " # Kill any existing llama stack server processes\n", - " os.system(\"ps aux | grep -v grep | grep llama_stack.distribution.server.server | awk '{print $2}' | xargs kill -9\")\n" + " os.system(\"ps aux | grep -v grep | grep llama_stack.core.server.server | awk '{print $2}' | xargs kill -9\")\n" ] }, { diff --git a/docs/getting_started_llama_api.ipynb b/docs/getting_started_llama_api.ipynb index e6c74986b..7950cbd12 100644 --- a/docs/getting_started_llama_api.ipynb +++ b/docs/getting_started_llama_api.ipynb @@ -265,7 +265,7 @@ "# use this helper if needed to kill the server \n", "def kill_llama_stack_server():\n", " # Kill any existing llama stack server processes\n", - " os.system(\"ps aux | grep -v grep | grep llama_stack.distribution.server.server | awk '{print $2}' | xargs kill -9\")\n" + " os.system(\"ps aux | grep -v grep | grep llama_stack.core.server.server | awk '{print $2}' | xargs kill -9\")\n" ] }, { diff --git a/docs/notebooks/Alpha_Llama_Stack_Post_Training.ipynb b/docs/notebooks/Alpha_Llama_Stack_Post_Training.ipynb index b7d769b51..59e516fd0 100644 --- a/docs/notebooks/Alpha_Llama_Stack_Post_Training.ipynb +++ b/docs/notebooks/Alpha_Llama_Stack_Post_Training.ipynb @@ -3216,19 +3216,19 @@ "INFO:datasets:Duckdb version 1.1.3 available.\n", "INFO:datasets:TensorFlow version 2.18.0 available.\n", "INFO:datasets:JAX version 0.4.33 available.\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: basic::equality served by basic\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: basic::subset_of served by basic\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: basic::regex_parser_multiple_choice_answer served by basic\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::factuality served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::answer-correctness served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::answer-relevancy served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::answer-similarity served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::faithfulness served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::context-entity-recall served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::context-precision served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::context-recall served by braintrust\n", - "INFO:llama_stack.distribution.stack:Scoring_fns: braintrust::context-relevancy served by braintrust\n", - "INFO:llama_stack.distribution.stack:\n" + "INFO:llama_stack.core.stack:Scoring_fns: basic::equality served by basic\n", + "INFO:llama_stack.core.stack:Scoring_fns: basic::subset_of served by basic\n", + "INFO:llama_stack.core.stack:Scoring_fns: basic::regex_parser_multiple_choice_answer served by basic\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::factuality served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::answer-correctness served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::answer-relevancy served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::answer-similarity served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::faithfulness served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::context-entity-recall served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::context-precision served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::context-recall served by braintrust\n", + "INFO:llama_stack.core.stack:Scoring_fns: braintrust::context-relevancy served by braintrust\n", + "INFO:llama_stack.core.stack:\n" ] }, { @@ -3448,7 +3448,7 @@ "\n", "os.environ['OPENAI_API_KEY'] = userdata.get('OPENAI_API_KEY')\n", "\n", - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "client = LlamaStackAsLibraryClient(\"experimental-post-training\")\n", "_ = client.initialize()" ] diff --git a/docs/notebooks/Llama_Stack_Agent_Workflows.ipynb b/docs/notebooks/Llama_Stack_Agent_Workflows.ipynb index cad28ab82..0edef4204 100644 --- a/docs/notebooks/Llama_Stack_Agent_Workflows.ipynb +++ b/docs/notebooks/Llama_Stack_Agent_Workflows.ipynb @@ -48,7 +48,7 @@ "outputs": [], "source": [ "from llama_stack_client import LlamaStackClient, Agent\n", - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "from rich.pretty import pprint\n", "import json\n", "import uuid\n", diff --git a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb index 93f78d268..72840ceaf 100644 --- a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb +++ b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb @@ -661,7 +661,7 @@ "except ImportError:\n", " print(\"Not in Google Colab environment\")\n", "\n", - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "\n", "client = LlamaStackAsLibraryClient(\"together\")\n", "_ = client.initialize()" diff --git a/docs/notebooks/Llama_Stack_RAG_Lifecycle.ipynb b/docs/notebooks/Llama_Stack_RAG_Lifecycle.ipynb index e70cc3bbe..769c91dfd 100644 --- a/docs/notebooks/Llama_Stack_RAG_Lifecycle.ipynb +++ b/docs/notebooks/Llama_Stack_RAG_Lifecycle.ipynb @@ -35,7 +35,7 @@ ], "source": [ "from llama_stack_client import LlamaStackClient, Agent\n", - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "from rich.pretty import pprint\n", "import json\n", "import uuid\n", diff --git a/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb b/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb index 583870404..0db0b26b6 100644 --- a/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb +++ b/docs/notebooks/nvidia/beginner_e2e/Llama_Stack_NVIDIA_E2E_Flow.ipynb @@ -194,7 +194,7 @@ "metadata": {}, "outputs": [], "source": [ - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "\n", "client = LlamaStackAsLibraryClient(\"nvidia\")\n", "client.initialize()" diff --git a/docs/notebooks/nvidia/tool_calling/2_finetuning_and_inference.ipynb b/docs/notebooks/nvidia/tool_calling/2_finetuning_and_inference.ipynb index 647a16b6d..a80720a5f 100644 --- a/docs/notebooks/nvidia/tool_calling/2_finetuning_and_inference.ipynb +++ b/docs/notebooks/nvidia/tool_calling/2_finetuning_and_inference.ipynb @@ -56,7 +56,7 @@ "metadata": {}, "outputs": [], "source": [ - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "\n", "client = LlamaStackAsLibraryClient(\"nvidia\")\n", "client.initialize()" diff --git a/docs/notebooks/nvidia/tool_calling/3_model_evaluation.ipynb b/docs/notebooks/nvidia/tool_calling/3_model_evaluation.ipynb index 5a1316adb..91d1db88f 100644 --- a/docs/notebooks/nvidia/tool_calling/3_model_evaluation.ipynb +++ b/docs/notebooks/nvidia/tool_calling/3_model_evaluation.ipynb @@ -56,7 +56,7 @@ "metadata": {}, "outputs": [], "source": [ - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "\n", "client = LlamaStackAsLibraryClient(\"nvidia\")\n", "client.initialize()" diff --git a/docs/notebooks/nvidia/tool_calling/4_adding_safety_guardrails.ipynb b/docs/notebooks/nvidia/tool_calling/4_adding_safety_guardrails.ipynb index 699a561f9..25bcd0b69 100644 --- a/docs/notebooks/nvidia/tool_calling/4_adding_safety_guardrails.ipynb +++ b/docs/notebooks/nvidia/tool_calling/4_adding_safety_guardrails.ipynb @@ -56,7 +56,7 @@ "metadata": {}, "outputs": [], "source": [ - "from llama_stack.distribution.library_client import LlamaStackAsLibraryClient\n", + "from llama_stack.core.library_client import LlamaStackAsLibraryClient\n", "\n", "client = LlamaStackAsLibraryClient(\"nvidia\")\n", "client.initialize()" diff --git a/docs/openapi_generator/README.md b/docs/openapi_generator/README.md index 7888e7828..85021d911 100644 --- a/docs/openapi_generator/README.md +++ b/docs/openapi_generator/README.md @@ -1 +1 @@ -The RFC Specification (OpenAPI format) is generated from the set of API endpoints located in `llama_stack/distribution/server/endpoints.py` using the `generate.py` utility. +The RFC Specification (OpenAPI format) is generated from the set of API endpoints located in `llama_stack.core/server/endpoints.py` using the `generate.py` utility. diff --git a/docs/openapi_generator/generate.py b/docs/openapi_generator/generate.py index 9fc375175..c27bc6440 100644 --- a/docs/openapi_generator/generate.py +++ b/docs/openapi_generator/generate.py @@ -17,7 +17,7 @@ import fire import ruamel.yaml as yaml from llama_stack.apis.version import LLAMA_STACK_API_VERSION # noqa: E402 -from llama_stack.distribution.stack import LlamaStack # noqa: E402 +from llama_stack.core.stack import LlamaStack # noqa: E402 from .pyopenapi.options import Options # noqa: E402 from .pyopenapi.specification import Info, Server # noqa: E402 diff --git a/docs/openapi_generator/pyopenapi/utility.py b/docs/openapi_generator/pyopenapi/utility.py index 57f92403d..d302b114f 100644 --- a/docs/openapi_generator/pyopenapi/utility.py +++ b/docs/openapi_generator/pyopenapi/utility.py @@ -12,7 +12,7 @@ from typing import TextIO from typing import Any, List, Optional, Union, get_type_hints, get_origin, get_args from llama_stack.strong_typing.schema import object_to_json, StrictJsonType -from llama_stack.distribution.resolver import api_protocol_map +from llama_stack.core.resolver import api_protocol_map from .generator import Generator from .options import Options diff --git a/docs/original_rfc.md b/docs/original_rfc.md index dc95a04cb..e9191cb6d 100644 --- a/docs/original_rfc.md +++ b/docs/original_rfc.md @@ -73,7 +73,7 @@ The API is defined in the [YAML](_static/llama-stack-spec.yaml) and [HTML](_stat To prove out the API, we implemented a handful of use cases to make things more concrete. The [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps) repository contains [6 different examples](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) ranging from very basic to a multi turn agent. -There is also a sample inference endpoint implementation in the [llama-stack](https://github.com/meta-llama/llama-stack/blob/main/llama_stack/distribution/server/server.py) repository. +There is also a sample inference endpoint implementation in the [llama-stack](https://github.com/meta-llama/llama-stack/blob/main/llama_stack.core/server/server.py) repository. ## Limitations diff --git a/docs/quick_start.ipynb b/docs/quick_start.ipynb index c3049a70f..e99276427 100644 --- a/docs/quick_start.ipynb +++ b/docs/quick_start.ipynb @@ -187,7 +187,7 @@ "# use this helper if needed to kill the server \n", "def kill_llama_stack_server():\n", " # Kill any existing llama stack server processes\n", - " os.system(\"ps aux | grep -v grep | grep llama_stack.distribution.server.server | awk '{print $2}' | xargs kill -9\")\n" + " os.system(\"ps aux | grep -v grep | grep llama_stack.core.server.server | awk '{print $2}' | xargs kill -9\")\n" ] }, { diff --git a/docs/source/apis/external.md b/docs/source/apis/external.md index 025267c33..cc13deb9b 100644 --- a/docs/source/apis/external.md +++ b/docs/source/apis/external.md @@ -355,7 +355,7 @@ server: 8. Run the server: ```bash -python -m llama_stack.distribution.server.server --yaml-config ~/.llama/run-byoa.yaml +python -m llama_stack.core.server.server --yaml-config ~/.llama/run-byoa.yaml ``` 9. Test the API: diff --git a/docs/source/building_applications/playground/index.md b/docs/source/building_applications/playground/index.md index 85895f6a5..399b41203 100644 --- a/docs/source/building_applications/playground/index.md +++ b/docs/source/building_applications/playground/index.md @@ -103,5 +103,5 @@ llama stack run together 2. Start Streamlit UI ```bash -uv run --with ".[ui]" streamlit run llama_stack/distribution/ui/app.py +uv run --with ".[ui]" streamlit run llama_stack.core/ui/app.py ``` diff --git a/docs/source/deploying/kubernetes_deployment.md b/docs/source/deploying/kubernetes_deployment.md index 7e9791d8d..4bdd87b24 100644 --- a/docs/source/deploying/kubernetes_deployment.md +++ b/docs/source/deploying/kubernetes_deployment.md @@ -174,7 +174,7 @@ spec: - name: llama-stack image: localhost/llama-stack-run-k8s:latest imagePullPolicy: IfNotPresent - command: ["python", "-m", "llama_stack.distribution.server.server", "--config", "/app/config.yaml"] + command: ["python", "-m", "llama_stack.core.server.server", "--config", "/app/config.yaml"] ports: - containerPort: 5000 volumeMounts: diff --git a/docs/source/distributions/building_distro.md b/docs/source/distributions/building_distro.md index cd2c6b6a8..fecfbac38 100644 --- a/docs/source/distributions/building_distro.md +++ b/docs/source/distributions/building_distro.md @@ -59,7 +59,7 @@ Build a Llama stack container options: -h, --help show this help message and exit - --config CONFIG Path to a config file to use for the build. You can find example configs in llama_stack/distributions/**/build.yaml. If this argument is not provided, you will + --config CONFIG Path to a config file to use for the build. You can find example configs in llama_stack.cores/**/build.yaml. If this argument is not provided, you will be prompted to enter information interactively (default: None) --template TEMPLATE Name of the example template config to use for build. You may use `llama stack build --list-templates` to check out the available templates (default: None) --list-templates Show the available templates for building a Llama Stack distribution (default: False) diff --git a/docs/source/distributions/importing_as_library.md b/docs/source/distributions/importing_as_library.md index 3427356a7..0f2acc974 100644 --- a/docs/source/distributions/importing_as_library.md +++ b/docs/source/distributions/importing_as_library.md @@ -10,7 +10,7 @@ llama stack build --template starter --image-type venv ``` ```python -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient client = LlamaStackAsLibraryClient( "starter", diff --git a/docs/source/distributions/k8s/stack-k8s.yaml.template b/docs/source/distributions/k8s/stack-k8s.yaml.template index 912445f68..ad5d2c716 100644 --- a/docs/source/distributions/k8s/stack-k8s.yaml.template +++ b/docs/source/distributions/k8s/stack-k8s.yaml.template @@ -52,7 +52,7 @@ spec: value: "${SAFETY_MODEL}" - name: TAVILY_SEARCH_API_KEY value: "${TAVILY_SEARCH_API_KEY}" - command: ["python", "-m", "llama_stack.distribution.server.server", "--config", "/etc/config/stack_run_config.yaml", "--port", "8321"] + command: ["python", "-m", "llama_stack.core.server.server", "--config", "/etc/config/stack_run_config.yaml", "--port", "8321"] ports: - containerPort: 8321 volumeMounts: diff --git a/docs/zero_to_hero_guide/06_Safety101.ipynb b/docs/zero_to_hero_guide/06_Safety101.ipynb index 5d7763924..91b809621 100644 --- a/docs/zero_to_hero_guide/06_Safety101.ipynb +++ b/docs/zero_to_hero_guide/06_Safety101.ipynb @@ -66,7 +66,7 @@ "from pydantic import BaseModel\n", "from termcolor import cprint\n", "\n", - "from llama_stack.distribution.datatypes import RemoteProviderConfig\n", + "from llama_stack.core.datatypes import RemoteProviderConfig\n", "from llama_stack.apis.safety import Safety\n", "from llama_stack_client import LlamaStackClient\n", "\n", diff --git a/llama_stack/__init__.py b/llama_stack/__init__.py index 98f2441c0..1c2ce7123 100644 --- a/llama_stack/__init__.py +++ b/llama_stack/__init__.py @@ -4,7 +4,7 @@ # This source code is licensed under the terms described in the LICENSE file in # the root directory of this source tree. -from llama_stack.distribution.library_client import ( # noqa: F401 +from llama_stack.core.library_client import ( # noqa: F401 AsyncLlamaStackAsLibraryClient, LlamaStackAsLibraryClient, ) diff --git a/llama_stack/cli/download.py b/llama_stack/cli/download.py index 30b6e11e9..70cb9f4db 100644 --- a/llama_stack/cli/download.py +++ b/llama_stack/cli/download.py @@ -323,7 +323,7 @@ def _hf_download( from huggingface_hub import snapshot_download from huggingface_hub.utils import GatedRepoError, RepositoryNotFoundError - from llama_stack.distribution.utils.model_utils import model_local_dir + from llama_stack.core.utils.model_utils import model_local_dir repo_id = model.huggingface_repo if repo_id is None: @@ -361,7 +361,7 @@ def _meta_download( info: "LlamaDownloadInfo", max_concurrent_downloads: int, ): - from llama_stack.distribution.utils.model_utils import model_local_dir + from llama_stack.core.utils.model_utils import model_local_dir output_dir = Path(model_local_dir(model.descriptor())) os.makedirs(output_dir, exist_ok=True) @@ -403,7 +403,7 @@ class Manifest(BaseModel): def _download_from_manifest(manifest_file: str, max_concurrent_downloads: int): - from llama_stack.distribution.utils.model_utils import model_local_dir + from llama_stack.core.utils.model_utils import model_local_dir with open(manifest_file) as f: d = json.load(f) diff --git a/llama_stack/cli/model/list.py b/llama_stack/cli/model/list.py index cf84dd526..f46a8c88d 100644 --- a/llama_stack/cli/model/list.py +++ b/llama_stack/cli/model/list.py @@ -11,7 +11,7 @@ from pathlib import Path from llama_stack.cli.subcommand import Subcommand from llama_stack.cli.table import print_table -from llama_stack.distribution.utils.config_dirs import DEFAULT_CHECKPOINT_DIR +from llama_stack.core.utils.config_dirs import DEFAULT_CHECKPOINT_DIR from llama_stack.models.llama.sku_list import all_registered_models diff --git a/llama_stack/cli/model/remove.py b/llama_stack/cli/model/remove.py index 98710d82b..138e06a2a 100644 --- a/llama_stack/cli/model/remove.py +++ b/llama_stack/cli/model/remove.py @@ -9,7 +9,7 @@ import os import shutil from llama_stack.cli.subcommand import Subcommand -from llama_stack.distribution.utils.config_dirs import DEFAULT_CHECKPOINT_DIR +from llama_stack.core.utils.config_dirs import DEFAULT_CHECKPOINT_DIR from llama_stack.models.llama.sku_list import resolve_model diff --git a/llama_stack/cli/stack/_build.py b/llama_stack/cli/stack/_build.py index 3718e456c..ab3662f01 100644 --- a/llama_stack/cli/stack/_build.py +++ b/llama_stack/cli/stack/_build.py @@ -23,27 +23,27 @@ from termcolor import colored, cprint from llama_stack.cli.stack.utils import ImageType from llama_stack.cli.table import print_table -from llama_stack.distribution.build import ( +from llama_stack.core.build import ( SERVER_DEPENDENCIES, build_image, get_provider_dependencies, ) -from llama_stack.distribution.configure import parse_and_maybe_upgrade_config -from llama_stack.distribution.datatypes import ( +from llama_stack.core.configure import parse_and_maybe_upgrade_config +from llama_stack.core.datatypes import ( BuildConfig, BuildProvider, DistributionSpec, Provider, StackRunConfig, ) -from llama_stack.distribution.distribution import get_provider_registry -from llama_stack.distribution.external import load_external_apis -from llama_stack.distribution.resolver import InvalidProviderError -from llama_stack.distribution.stack import replace_env_vars -from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR, EXTERNAL_PROVIDERS_DIR -from llama_stack.distribution.utils.dynamic import instantiate_class_type -from llama_stack.distribution.utils.exec import formulate_run_args, run_command -from llama_stack.distribution.utils.image_types import LlamaStackImageType +from llama_stack.core.distribution import get_provider_registry +from llama_stack.core.external import load_external_apis +from llama_stack.core.resolver import InvalidProviderError +from llama_stack.core.stack import replace_env_vars +from llama_stack.core.utils.config_dirs import DISTRIBS_BASE_DIR, EXTERNAL_PROVIDERS_DIR +from llama_stack.core.utils.dynamic import instantiate_class_type +from llama_stack.core.utils.exec import formulate_run_args, run_command +from llama_stack.core.utils.image_types import LlamaStackImageType from llama_stack.providers.datatypes import Api TEMPLATES_PATH = Path(__file__).parent.parent.parent / "templates" diff --git a/llama_stack/cli/stack/build.py b/llama_stack/cli/stack/build.py index 2c402beeb..279806bdf 100644 --- a/llama_stack/cli/stack/build.py +++ b/llama_stack/cli/stack/build.py @@ -27,7 +27,7 @@ class StackBuild(Subcommand): "--config", type=str, default=None, - help="Path to a config file to use for the build. You can find example configs in llama_stack/distributions/**/build.yaml. If this argument is not provided, you will be prompted to enter information interactively", + help="Path to a config file to use for the build. You can find example configs in llama_stack.cores/**/build.yaml. If this argument is not provided, you will be prompted to enter information interactively", ) self.parser.add_argument( diff --git a/llama_stack/cli/stack/list_apis.py b/llama_stack/cli/stack/list_apis.py index cac803f92..6eed5ca51 100644 --- a/llama_stack/cli/stack/list_apis.py +++ b/llama_stack/cli/stack/list_apis.py @@ -26,7 +26,7 @@ class StackListApis(Subcommand): def _run_apis_list_cmd(self, args: argparse.Namespace) -> None: from llama_stack.cli.table import print_table - from llama_stack.distribution.distribution import stack_apis + from llama_stack.core.distribution import stack_apis # eventually, this should query a registry at llama.meta.com/llamastack/distributions headers = [ diff --git a/llama_stack/cli/stack/list_providers.py b/llama_stack/cli/stack/list_providers.py index deebd937b..b78b3c31f 100644 --- a/llama_stack/cli/stack/list_providers.py +++ b/llama_stack/cli/stack/list_providers.py @@ -23,7 +23,7 @@ class StackListProviders(Subcommand): @property def providable_apis(self): - from llama_stack.distribution.distribution import providable_apis + from llama_stack.core.distribution import providable_apis return [api.value for api in providable_apis()] @@ -38,7 +38,7 @@ class StackListProviders(Subcommand): def _run_providers_list_cmd(self, args: argparse.Namespace) -> None: from llama_stack.cli.table import print_table - from llama_stack.distribution.distribution import Api, get_provider_registry + from llama_stack.core.distribution import Api, get_provider_registry all_providers = get_provider_registry() if args.api: diff --git a/llama_stack/cli/stack/run.py b/llama_stack/cli/stack/run.py index 3cb2e213c..9a366a8b2 100644 --- a/llama_stack/cli/stack/run.py +++ b/llama_stack/cli/stack/run.py @@ -85,8 +85,8 @@ class StackRun(Subcommand): def _run_stack_run_cmd(self, args: argparse.Namespace) -> None: import yaml - from llama_stack.distribution.configure import parse_and_maybe_upgrade_config - from llama_stack.distribution.utils.exec import formulate_run_args, run_command + from llama_stack.core.configure import parse_and_maybe_upgrade_config + from llama_stack.core.utils.exec import formulate_run_args, run_command if args.enable_ui: self._start_ui_development_server(args.port) @@ -94,7 +94,7 @@ class StackRun(Subcommand): if args.config: try: - from llama_stack.distribution.utils.config_resolution import Mode, resolve_config_or_template + from llama_stack.core.utils.config_resolution import Mode, resolve_config_or_template config_file = resolve_config_or_template(args.config, Mode.RUN) except ValueError as e: @@ -127,7 +127,7 @@ class StackRun(Subcommand): # using the current environment packages. if not image_type and not image_name: logger.info("No image type or image name provided. Assuming environment packages.") - from llama_stack.distribution.server.server import main as server_main + from llama_stack.core.server.server import main as server_main # Build the server args from the current args passed to the CLI server_args = argparse.Namespace() diff --git a/llama_stack/cli/verify_download.py b/llama_stack/cli/verify_download.py index 3a1af3cbc..b7f4cfdb5 100644 --- a/llama_stack/cli/verify_download.py +++ b/llama_stack/cli/verify_download.py @@ -107,7 +107,7 @@ def verify_files(model_dir: Path, checksums: dict[str, str], console: Console) - def run_verify_cmd(args: argparse.Namespace, parser: argparse.ArgumentParser): - from llama_stack.distribution.utils.model_utils import model_local_dir + from llama_stack.core.utils.model_utils import model_local_dir console = Console() model_dir = Path(model_local_dir(args.model_id)) diff --git a/llama_stack/distribution/__init__.py b/llama_stack/core/__init__.py similarity index 100% rename from llama_stack/distribution/__init__.py rename to llama_stack/core/__init__.py diff --git a/llama_stack/distribution/access_control/__init__.py b/llama_stack/core/access_control/__init__.py similarity index 100% rename from llama_stack/distribution/access_control/__init__.py rename to llama_stack/core/access_control/__init__.py diff --git a/llama_stack/distribution/access_control/access_control.py b/llama_stack/core/access_control/access_control.py similarity index 98% rename from llama_stack/distribution/access_control/access_control.py rename to llama_stack/core/access_control/access_control.py index 64c0122c1..bde5cfd76 100644 --- a/llama_stack/distribution/access_control/access_control.py +++ b/llama_stack/core/access_control/access_control.py @@ -6,7 +6,7 @@ from typing import Any -from llama_stack.distribution.datatypes import User +from llama_stack.core.datatypes import User from .conditions import ( Condition, diff --git a/llama_stack/distribution/access_control/conditions.py b/llama_stack/core/access_control/conditions.py similarity index 100% rename from llama_stack/distribution/access_control/conditions.py rename to llama_stack/core/access_control/conditions.py diff --git a/llama_stack/distribution/access_control/datatypes.py b/llama_stack/core/access_control/datatypes.py similarity index 100% rename from llama_stack/distribution/access_control/datatypes.py rename to llama_stack/core/access_control/datatypes.py diff --git a/llama_stack/distribution/build.py b/llama_stack/core/build.py similarity index 93% rename from llama_stack/distribution/build.py rename to llama_stack/core/build.py index b4eaac1c7..a6544435f 100644 --- a/llama_stack/distribution/build.py +++ b/llama_stack/core/build.py @@ -12,11 +12,11 @@ from pathlib import Path from pydantic import BaseModel from termcolor import cprint -from llama_stack.distribution.datatypes import BuildConfig -from llama_stack.distribution.distribution import get_provider_registry -from llama_stack.distribution.external import load_external_apis -from llama_stack.distribution.utils.exec import run_command -from llama_stack.distribution.utils.image_types import LlamaStackImageType +from llama_stack.core.datatypes import BuildConfig +from llama_stack.core.distribution import get_provider_registry +from llama_stack.core.external import load_external_apis +from llama_stack.core.utils.exec import run_command +from llama_stack.core.utils.image_types import LlamaStackImageType from llama_stack.providers.datatypes import Api from llama_stack.templates.template import DistributionTemplate @@ -122,7 +122,7 @@ def build_image( normal_deps.extend(api_spec.pip_packages) if build_config.image_type == LlamaStackImageType.CONTAINER.value: - script = str(importlib.resources.files("llama_stack") / "distribution/build_container.sh") + script = str(importlib.resources.files("llama_stack") / "core/build_container.sh") args = [ script, "--template-or-config", @@ -139,7 +139,7 @@ def build_image( if run_config is not None: args.extend(["--run-config", run_config]) elif build_config.image_type == LlamaStackImageType.CONDA.value: - script = str(importlib.resources.files("llama_stack") / "distribution/build_conda_env.sh") + script = str(importlib.resources.files("llama_stack") / "core/build_conda_env.sh") args = [ script, "--env-name", @@ -150,7 +150,7 @@ def build_image( " ".join(normal_deps), ] elif build_config.image_type == LlamaStackImageType.VENV.value: - script = str(importlib.resources.files("llama_stack") / "distribution/build_venv.sh") + script = str(importlib.resources.files("llama_stack") / "core/build_venv.sh") args = [ script, "--env-name", diff --git a/llama_stack/distribution/build_conda_env.sh b/llama_stack/core/build_conda_env.sh similarity index 100% rename from llama_stack/distribution/build_conda_env.sh rename to llama_stack/core/build_conda_env.sh diff --git a/llama_stack/distribution/build_container.sh b/llama_stack/core/build_container.sh similarity index 98% rename from llama_stack/distribution/build_container.sh rename to llama_stack/core/build_container.sh index 50d8e4925..1376aaa28 100755 --- a/llama_stack/distribution/build_container.sh +++ b/llama_stack/core/build_container.sh @@ -327,12 +327,12 @@ EOF # If a run config is provided, we use the --config flag if [[ -n "$run_config" ]]; then add_to_container << EOF -ENTRYPOINT ["python", "-m", "llama_stack.distribution.server.server", "--config", "$RUN_CONFIG_PATH"] +ENTRYPOINT ["python", "-m", "llama_stack.core.server.server", "--config", "$RUN_CONFIG_PATH"] EOF # If a template is provided (not a yaml file), we use the --template flag elif [[ "$template_or_config" != *.yaml ]]; then add_to_container << EOF -ENTRYPOINT ["python", "-m", "llama_stack.distribution.server.server", "--template", "$template_or_config"] +ENTRYPOINT ["python", "-m", "llama_stack.core.server.server", "--template", "$template_or_config"] EOF fi diff --git a/llama_stack/distribution/build_venv.sh b/llama_stack/core/build_venv.sh similarity index 100% rename from llama_stack/distribution/build_venv.sh rename to llama_stack/core/build_venv.sh diff --git a/llama_stack/distribution/client.py b/llama_stack/core/client.py similarity index 100% rename from llama_stack/distribution/client.py rename to llama_stack/core/client.py diff --git a/llama_stack/distribution/common.sh b/llama_stack/core/common.sh similarity index 100% rename from llama_stack/distribution/common.sh rename to llama_stack/core/common.sh diff --git a/llama_stack/distribution/configure.py b/llama_stack/core/configure.py similarity index 93% rename from llama_stack/distribution/configure.py rename to llama_stack/core/configure.py index 20be040a0..9e18b438c 100644 --- a/llama_stack/distribution/configure.py +++ b/llama_stack/core/configure.py @@ -7,20 +7,20 @@ import logging import textwrap from typing import Any -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( LLAMA_STACK_RUN_CONFIG_VERSION, DistributionSpec, Provider, StackRunConfig, ) -from llama_stack.distribution.distribution import ( +from llama_stack.core.distribution import ( builtin_automatically_routed_apis, get_provider_registry, ) -from llama_stack.distribution.stack import cast_image_name_to_string, replace_env_vars -from llama_stack.distribution.utils.config_dirs import EXTERNAL_PROVIDERS_DIR -from llama_stack.distribution.utils.dynamic import instantiate_class_type -from llama_stack.distribution.utils.prompt_for_config import prompt_for_config +from llama_stack.core.stack import cast_image_name_to_string, replace_env_vars +from llama_stack.core.utils.config_dirs import EXTERNAL_PROVIDERS_DIR +from llama_stack.core.utils.dynamic import instantiate_class_type +from llama_stack.core.utils.prompt_for_config import prompt_for_config from llama_stack.providers.datatypes import Api, ProviderSpec logger = logging.getLogger(__name__) diff --git a/llama_stack/distribution/datatypes.py b/llama_stack/core/datatypes.py similarity index 99% rename from llama_stack/distribution/datatypes.py rename to llama_stack/core/datatypes.py index 60c317337..c66d510cc 100644 --- a/llama_stack/distribution/datatypes.py +++ b/llama_stack/core/datatypes.py @@ -24,7 +24,7 @@ from llama_stack.apis.shields import Shield, ShieldInput from llama_stack.apis.tools import Tool, ToolGroup, ToolGroupInput, ToolRuntime from llama_stack.apis.vector_dbs import VectorDB, VectorDBInput from llama_stack.apis.vector_io import VectorIO -from llama_stack.distribution.access_control.datatypes import AccessRule +from llama_stack.core.access_control.datatypes import AccessRule from llama_stack.providers.datatypes import Api, ProviderSpec from llama_stack.providers.utils.kvstore.config import KVStoreConfig, SqliteKVStoreConfig from llama_stack.providers.utils.sqlstore.sqlstore import SqlStoreConfig diff --git a/llama_stack/distribution/distribution.py b/llama_stack/core/distribution.py similarity index 98% rename from llama_stack/distribution/distribution.py rename to llama_stack/core/distribution.py index 6e7297e32..977eb5393 100644 --- a/llama_stack/distribution/distribution.py +++ b/llama_stack/core/distribution.py @@ -12,8 +12,8 @@ from typing import Any import yaml from pydantic import BaseModel -from llama_stack.distribution.datatypes import BuildConfig, DistributionSpec -from llama_stack.distribution.external import load_external_apis +from llama_stack.core.datatypes import BuildConfig, DistributionSpec +from llama_stack.core.external import load_external_apis from llama_stack.log import get_logger from llama_stack.providers.datatypes import ( AdapterSpec, diff --git a/llama_stack/distribution/external.py b/llama_stack/core/external.py similarity index 96% rename from llama_stack/distribution/external.py rename to llama_stack/core/external.py index 0a7da16b1..12e9824ad 100644 --- a/llama_stack/distribution/external.py +++ b/llama_stack/core/external.py @@ -8,7 +8,7 @@ import yaml from llama_stack.apis.datatypes import Api, ExternalApiSpec -from llama_stack.distribution.datatypes import BuildConfig, StackRunConfig +from llama_stack.core.datatypes import BuildConfig, StackRunConfig from llama_stack.log import get_logger logger = get_logger(name=__name__, category="core") diff --git a/llama_stack/distribution/inspect.py b/llama_stack/core/inspect.py similarity index 93% rename from llama_stack/distribution/inspect.py rename to llama_stack/core/inspect.py index f62de4f6b..37dab4199 100644 --- a/llama_stack/distribution/inspect.py +++ b/llama_stack/core/inspect.py @@ -15,9 +15,9 @@ from llama_stack.apis.inspect import ( RouteInfo, VersionInfo, ) -from llama_stack.distribution.datatypes import StackRunConfig -from llama_stack.distribution.external import load_external_apis -from llama_stack.distribution.server.routes import get_all_api_routes +from llama_stack.core.datatypes import StackRunConfig +from llama_stack.core.external import load_external_apis +from llama_stack.core.server.routes import get_all_api_routes from llama_stack.providers.datatypes import HealthStatus diff --git a/llama_stack/distribution/library_client.py b/llama_stack/core/library_client.py similarity index 96% rename from llama_stack/distribution/library_client.py rename to llama_stack/core/library_client.py index 43ee7f417..d5020d16b 100644 --- a/llama_stack/distribution/library_client.py +++ b/llama_stack/core/library_client.py @@ -31,23 +31,23 @@ from pydantic import BaseModel, TypeAdapter from rich.console import Console from termcolor import cprint -from llama_stack.distribution.build import print_pip_install_help -from llama_stack.distribution.configure import parse_and_maybe_upgrade_config -from llama_stack.distribution.datatypes import Api, BuildConfig, BuildProvider, DistributionSpec -from llama_stack.distribution.request_headers import ( +from llama_stack.core.build import print_pip_install_help +from llama_stack.core.configure import parse_and_maybe_upgrade_config +from llama_stack.core.datatypes import Api, BuildConfig, BuildProvider, DistributionSpec +from llama_stack.core.request_headers import ( PROVIDER_DATA_VAR, request_provider_data_context, ) -from llama_stack.distribution.resolver import ProviderRegistry -from llama_stack.distribution.server.routes import RouteImpls, find_matching_route, initialize_route_impls -from llama_stack.distribution.stack import ( +from llama_stack.core.resolver import ProviderRegistry +from llama_stack.core.server.routes import RouteImpls, find_matching_route, initialize_route_impls +from llama_stack.core.stack import ( construct_stack, get_stack_run_config_from_template, replace_env_vars, ) -from llama_stack.distribution.utils.config import redact_sensitive_fields -from llama_stack.distribution.utils.context import preserve_contexts_async_generator -from llama_stack.distribution.utils.exec import in_notebook +from llama_stack.core.utils.config import redact_sensitive_fields +from llama_stack.core.utils.context import preserve_contexts_async_generator +from llama_stack.core.utils.exec import in_notebook from llama_stack.providers.utils.telemetry.tracing import ( CURRENT_TRACE_CONTEXT, end_trace, diff --git a/llama_stack/distribution/providers.py b/llama_stack/core/providers.py similarity index 100% rename from llama_stack/distribution/providers.py rename to llama_stack/core/providers.py diff --git a/llama_stack/distribution/request_headers.py b/llama_stack/core/request_headers.py similarity index 98% rename from llama_stack/distribution/request_headers.py rename to llama_stack/core/request_headers.py index 509c2be44..35ac72775 100644 --- a/llama_stack/distribution/request_headers.py +++ b/llama_stack/core/request_headers.py @@ -10,7 +10,7 @@ import logging from contextlib import AbstractContextManager from typing import Any -from llama_stack.distribution.datatypes import User +from llama_stack.core.datatypes import User from .utils.dynamic import instantiate_class_type diff --git a/llama_stack/distribution/resolver.py b/llama_stack/core/resolver.py similarity index 97% rename from llama_stack/distribution/resolver.py rename to llama_stack/core/resolver.py index db6856ed2..70c78fb01 100644 --- a/llama_stack/distribution/resolver.py +++ b/llama_stack/core/resolver.py @@ -27,18 +27,18 @@ from llama_stack.apis.telemetry import Telemetry from llama_stack.apis.tools import ToolGroups, ToolRuntime from llama_stack.apis.vector_dbs import VectorDBs from llama_stack.apis.vector_io import VectorIO -from llama_stack.distribution.client import get_client_impl -from llama_stack.distribution.datatypes import ( +from llama_stack.core.client import get_client_impl +from llama_stack.core.datatypes import ( AccessRule, AutoRoutedProviderSpec, Provider, RoutingTableProviderSpec, StackRunConfig, ) -from llama_stack.distribution.distribution import builtin_automatically_routed_apis -from llama_stack.distribution.external import load_external_apis -from llama_stack.distribution.store import DistributionRegistry -from llama_stack.distribution.utils.dynamic import instantiate_class_type +from llama_stack.core.distribution import builtin_automatically_routed_apis +from llama_stack.core.external import load_external_apis +from llama_stack.core.store import DistributionRegistry +from llama_stack.core.utils.dynamic import instantiate_class_type from llama_stack.log import get_logger from llama_stack.providers.datatypes import ( Api, @@ -183,7 +183,7 @@ def specs_for_autorouted_apis(apis_to_serve: list[str] | set[str]) -> dict[str, spec=RoutingTableProviderSpec( api=info.routing_table_api, router_api=info.router_api, - module="llama_stack.distribution.routers", + module="llama_stack.core.routers", api_dependencies=[], deps__=[f"inner-{info.router_api.value}"], ), @@ -197,7 +197,7 @@ def specs_for_autorouted_apis(apis_to_serve: list[str] | set[str]) -> dict[str, config={}, spec=AutoRoutedProviderSpec( api=info.router_api, - module="llama_stack.distribution.routers", + module="llama_stack.core.routers", routing_table_api=info.routing_table_api, api_dependencies=[info.routing_table_api], # Add telemetry as an optional dependency to all auto-routed providers diff --git a/llama_stack/distribution/routers/__init__.py b/llama_stack/core/routers/__init__.py similarity index 94% rename from llama_stack/distribution/routers/__init__.py rename to llama_stack/core/routers/__init__.py index 8671a62e1..1faace34a 100644 --- a/llama_stack/distribution/routers/__init__.py +++ b/llama_stack/core/routers/__init__.py @@ -6,9 +6,9 @@ from typing import Any -from llama_stack.distribution.datatypes import AccessRule, RoutedProtocol -from llama_stack.distribution.stack import StackRunConfig -from llama_stack.distribution.store import DistributionRegistry +from llama_stack.core.datatypes import AccessRule, RoutedProtocol +from llama_stack.core.stack import StackRunConfig +from llama_stack.core.store import DistributionRegistry from llama_stack.providers.datatypes import Api, RoutingTable from llama_stack.providers.utils.inference.inference_store import InferenceStore diff --git a/llama_stack/distribution/routers/datasets.py b/llama_stack/core/routers/datasets.py similarity index 100% rename from llama_stack/distribution/routers/datasets.py rename to llama_stack/core/routers/datasets.py diff --git a/llama_stack/distribution/routers/eval_scoring.py b/llama_stack/core/routers/eval_scoring.py similarity index 100% rename from llama_stack/distribution/routers/eval_scoring.py rename to llama_stack/core/routers/eval_scoring.py diff --git a/llama_stack/distribution/routers/inference.py b/llama_stack/core/routers/inference.py similarity index 100% rename from llama_stack/distribution/routers/inference.py rename to llama_stack/core/routers/inference.py diff --git a/llama_stack/distribution/routers/safety.py b/llama_stack/core/routers/safety.py similarity index 100% rename from llama_stack/distribution/routers/safety.py rename to llama_stack/core/routers/safety.py diff --git a/llama_stack/distribution/routers/tool_runtime.py b/llama_stack/core/routers/tool_runtime.py similarity index 100% rename from llama_stack/distribution/routers/tool_runtime.py rename to llama_stack/core/routers/tool_runtime.py diff --git a/llama_stack/distribution/routers/vector_io.py b/llama_stack/core/routers/vector_io.py similarity index 100% rename from llama_stack/distribution/routers/vector_io.py rename to llama_stack/core/routers/vector_io.py diff --git a/llama_stack/distribution/routing_tables/__init__.py b/llama_stack/core/routing_tables/__init__.py similarity index 100% rename from llama_stack/distribution/routing_tables/__init__.py rename to llama_stack/core/routing_tables/__init__.py diff --git a/llama_stack/distribution/routing_tables/benchmarks.py b/llama_stack/core/routing_tables/benchmarks.py similarity index 97% rename from llama_stack/distribution/routing_tables/benchmarks.py rename to llama_stack/core/routing_tables/benchmarks.py index 815483494..74bee8040 100644 --- a/llama_stack/distribution/routing_tables/benchmarks.py +++ b/llama_stack/core/routing_tables/benchmarks.py @@ -7,7 +7,7 @@ from typing import Any from llama_stack.apis.benchmarks import Benchmark, Benchmarks, ListBenchmarksResponse -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( BenchmarkWithOwner, ) from llama_stack.log import get_logger diff --git a/llama_stack/distribution/routing_tables/common.py b/llama_stack/core/routing_tables/common.py similarity index 96% rename from llama_stack/distribution/routing_tables/common.py rename to llama_stack/core/routing_tables/common.py index a759ea8dd..4be3de42d 100644 --- a/llama_stack/distribution/routing_tables/common.py +++ b/llama_stack/core/routing_tables/common.py @@ -10,16 +10,16 @@ from llama_stack.apis.common.errors import ModelNotFoundError from llama_stack.apis.models import Model from llama_stack.apis.resource import ResourceType from llama_stack.apis.scoring_functions import ScoringFn -from llama_stack.distribution.access_control.access_control import AccessDeniedError, is_action_allowed -from llama_stack.distribution.access_control.datatypes import Action -from llama_stack.distribution.datatypes import ( +from llama_stack.core.access_control.access_control import AccessDeniedError, is_action_allowed +from llama_stack.core.access_control.datatypes import Action +from llama_stack.core.datatypes import ( AccessRule, RoutableObject, RoutableObjectWithProvider, RoutedProtocol, ) -from llama_stack.distribution.request_headers import get_authenticated_user -from llama_stack.distribution.store import DistributionRegistry +from llama_stack.core.request_headers import get_authenticated_user +from llama_stack.core.store import DistributionRegistry from llama_stack.log import get_logger from llama_stack.providers.datatypes import Api, RoutingTable diff --git a/llama_stack/distribution/routing_tables/datasets.py b/llama_stack/core/routing_tables/datasets.py similarity index 98% rename from llama_stack/distribution/routing_tables/datasets.py rename to llama_stack/core/routing_tables/datasets.py index 89da7d081..fc6a75df4 100644 --- a/llama_stack/distribution/routing_tables/datasets.py +++ b/llama_stack/core/routing_tables/datasets.py @@ -19,7 +19,7 @@ from llama_stack.apis.datasets import ( URIDataSource, ) from llama_stack.apis.resource import ResourceType -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( DatasetWithOwner, ) from llama_stack.log import get_logger diff --git a/llama_stack/distribution/routing_tables/models.py b/llama_stack/core/routing_tables/models.py similarity index 99% rename from llama_stack/distribution/routing_tables/models.py rename to llama_stack/core/routing_tables/models.py index ae1fe2882..c76619271 100644 --- a/llama_stack/distribution/routing_tables/models.py +++ b/llama_stack/core/routing_tables/models.py @@ -9,7 +9,7 @@ from typing import Any from llama_stack.apis.common.errors import ModelNotFoundError from llama_stack.apis.models import ListModelsResponse, Model, Models, ModelType, OpenAIListModelsResponse, OpenAIModel -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( ModelWithOwner, RegistryEntrySource, ) diff --git a/llama_stack/distribution/routing_tables/scoring_functions.py b/llama_stack/core/routing_tables/scoring_functions.py similarity index 97% rename from llama_stack/distribution/routing_tables/scoring_functions.py rename to llama_stack/core/routing_tables/scoring_functions.py index 742cc3ca6..5874ba941 100644 --- a/llama_stack/distribution/routing_tables/scoring_functions.py +++ b/llama_stack/core/routing_tables/scoring_functions.py @@ -12,7 +12,7 @@ from llama_stack.apis.scoring_functions import ( ScoringFnParams, ScoringFunctions, ) -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( ScoringFnWithOwner, ) from llama_stack.log import get_logger diff --git a/llama_stack/distribution/routing_tables/shields.py b/llama_stack/core/routing_tables/shields.py similarity index 97% rename from llama_stack/distribution/routing_tables/shields.py rename to llama_stack/core/routing_tables/shields.py index 5215981b9..0c592601a 100644 --- a/llama_stack/distribution/routing_tables/shields.py +++ b/llama_stack/core/routing_tables/shields.py @@ -8,7 +8,7 @@ from typing import Any from llama_stack.apis.resource import ResourceType from llama_stack.apis.shields import ListShieldsResponse, Shield, Shields -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( ShieldWithOwner, ) from llama_stack.log import get_logger diff --git a/llama_stack/distribution/routing_tables/toolgroups.py b/llama_stack/core/routing_tables/toolgroups.py similarity index 98% rename from llama_stack/distribution/routing_tables/toolgroups.py rename to llama_stack/core/routing_tables/toolgroups.py index 22c4e109a..a6d15796a 100644 --- a/llama_stack/distribution/routing_tables/toolgroups.py +++ b/llama_stack/core/routing_tables/toolgroups.py @@ -8,7 +8,7 @@ from typing import Any from llama_stack.apis.common.content_types import URL from llama_stack.apis.tools import ListToolGroupsResponse, ListToolsResponse, Tool, ToolGroup, ToolGroups -from llama_stack.distribution.datatypes import ToolGroupWithOwner +from llama_stack.core.datatypes import ToolGroupWithOwner from llama_stack.log import get_logger from .common import CommonRoutingTableImpl diff --git a/llama_stack/distribution/routing_tables/vector_dbs.py b/llama_stack/core/routing_tables/vector_dbs.py similarity index 99% rename from llama_stack/distribution/routing_tables/vector_dbs.py rename to llama_stack/core/routing_tables/vector_dbs.py index aa61ea2fd..5ea4e9854 100644 --- a/llama_stack/distribution/routing_tables/vector_dbs.py +++ b/llama_stack/core/routing_tables/vector_dbs.py @@ -23,7 +23,7 @@ from llama_stack.apis.vector_io.vector_io import ( VectorStoreObject, VectorStoreSearchResponsePage, ) -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( VectorDBWithOwner, ) from llama_stack.log import get_logger diff --git a/llama_stack/distribution/server/__init__.py b/llama_stack/core/server/__init__.py similarity index 100% rename from llama_stack/distribution/server/__init__.py rename to llama_stack/core/server/__init__.py diff --git a/llama_stack/distribution/server/auth.py b/llama_stack/core/server/auth.py similarity index 95% rename from llama_stack/distribution/server/auth.py rename to llama_stack/core/server/auth.py index 87c1a2ab6..e4fb4ff2b 100644 --- a/llama_stack/distribution/server/auth.py +++ b/llama_stack/core/server/auth.py @@ -9,10 +9,10 @@ import json import httpx from aiohttp import hdrs -from llama_stack.distribution.datatypes import AuthenticationConfig, User -from llama_stack.distribution.request_headers import user_from_scope -from llama_stack.distribution.server.auth_providers import create_auth_provider -from llama_stack.distribution.server.routes import find_matching_route, initialize_route_impls +from llama_stack.core.datatypes import AuthenticationConfig, User +from llama_stack.core.request_headers import user_from_scope +from llama_stack.core.server.auth_providers import create_auth_provider +from llama_stack.core.server.routes import find_matching_route, initialize_route_impls from llama_stack.log import get_logger logger = get_logger(name=__name__, category="auth") diff --git a/llama_stack/distribution/server/auth_providers.py b/llama_stack/core/server/auth_providers.py similarity index 99% rename from llama_stack/distribution/server/auth_providers.py rename to llama_stack/core/server/auth_providers.py index 9b0e182f5..73d5581c2 100644 --- a/llama_stack/distribution/server/auth_providers.py +++ b/llama_stack/core/server/auth_providers.py @@ -14,7 +14,7 @@ import httpx from jose import jwt from pydantic import BaseModel, Field -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( AuthenticationConfig, CustomAuthConfig, GitHubTokenAuthConfig, diff --git a/llama_stack/distribution/server/quota.py b/llama_stack/core/server/quota.py similarity index 100% rename from llama_stack/distribution/server/quota.py rename to llama_stack/core/server/quota.py diff --git a/llama_stack/distribution/server/routes.py b/llama_stack/core/server/routes.py similarity index 98% rename from llama_stack/distribution/server/routes.py rename to llama_stack/core/server/routes.py index ca6f629af..7baf20da5 100644 --- a/llama_stack/distribution/server/routes.py +++ b/llama_stack/core/server/routes.py @@ -15,7 +15,7 @@ from starlette.routing import Route from llama_stack.apis.datatypes import Api, ExternalApiSpec from llama_stack.apis.tools import RAGToolRuntime, SpecialToolGroup from llama_stack.apis.version import LLAMA_STACK_API_VERSION -from llama_stack.distribution.resolver import api_protocol_map +from llama_stack.core.resolver import api_protocol_map from llama_stack.schema_utils import WebMethod EndpointFunc = Callable[..., Any] diff --git a/llama_stack/distribution/server/server.py b/llama_stack/core/server/server.py similarity index 96% rename from llama_stack/distribution/server/server.py rename to llama_stack/core/server/server.py index 96a0d60e7..481c6c321 100644 --- a/llama_stack/distribution/server/server.py +++ b/llama_stack/core/server/server.py @@ -33,35 +33,35 @@ from pydantic import BaseModel, ValidationError from llama_stack.apis.common.responses import PaginatedResponse from llama_stack.cli.utils import add_config_template_args, get_config_from_args -from llama_stack.distribution.access_control.access_control import AccessDeniedError -from llama_stack.distribution.datatypes import ( +from llama_stack.core.access_control.access_control import AccessDeniedError +from llama_stack.core.datatypes import ( AuthenticationRequiredError, LoggingConfig, StackRunConfig, ) -from llama_stack.distribution.distribution import builtin_automatically_routed_apis -from llama_stack.distribution.external import ExternalApiSpec, load_external_apis -from llama_stack.distribution.request_headers import ( +from llama_stack.core.distribution import builtin_automatically_routed_apis +from llama_stack.core.external import ExternalApiSpec, load_external_apis +from llama_stack.core.request_headers import ( PROVIDER_DATA_VAR, request_provider_data_context, user_from_scope, ) -from llama_stack.distribution.resolver import InvalidProviderError -from llama_stack.distribution.server.routes import ( +from llama_stack.core.resolver import InvalidProviderError +from llama_stack.core.server.routes import ( find_matching_route, get_all_api_routes, initialize_route_impls, ) -from llama_stack.distribution.stack import ( +from llama_stack.core.stack import ( cast_image_name_to_string, construct_stack, replace_env_vars, shutdown_stack, validate_env_pair, ) -from llama_stack.distribution.utils.config import redact_sensitive_fields -from llama_stack.distribution.utils.config_resolution import Mode, resolve_config_or_template -from llama_stack.distribution.utils.context import preserve_contexts_async_generator +from llama_stack.core.utils.config import redact_sensitive_fields +from llama_stack.core.utils.config_resolution import Mode, resolve_config_or_template +from llama_stack.core.utils.context import preserve_contexts_async_generator from llama_stack.log import get_logger from llama_stack.providers.datatypes import Api from llama_stack.providers.inline.telemetry.meta_reference.config import TelemetryConfig diff --git a/llama_stack/distribution/stack.py b/llama_stack/core/stack.py similarity index 96% rename from llama_stack/distribution/stack.py rename to llama_stack/core/stack.py index 1dbcbb7fa..c7fe67503 100644 --- a/llama_stack/distribution/stack.py +++ b/llama_stack/core/stack.py @@ -34,14 +34,14 @@ from llama_stack.apis.telemetry import Telemetry from llama_stack.apis.tools import RAGToolRuntime, ToolGroups, ToolRuntime from llama_stack.apis.vector_dbs import VectorDBs from llama_stack.apis.vector_io import VectorIO -from llama_stack.distribution.datatypes import Provider, StackRunConfig -from llama_stack.distribution.distribution import get_provider_registry -from llama_stack.distribution.inspect import DistributionInspectConfig, DistributionInspectImpl -from llama_stack.distribution.providers import ProviderImpl, ProviderImplConfig -from llama_stack.distribution.resolver import ProviderRegistry, resolve_impls -from llama_stack.distribution.routing_tables.common import CommonRoutingTableImpl -from llama_stack.distribution.store.registry import create_dist_registry -from llama_stack.distribution.utils.dynamic import instantiate_class_type +from llama_stack.core.datatypes import Provider, StackRunConfig +from llama_stack.core.distribution import get_provider_registry +from llama_stack.core.inspect import DistributionInspectConfig, DistributionInspectImpl +from llama_stack.core.providers import ProviderImpl, ProviderImplConfig +from llama_stack.core.resolver import ProviderRegistry, resolve_impls +from llama_stack.core.routing_tables.common import CommonRoutingTableImpl +from llama_stack.core.store.registry import create_dist_registry +from llama_stack.core.utils.dynamic import instantiate_class_type from llama_stack.log import get_logger from llama_stack.providers.datatypes import Api diff --git a/llama_stack/distribution/start_stack.sh b/llama_stack/core/start_stack.sh similarity index 98% rename from llama_stack/distribution/start_stack.sh rename to llama_stack/core/start_stack.sh index 77a7dc92e..ff8275aeb 100755 --- a/llama_stack/distribution/start_stack.sh +++ b/llama_stack/core/start_stack.sh @@ -122,7 +122,7 @@ if [[ "$env_type" == "venv" || "$env_type" == "conda" ]]; then yaml_config_arg="" fi - $PYTHON_BINARY -m llama_stack.distribution.server.server \ + $PYTHON_BINARY -m llama_stack.core.server.server \ $yaml_config_arg \ --port "$port" \ $env_vars \ diff --git a/llama_stack/distribution/store/__init__.py b/llama_stack/core/store/__init__.py similarity index 100% rename from llama_stack/distribution/store/__init__.py rename to llama_stack/core/store/__init__.py diff --git a/llama_stack/distribution/store/registry.py b/llama_stack/core/store/registry.py similarity index 98% rename from llama_stack/distribution/store/registry.py rename to llama_stack/core/store/registry.py index cd7cd9f00..4b60e1001 100644 --- a/llama_stack/distribution/store/registry.py +++ b/llama_stack/core/store/registry.py @@ -10,8 +10,8 @@ from typing import Protocol import pydantic -from llama_stack.distribution.datatypes import RoutableObjectWithProvider -from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR +from llama_stack.core.datatypes import RoutableObjectWithProvider +from llama_stack.core.utils.config_dirs import DISTRIBS_BASE_DIR from llama_stack.log import get_logger from llama_stack.providers.utils.kvstore import KVStore, kvstore_impl from llama_stack.providers.utils.kvstore.config import KVStoreConfig, SqliteKVStoreConfig diff --git a/llama_stack/distribution/ui/Containerfile b/llama_stack/core/ui/Containerfile similarity index 100% rename from llama_stack/distribution/ui/Containerfile rename to llama_stack/core/ui/Containerfile diff --git a/llama_stack/distribution/ui/README.md b/llama_stack/core/ui/README.md similarity index 96% rename from llama_stack/distribution/ui/README.md rename to llama_stack/core/ui/README.md index 51c2d2bc2..5f4a9e250 100644 --- a/llama_stack/distribution/ui/README.md +++ b/llama_stack/core/ui/README.md @@ -36,7 +36,7 @@ llama-stack-client benchmarks register \ 3. Start Streamlit UI ```bash -uv run --with ".[ui]" streamlit run llama_stack/distribution/ui/app.py +uv run --with ".[ui]" streamlit run llama_stack.core/ui/app.py ``` ## Environment Variables diff --git a/llama_stack/distribution/ui/__init__.py b/llama_stack/core/ui/__init__.py similarity index 100% rename from llama_stack/distribution/ui/__init__.py rename to llama_stack/core/ui/__init__.py diff --git a/llama_stack/distribution/ui/app.py b/llama_stack/core/ui/app.py similarity index 100% rename from llama_stack/distribution/ui/app.py rename to llama_stack/core/ui/app.py diff --git a/llama_stack/distribution/ui/modules/__init__.py b/llama_stack/core/ui/modules/__init__.py similarity index 100% rename from llama_stack/distribution/ui/modules/__init__.py rename to llama_stack/core/ui/modules/__init__.py diff --git a/llama_stack/distribution/ui/modules/api.py b/llama_stack/core/ui/modules/api.py similarity index 100% rename from llama_stack/distribution/ui/modules/api.py rename to llama_stack/core/ui/modules/api.py diff --git a/llama_stack/distribution/ui/modules/utils.py b/llama_stack/core/ui/modules/utils.py similarity index 100% rename from llama_stack/distribution/ui/modules/utils.py rename to llama_stack/core/ui/modules/utils.py diff --git a/llama_stack/distribution/ui/page/__init__.py b/llama_stack/core/ui/page/__init__.py similarity index 100% rename from llama_stack/distribution/ui/page/__init__.py rename to llama_stack/core/ui/page/__init__.py diff --git a/llama_stack/distribution/ui/page/distribution/__init__.py b/llama_stack/core/ui/page/distribution/__init__.py similarity index 100% rename from llama_stack/distribution/ui/page/distribution/__init__.py rename to llama_stack/core/ui/page/distribution/__init__.py diff --git a/llama_stack/distribution/ui/page/distribution/datasets.py b/llama_stack/core/ui/page/distribution/datasets.py similarity index 88% rename from llama_stack/distribution/ui/page/distribution/datasets.py rename to llama_stack/core/ui/page/distribution/datasets.py index 6842b29a7..aab0901ac 100644 --- a/llama_stack/distribution/ui/page/distribution/datasets.py +++ b/llama_stack/core/ui/page/distribution/datasets.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def datasets(): diff --git a/llama_stack/distribution/ui/page/distribution/eval_tasks.py b/llama_stack/core/ui/page/distribution/eval_tasks.py similarity index 90% rename from llama_stack/distribution/ui/page/distribution/eval_tasks.py rename to llama_stack/core/ui/page/distribution/eval_tasks.py index 492be4700..1a0ce502b 100644 --- a/llama_stack/distribution/ui/page/distribution/eval_tasks.py +++ b/llama_stack/core/ui/page/distribution/eval_tasks.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def benchmarks(): diff --git a/llama_stack/distribution/ui/page/distribution/models.py b/llama_stack/core/ui/page/distribution/models.py similarity index 87% rename from llama_stack/distribution/ui/page/distribution/models.py rename to llama_stack/core/ui/page/distribution/models.py index f29459098..f84508746 100644 --- a/llama_stack/distribution/ui/page/distribution/models.py +++ b/llama_stack/core/ui/page/distribution/models.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def models(): diff --git a/llama_stack/distribution/ui/page/distribution/providers.py b/llama_stack/core/ui/page/distribution/providers.py similarity index 91% rename from llama_stack/distribution/ui/page/distribution/providers.py rename to llama_stack/core/ui/page/distribution/providers.py index c660cb986..3ec6026d1 100644 --- a/llama_stack/distribution/ui/page/distribution/providers.py +++ b/llama_stack/core/ui/page/distribution/providers.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def providers(): diff --git a/llama_stack/distribution/ui/page/distribution/resources.py b/llama_stack/core/ui/page/distribution/resources.py similarity index 70% rename from llama_stack/distribution/ui/page/distribution/resources.py rename to llama_stack/core/ui/page/distribution/resources.py index 5e10e6e80..c56fcfff3 100644 --- a/llama_stack/distribution/ui/page/distribution/resources.py +++ b/llama_stack/core/ui/page/distribution/resources.py @@ -6,12 +6,12 @@ from streamlit_option_menu import option_menu -from llama_stack.distribution.ui.page.distribution.datasets import datasets -from llama_stack.distribution.ui.page.distribution.eval_tasks import benchmarks -from llama_stack.distribution.ui.page.distribution.models import models -from llama_stack.distribution.ui.page.distribution.scoring_functions import scoring_functions -from llama_stack.distribution.ui.page.distribution.shields import shields -from llama_stack.distribution.ui.page.distribution.vector_dbs import vector_dbs +from llama_stack.core.ui.page.distribution.datasets import datasets +from llama_stack.core.ui.page.distribution.eval_tasks import benchmarks +from llama_stack.core.ui.page.distribution.models import models +from llama_stack.core.ui.page.distribution.scoring_functions import scoring_functions +from llama_stack.core.ui.page.distribution.shields import shields +from llama_stack.core.ui.page.distribution.vector_dbs import vector_dbs def resources_page(): diff --git a/llama_stack/distribution/ui/page/distribution/scoring_functions.py b/llama_stack/core/ui/page/distribution/scoring_functions.py similarity index 89% rename from llama_stack/distribution/ui/page/distribution/scoring_functions.py rename to llama_stack/core/ui/page/distribution/scoring_functions.py index 193146356..2a5196fa9 100644 --- a/llama_stack/distribution/ui/page/distribution/scoring_functions.py +++ b/llama_stack/core/ui/page/distribution/scoring_functions.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def scoring_functions(): diff --git a/llama_stack/distribution/ui/page/distribution/shields.py b/llama_stack/core/ui/page/distribution/shields.py similarity index 88% rename from llama_stack/distribution/ui/page/distribution/shields.py rename to llama_stack/core/ui/page/distribution/shields.py index 67d66d64f..ecce2f12b 100644 --- a/llama_stack/distribution/ui/page/distribution/shields.py +++ b/llama_stack/core/ui/page/distribution/shields.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def shields(): diff --git a/llama_stack/distribution/ui/page/distribution/vector_dbs.py b/llama_stack/core/ui/page/distribution/vector_dbs.py similarity index 90% rename from llama_stack/distribution/ui/page/distribution/vector_dbs.py rename to llama_stack/core/ui/page/distribution/vector_dbs.py index 49a4f25bb..e81077d2a 100644 --- a/llama_stack/distribution/ui/page/distribution/vector_dbs.py +++ b/llama_stack/core/ui/page/distribution/vector_dbs.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def vector_dbs(): diff --git a/llama_stack/distribution/ui/page/evaluations/__init__.py b/llama_stack/core/ui/page/evaluations/__init__.py similarity index 100% rename from llama_stack/distribution/ui/page/evaluations/__init__.py rename to llama_stack/core/ui/page/evaluations/__init__.py diff --git a/llama_stack/distribution/ui/page/evaluations/app_eval.py b/llama_stack/core/ui/page/evaluations/app_eval.py similarity index 97% rename from llama_stack/distribution/ui/page/evaluations/app_eval.py rename to llama_stack/core/ui/page/evaluations/app_eval.py index d7bc6388c..07e6349c9 100644 --- a/llama_stack/distribution/ui/page/evaluations/app_eval.py +++ b/llama_stack/core/ui/page/evaluations/app_eval.py @@ -9,8 +9,8 @@ import json import pandas as pd import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api -from llama_stack.distribution.ui.modules.utils import process_dataset +from llama_stack.core.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.utils import process_dataset def application_evaluation_page(): diff --git a/llama_stack/distribution/ui/page/evaluations/native_eval.py b/llama_stack/core/ui/page/evaluations/native_eval.py similarity index 99% rename from llama_stack/distribution/ui/page/evaluations/native_eval.py rename to llama_stack/core/ui/page/evaluations/native_eval.py index 97f875e17..2bef63b2f 100644 --- a/llama_stack/distribution/ui/page/evaluations/native_eval.py +++ b/llama_stack/core/ui/page/evaluations/native_eval.py @@ -9,7 +9,7 @@ import json import pandas as pd import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api def select_benchmark_1(): diff --git a/llama_stack/distribution/ui/page/playground/__init__.py b/llama_stack/core/ui/page/playground/__init__.py similarity index 100% rename from llama_stack/distribution/ui/page/playground/__init__.py rename to llama_stack/core/ui/page/playground/__init__.py diff --git a/llama_stack/distribution/ui/page/playground/chat.py b/llama_stack/core/ui/page/playground/chat.py similarity index 98% rename from llama_stack/distribution/ui/page/playground/chat.py rename to llama_stack/core/ui/page/playground/chat.py index fcaf08795..d391d0fb7 100644 --- a/llama_stack/distribution/ui/page/playground/chat.py +++ b/llama_stack/core/ui/page/playground/chat.py @@ -6,7 +6,7 @@ import streamlit as st -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api # Sidebar configurations with st.sidebar: diff --git a/llama_stack/distribution/ui/page/playground/rag.py b/llama_stack/core/ui/page/playground/rag.py similarity index 98% rename from llama_stack/distribution/ui/page/playground/rag.py rename to llama_stack/core/ui/page/playground/rag.py index 696d89bc2..2ffae1c33 100644 --- a/llama_stack/distribution/ui/page/playground/rag.py +++ b/llama_stack/core/ui/page/playground/rag.py @@ -10,8 +10,8 @@ import streamlit as st from llama_stack_client import Agent, AgentEventLogger, RAGDocument from llama_stack.apis.common.content_types import ToolCallDelta -from llama_stack.distribution.ui.modules.api import llama_stack_api -from llama_stack.distribution.ui.modules.utils import data_url_from_file +from llama_stack.core.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.utils import data_url_from_file def rag_chat_page(): diff --git a/llama_stack/distribution/ui/page/playground/tools.py b/llama_stack/core/ui/page/playground/tools.py similarity index 99% rename from llama_stack/distribution/ui/page/playground/tools.py rename to llama_stack/core/ui/page/playground/tools.py index 149d8cce9..602c9eea1 100644 --- a/llama_stack/distribution/ui/page/playground/tools.py +++ b/llama_stack/core/ui/page/playground/tools.py @@ -13,7 +13,7 @@ from llama_stack_client import Agent from llama_stack_client.lib.agents.react.agent import ReActAgent from llama_stack_client.lib.agents.react.tool_parser import ReActOutput -from llama_stack.distribution.ui.modules.api import llama_stack_api +from llama_stack.core.ui.modules.api import llama_stack_api class AgentType(enum.Enum): diff --git a/llama_stack/distribution/ui/requirements.txt b/llama_stack/core/ui/requirements.txt similarity index 100% rename from llama_stack/distribution/ui/requirements.txt rename to llama_stack/core/ui/requirements.txt diff --git a/llama_stack/distribution/utils/__init__.py b/llama_stack/core/utils/__init__.py similarity index 100% rename from llama_stack/distribution/utils/__init__.py rename to llama_stack/core/utils/__init__.py diff --git a/llama_stack/distribution/utils/config.py b/llama_stack/core/utils/config.py similarity index 100% rename from llama_stack/distribution/utils/config.py rename to llama_stack/core/utils/config.py diff --git a/llama_stack/distribution/utils/config_dirs.py b/llama_stack/core/utils/config_dirs.py similarity index 100% rename from llama_stack/distribution/utils/config_dirs.py rename to llama_stack/core/utils/config_dirs.py diff --git a/llama_stack/distribution/utils/config_resolution.py b/llama_stack/core/utils/config_resolution.py similarity index 96% rename from llama_stack/distribution/utils/config_resolution.py rename to llama_stack/core/utils/config_resolution.py index 7e8de1242..445050944 100644 --- a/llama_stack/distribution/utils/config_resolution.py +++ b/llama_stack/core/utils/config_resolution.py @@ -7,7 +7,7 @@ from enum import StrEnum from pathlib import Path -from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR +from llama_stack.core.utils.config_dirs import DISTRIBS_BASE_DIR from llama_stack.log import get_logger logger = get_logger(name=__name__, category="config_resolution") @@ -74,7 +74,7 @@ def _get_template_config_path(template_name: str, mode: Mode) -> Path: def _format_resolution_error(config_or_template: str, mode: Mode) -> str: """Format a helpful error message for resolution failures.""" - from llama_stack.distribution.utils.config_dirs import DISTRIBS_BASE_DIR + from llama_stack.core.utils.config_dirs import DISTRIBS_BASE_DIR template_path = _get_template_config_path(config_or_template, mode) distrib_path = DISTRIBS_BASE_DIR / f"llamastack-{config_or_template}" / f"{config_or_template}-{mode}.yaml" diff --git a/llama_stack/distribution/utils/context.py b/llama_stack/core/utils/context.py similarity index 100% rename from llama_stack/distribution/utils/context.py rename to llama_stack/core/utils/context.py diff --git a/llama_stack/distribution/utils/dynamic.py b/llama_stack/core/utils/dynamic.py similarity index 100% rename from llama_stack/distribution/utils/dynamic.py rename to llama_stack/core/utils/dynamic.py diff --git a/llama_stack/distribution/utils/exec.py b/llama_stack/core/utils/exec.py similarity index 96% rename from llama_stack/distribution/utils/exec.py rename to llama_stack/core/utils/exec.py index c646ae821..4272db3b8 100644 --- a/llama_stack/distribution/utils/exec.py +++ b/llama_stack/core/utils/exec.py @@ -18,7 +18,7 @@ import importlib import json from pathlib import Path -from llama_stack.distribution.utils.image_types import LlamaStackImageType +from llama_stack.core.utils.image_types import LlamaStackImageType def formulate_run_args(image_type: str, image_name: str) -> list[str]: @@ -79,7 +79,7 @@ def formulate_run_args(image_type: str, image_name: str) -> list[str]: return cprint(f"Using virtual environment: {env_name}", file=sys.stderr) - script = importlib.resources.files("llama_stack") / "distribution/start_stack.sh" + script = importlib.resources.files("llama_stack") / "core/start_stack.sh" run_args = [ script, image_type, diff --git a/llama_stack/distribution/utils/image_types.py b/llama_stack/core/utils/image_types.py similarity index 100% rename from llama_stack/distribution/utils/image_types.py rename to llama_stack/core/utils/image_types.py diff --git a/llama_stack/distribution/utils/model_utils.py b/llama_stack/core/utils/model_utils.py similarity index 100% rename from llama_stack/distribution/utils/model_utils.py rename to llama_stack/core/utils/model_utils.py diff --git a/llama_stack/distribution/utils/prompt_for_config.py b/llama_stack/core/utils/prompt_for_config.py similarity index 100% rename from llama_stack/distribution/utils/prompt_for_config.py rename to llama_stack/core/utils/prompt_for_config.py diff --git a/llama_stack/distribution/utils/serialize.py b/llama_stack/core/utils/serialize.py similarity index 100% rename from llama_stack/distribution/utils/serialize.py rename to llama_stack/core/utils/serialize.py diff --git a/llama_stack/log.py b/llama_stack/log.py index fb6fa85f9..ab53e08c0 100644 --- a/llama_stack/log.py +++ b/llama_stack/log.py @@ -15,7 +15,7 @@ from rich.errors import MarkupError from rich.logging import RichHandler from termcolor import cprint -from .distribution.datatypes import LoggingConfig +from llama_stack.core.datatypes import LoggingConfig # Default log level DEFAULT_LOG_LEVEL = logging.INFO diff --git a/llama_stack/providers/inline/agents/meta_reference/__init__.py b/llama_stack/providers/inline/agents/meta_reference/__init__.py index 4a77e65b9..334c32e15 100644 --- a/llama_stack/providers/inline/agents/meta_reference/__init__.py +++ b/llama_stack/providers/inline/agents/meta_reference/__init__.py @@ -6,7 +6,7 @@ from typing import Any -from llama_stack.distribution.datatypes import AccessRule, Api +from llama_stack.core.datatypes import AccessRule, Api from .config import MetaReferenceAgentsImplConfig diff --git a/llama_stack/providers/inline/agents/meta_reference/agent_instance.py b/llama_stack/providers/inline/agents/meta_reference/agent_instance.py index 3c34c71fb..fb22e10cc 100644 --- a/llama_stack/providers/inline/agents/meta_reference/agent_instance.py +++ b/llama_stack/providers/inline/agents/meta_reference/agent_instance.py @@ -61,7 +61,7 @@ from llama_stack.apis.inference import ( from llama_stack.apis.safety import Safety from llama_stack.apis.tools import ToolGroups, ToolInvocationResult, ToolRuntime from llama_stack.apis.vector_io import VectorIO -from llama_stack.distribution.datatypes import AccessRule +from llama_stack.core.datatypes import AccessRule from llama_stack.log import get_logger from llama_stack.models.llama.datatypes import ( BuiltinTool, diff --git a/llama_stack/providers/inline/agents/meta_reference/agents.py b/llama_stack/providers/inline/agents/meta_reference/agents.py index 4d0c429bd..b326a95b4 100644 --- a/llama_stack/providers/inline/agents/meta_reference/agents.py +++ b/llama_stack/providers/inline/agents/meta_reference/agents.py @@ -41,7 +41,7 @@ from llama_stack.apis.inference import ( from llama_stack.apis.safety import Safety from llama_stack.apis.tools import ToolGroups, ToolRuntime from llama_stack.apis.vector_io import VectorIO -from llama_stack.distribution.datatypes import AccessRule +from llama_stack.core.datatypes import AccessRule from llama_stack.providers.utils.kvstore import InmemoryKVStoreImpl, kvstore_impl from llama_stack.providers.utils.pagination import paginate_records from llama_stack.providers.utils.responses.responses_store import ResponsesStore diff --git a/llama_stack/providers/inline/agents/meta_reference/persistence.py b/llama_stack/providers/inline/agents/meta_reference/persistence.py index 437d617ad..8d4713cd2 100644 --- a/llama_stack/providers/inline/agents/meta_reference/persistence.py +++ b/llama_stack/providers/inline/agents/meta_reference/persistence.py @@ -10,10 +10,10 @@ import uuid from datetime import UTC, datetime from llama_stack.apis.agents import AgentConfig, Session, ToolExecutionStep, Turn -from llama_stack.distribution.access_control.access_control import AccessDeniedError, is_action_allowed -from llama_stack.distribution.access_control.datatypes import AccessRule -from llama_stack.distribution.datatypes import User -from llama_stack.distribution.request_headers import get_authenticated_user +from llama_stack.core.access_control.access_control import AccessDeniedError, is_action_allowed +from llama_stack.core.access_control.datatypes import AccessRule +from llama_stack.core.datatypes import User +from llama_stack.core.request_headers import get_authenticated_user from llama_stack.providers.utils.kvstore import KVStore log = logging.getLogger(__name__) diff --git a/llama_stack/providers/inline/eval/meta_reference/__init__.py b/llama_stack/providers/inline/eval/meta_reference/__init__.py index 7afe7f33b..cf2578a72 100644 --- a/llama_stack/providers/inline/eval/meta_reference/__init__.py +++ b/llama_stack/providers/inline/eval/meta_reference/__init__.py @@ -5,7 +5,7 @@ # the root directory of this source tree. from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import MetaReferenceEvalConfig diff --git a/llama_stack/providers/inline/files/localfs/__init__.py b/llama_stack/providers/inline/files/localfs/__init__.py index 71664efad..363b6f04c 100644 --- a/llama_stack/providers/inline/files/localfs/__init__.py +++ b/llama_stack/providers/inline/files/localfs/__init__.py @@ -6,7 +6,7 @@ from typing import Any -from llama_stack.distribution.datatypes import AccessRule, Api +from llama_stack.core.datatypes import AccessRule, Api from .config import LocalfsFilesImplConfig from .files import LocalfsFilesImpl diff --git a/llama_stack/providers/inline/files/localfs/files.py b/llama_stack/providers/inline/files/localfs/files.py index 433762c5a..1e9dca3b5 100644 --- a/llama_stack/providers/inline/files/localfs/files.py +++ b/llama_stack/providers/inline/files/localfs/files.py @@ -19,7 +19,7 @@ from llama_stack.apis.files import ( OpenAIFileObject, OpenAIFilePurpose, ) -from llama_stack.distribution.datatypes import AccessRule +from llama_stack.core.datatypes import AccessRule from llama_stack.providers.utils.sqlstore.api import ColumnDefinition, ColumnType from llama_stack.providers.utils.sqlstore.authorized_sqlstore import AuthorizedSqlStore from llama_stack.providers.utils.sqlstore.sqlstore import sqlstore_impl diff --git a/llama_stack/providers/inline/inference/meta_reference/common.py b/llama_stack/providers/inline/inference/meta_reference/common.py index beb0d39d4..1e164430d 100644 --- a/llama_stack/providers/inline/inference/meta_reference/common.py +++ b/llama_stack/providers/inline/inference/meta_reference/common.py @@ -6,7 +6,7 @@ from pathlib import Path -from llama_stack.distribution.utils.model_utils import model_local_dir +from llama_stack.core.utils.model_utils import model_local_dir def model_checkpoint_dir(model_id) -> str: diff --git a/llama_stack/providers/inline/post_training/huggingface/__init__.py b/llama_stack/providers/inline/post_training/huggingface/__init__.py index cc1a671c1..96c45cc4f 100644 --- a/llama_stack/providers/inline/post_training/huggingface/__init__.py +++ b/llama_stack/providers/inline/post_training/huggingface/__init__.py @@ -6,7 +6,7 @@ from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import HuggingFacePostTrainingConfig diff --git a/llama_stack/providers/inline/post_training/torchtune/__init__.py b/llama_stack/providers/inline/post_training/torchtune/__init__.py index 7a2f9eba2..af4ebd92a 100644 --- a/llama_stack/providers/inline/post_training/torchtune/__init__.py +++ b/llama_stack/providers/inline/post_training/torchtune/__init__.py @@ -6,7 +6,7 @@ from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import TorchtunePostTrainingConfig diff --git a/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py b/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py index fed19428c..49e1c95b8 100644 --- a/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py +++ b/llama_stack/providers/inline/post_training/torchtune/recipes/lora_finetuning_single_device.py @@ -43,8 +43,8 @@ from llama_stack.apis.post_training import ( QATFinetuningConfig, TrainingConfig, ) -from llama_stack.distribution.utils.config_dirs import DEFAULT_CHECKPOINT_DIR -from llama_stack.distribution.utils.model_utils import model_local_dir +from llama_stack.core.utils.config_dirs import DEFAULT_CHECKPOINT_DIR +from llama_stack.core.utils.model_utils import model_local_dir from llama_stack.models.llama.sku_list import resolve_model from llama_stack.providers.inline.post_training.common.utils import evacuate_model_from_device from llama_stack.providers.inline.post_training.torchtune.common import utils diff --git a/llama_stack/providers/inline/safety/llama_guard/llama_guard.py b/llama_stack/providers/inline/safety/llama_guard/llama_guard.py index 0d1c4ffe1..c580adfad 100644 --- a/llama_stack/providers/inline/safety/llama_guard/llama_guard.py +++ b/llama_stack/providers/inline/safety/llama_guard/llama_guard.py @@ -21,7 +21,7 @@ from llama_stack.apis.safety import ( ViolationLevel, ) from llama_stack.apis.shields import Shield -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from llama_stack.models.llama.datatypes import Role from llama_stack.models.llama.sku_types import CoreModelId from llama_stack.providers.datatypes import ShieldsProtocolPrivate diff --git a/llama_stack/providers/inline/safety/prompt_guard/prompt_guard.py b/llama_stack/providers/inline/safety/prompt_guard/prompt_guard.py index ff87889ea..ee645a41d 100644 --- a/llama_stack/providers/inline/safety/prompt_guard/prompt_guard.py +++ b/llama_stack/providers/inline/safety/prompt_guard/prompt_guard.py @@ -18,7 +18,7 @@ from llama_stack.apis.safety import ( ViolationLevel, ) from llama_stack.apis.shields import Shield -from llama_stack.distribution.utils.model_utils import model_local_dir +from llama_stack.core.utils.model_utils import model_local_dir from llama_stack.providers.datatypes import ShieldsProtocolPrivate from llama_stack.providers.utils.inference.prompt_adapter import ( interleaved_content_as_str, diff --git a/llama_stack/providers/inline/scoring/basic/__init__.py b/llama_stack/providers/inline/scoring/basic/__init__.py index d9d150b1a..c996b9c2d 100644 --- a/llama_stack/providers/inline/scoring/basic/__init__.py +++ b/llama_stack/providers/inline/scoring/basic/__init__.py @@ -5,7 +5,7 @@ # the root directory of this source tree. from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import BasicScoringConfig diff --git a/llama_stack/providers/inline/scoring/basic/scoring.py b/llama_stack/providers/inline/scoring/basic/scoring.py index 09f89be5e..91b10daae 100644 --- a/llama_stack/providers/inline/scoring/basic/scoring.py +++ b/llama_stack/providers/inline/scoring/basic/scoring.py @@ -14,7 +14,7 @@ from llama_stack.apis.scoring import ( ScoringResult, ) from llama_stack.apis.scoring_functions import ScoringFn, ScoringFnParams -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from llama_stack.providers.datatypes import ScoringFunctionsProtocolPrivate from llama_stack.providers.utils.common.data_schema_validator import ( get_valid_schemas, diff --git a/llama_stack/providers/inline/scoring/braintrust/__init__.py b/llama_stack/providers/inline/scoring/braintrust/__init__.py index 8ea6e9b96..3b492ae3f 100644 --- a/llama_stack/providers/inline/scoring/braintrust/__init__.py +++ b/llama_stack/providers/inline/scoring/braintrust/__init__.py @@ -7,7 +7,7 @@ from typing import Any from pydantic import BaseModel -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import BraintrustScoringConfig diff --git a/llama_stack/providers/inline/scoring/braintrust/braintrust.py b/llama_stack/providers/inline/scoring/braintrust/braintrust.py index d6655d657..14810f706 100644 --- a/llama_stack/providers/inline/scoring/braintrust/braintrust.py +++ b/llama_stack/providers/inline/scoring/braintrust/braintrust.py @@ -29,8 +29,8 @@ from llama_stack.apis.scoring import ( ScoringResultRow, ) from llama_stack.apis.scoring_functions import ScoringFn, ScoringFnParams -from llama_stack.distribution.datatypes import Api -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.datatypes import Api +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.providers.datatypes import ScoringFunctionsProtocolPrivate from llama_stack.providers.utils.common.data_schema_validator import ( get_valid_schemas, diff --git a/llama_stack/providers/inline/scoring/llm_as_judge/__init__.py b/llama_stack/providers/inline/scoring/llm_as_judge/__init__.py index 88bf10737..76735fcb3 100644 --- a/llama_stack/providers/inline/scoring/llm_as_judge/__init__.py +++ b/llama_stack/providers/inline/scoring/llm_as_judge/__init__.py @@ -5,7 +5,7 @@ # the root directory of this source tree. from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import LlmAsJudgeScoringConfig diff --git a/llama_stack/providers/inline/scoring/llm_as_judge/scoring.py b/llama_stack/providers/inline/scoring/llm_as_judge/scoring.py index 2bd113a94..fd651877c 100644 --- a/llama_stack/providers/inline/scoring/llm_as_judge/scoring.py +++ b/llama_stack/providers/inline/scoring/llm_as_judge/scoring.py @@ -15,7 +15,7 @@ from llama_stack.apis.scoring import ( ScoringResult, ) from llama_stack.apis.scoring_functions import ScoringFn, ScoringFnParams -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from llama_stack.providers.datatypes import ScoringFunctionsProtocolPrivate from llama_stack.providers.utils.common.data_schema_validator import ( get_valid_schemas, diff --git a/llama_stack/providers/inline/telemetry/meta_reference/__init__.py b/llama_stack/providers/inline/telemetry/meta_reference/__init__.py index 09e97136a..21743b653 100644 --- a/llama_stack/providers/inline/telemetry/meta_reference/__init__.py +++ b/llama_stack/providers/inline/telemetry/meta_reference/__init__.py @@ -6,7 +6,7 @@ from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import TelemetryConfig, TelemetrySink diff --git a/llama_stack/providers/inline/telemetry/meta_reference/config.py b/llama_stack/providers/inline/telemetry/meta_reference/config.py index f2a7c2a6e..31ae80050 100644 --- a/llama_stack/providers/inline/telemetry/meta_reference/config.py +++ b/llama_stack/providers/inline/telemetry/meta_reference/config.py @@ -9,7 +9,7 @@ from typing import Any from pydantic import BaseModel, Field, field_validator -from llama_stack.distribution.utils.config_dirs import RUNTIME_BASE_DIR +from llama_stack.core.utils.config_dirs import RUNTIME_BASE_DIR class TelemetrySink(StrEnum): diff --git a/llama_stack/providers/inline/telemetry/meta_reference/telemetry.py b/llama_stack/providers/inline/telemetry/meta_reference/telemetry.py index c63fc23c2..623267172 100644 --- a/llama_stack/providers/inline/telemetry/meta_reference/telemetry.py +++ b/llama_stack/providers/inline/telemetry/meta_reference/telemetry.py @@ -36,7 +36,7 @@ from llama_stack.apis.telemetry import ( Trace, UnstructuredLogEvent, ) -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from llama_stack.providers.inline.telemetry.meta_reference.console_span_processor import ( ConsoleSpanProcessor, ) diff --git a/llama_stack/providers/remote/datasetio/nvidia/README.md b/llama_stack/providers/remote/datasetio/nvidia/README.md index 8b1e2e6ee..40f2922f2 100644 --- a/llama_stack/providers/remote/datasetio/nvidia/README.md +++ b/llama_stack/providers/remote/datasetio/nvidia/README.md @@ -34,7 +34,7 @@ os.environ["NVIDIA_API_KEY"] = "your-api-key" os.environ["NVIDIA_CUSTOMIZER_URL"] = "http://nemo.test" os.environ["NVIDIA_DATASET_NAMESPACE"] = "default" os.environ["NVIDIA_PROJECT_ID"] = "test-project" -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient client = LlamaStackAsLibraryClient("nvidia") client.initialize() diff --git a/llama_stack/providers/remote/eval/nvidia/__init__.py b/llama_stack/providers/remote/eval/nvidia/__init__.py index 55e3754f3..1314fdb83 100644 --- a/llama_stack/providers/remote/eval/nvidia/__init__.py +++ b/llama_stack/providers/remote/eval/nvidia/__init__.py @@ -5,7 +5,7 @@ # the root directory of this source tree. from typing import Any -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api from .config import NVIDIAEvalConfig diff --git a/llama_stack/providers/remote/inference/fireworks/fireworks.py b/llama_stack/providers/remote/inference/fireworks/fireworks.py index c76aa39f3..ca4c7b578 100644 --- a/llama_stack/providers/remote/inference/fireworks/fireworks.py +++ b/llama_stack/providers/remote/inference/fireworks/fireworks.py @@ -39,7 +39,7 @@ from llama_stack.apis.inference import ( ToolDefinition, ToolPromptFormat, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.log import get_logger from llama_stack.providers.utils.inference.model_registry import ( ModelRegistryHelper, diff --git a/llama_stack/providers/remote/inference/nvidia/NVIDIA.md b/llama_stack/providers/remote/inference/nvidia/NVIDIA.md index a353c67f5..0d1da0365 100644 --- a/llama_stack/providers/remote/inference/nvidia/NVIDIA.md +++ b/llama_stack/providers/remote/inference/nvidia/NVIDIA.md @@ -33,7 +33,7 @@ os.environ["NVIDIA_API_KEY"] = ( ) os.environ["NVIDIA_BASE_URL"] = "http://nim.test" # NIM URL -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient client = LlamaStackAsLibraryClient("nvidia") client.initialize() diff --git a/llama_stack/providers/remote/inference/passthrough/passthrough.py b/llama_stack/providers/remote/inference/passthrough/passthrough.py index d5b3a5973..2f1cd40f2 100644 --- a/llama_stack/providers/remote/inference/passthrough/passthrough.py +++ b/llama_stack/providers/remote/inference/passthrough/passthrough.py @@ -34,7 +34,7 @@ from llama_stack.apis.inference import ( ToolPromptFormat, ) from llama_stack.apis.models import Model -from llama_stack.distribution.library_client import convert_pydantic_to_json_value, convert_to_pydantic +from llama_stack.core.library_client import convert_pydantic_to_json_value, convert_to_pydantic from llama_stack.providers.utils.inference.model_registry import ModelRegistryHelper from llama_stack.providers.utils.inference.openai_compat import prepare_openai_completion_params diff --git a/llama_stack/providers/remote/inference/together/together.py b/llama_stack/providers/remote/inference/together/together.py index 46094c146..a06e4173b 100644 --- a/llama_stack/providers/remote/inference/together/together.py +++ b/llama_stack/providers/remote/inference/together/together.py @@ -38,7 +38,7 @@ from llama_stack.apis.inference import ( ToolDefinition, ToolPromptFormat, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.log import get_logger from llama_stack.providers.utils.inference.model_registry import ModelRegistryHelper from llama_stack.providers.utils.inference.openai_compat import ( diff --git a/llama_stack/providers/remote/post_training/nvidia/README.md b/llama_stack/providers/remote/post_training/nvidia/README.md index 3ef538d29..1cce78c89 100644 --- a/llama_stack/providers/remote/post_training/nvidia/README.md +++ b/llama_stack/providers/remote/post_training/nvidia/README.md @@ -40,7 +40,7 @@ os.environ["NVIDIA_DATASET_NAMESPACE"] = "default" os.environ["NVIDIA_PROJECT_ID"] = "test-project" os.environ["NVIDIA_OUTPUT_MODEL_DIR"] = "test-example-model@v1" -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient client = LlamaStackAsLibraryClient("nvidia") client.initialize() diff --git a/llama_stack/providers/remote/safety/nvidia/README.md b/llama_stack/providers/remote/safety/nvidia/README.md index 434db32fb..218963e47 100644 --- a/llama_stack/providers/remote/safety/nvidia/README.md +++ b/llama_stack/providers/remote/safety/nvidia/README.md @@ -32,7 +32,7 @@ import os os.environ["NVIDIA_API_KEY"] = "your-api-key" os.environ["NVIDIA_GUARDRAILS_URL"] = "http://guardrails.test" -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient client = LlamaStackAsLibraryClient("nvidia") client.initialize() diff --git a/llama_stack/providers/remote/safety/sambanova/sambanova.py b/llama_stack/providers/remote/safety/sambanova/sambanova.py index 1a65f6aa1..3e0d03956 100644 --- a/llama_stack/providers/remote/safety/sambanova/sambanova.py +++ b/llama_stack/providers/remote/safety/sambanova/sambanova.py @@ -19,7 +19,7 @@ from llama_stack.apis.safety import ( ViolationLevel, ) from llama_stack.apis.shields import Shield -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.providers.datatypes import ShieldsProtocolPrivate from llama_stack.providers.utils.inference.openai_compat import convert_message_to_openai_dict_new diff --git a/llama_stack/providers/remote/tool_runtime/bing_search/bing_search.py b/llama_stack/providers/remote/tool_runtime/bing_search/bing_search.py index 7e82cb6d4..e40903969 100644 --- a/llama_stack/providers/remote/tool_runtime/bing_search/bing_search.py +++ b/llama_stack/providers/remote/tool_runtime/bing_search/bing_search.py @@ -18,7 +18,7 @@ from llama_stack.apis.tools import ( ToolParameter, ToolRuntime, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.providers.datatypes import ToolGroupsProtocolPrivate from .config import BingSearchToolConfig diff --git a/llama_stack/providers/remote/tool_runtime/brave_search/brave_search.py b/llama_stack/providers/remote/tool_runtime/brave_search/brave_search.py index b96b9e59c..ba3b910d5 100644 --- a/llama_stack/providers/remote/tool_runtime/brave_search/brave_search.py +++ b/llama_stack/providers/remote/tool_runtime/brave_search/brave_search.py @@ -17,7 +17,7 @@ from llama_stack.apis.tools import ( ToolParameter, ToolRuntime, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.models.llama.datatypes import BuiltinTool from llama_stack.providers.datatypes import ToolGroupsProtocolPrivate diff --git a/llama_stack/providers/remote/tool_runtime/model_context_protocol/model_context_protocol.py b/llama_stack/providers/remote/tool_runtime/model_context_protocol/model_context_protocol.py index a9b252dfe..578bb6d34 100644 --- a/llama_stack/providers/remote/tool_runtime/model_context_protocol/model_context_protocol.py +++ b/llama_stack/providers/remote/tool_runtime/model_context_protocol/model_context_protocol.py @@ -15,7 +15,7 @@ from llama_stack.apis.tools import ( ToolInvocationResult, ToolRuntime, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.log import get_logger from llama_stack.providers.datatypes import ToolGroupsProtocolPrivate from llama_stack.providers.utils.tools.mcp import invoke_mcp_tool, list_mcp_tools diff --git a/llama_stack/providers/remote/tool_runtime/tavily_search/tavily_search.py b/llama_stack/providers/remote/tool_runtime/tavily_search/tavily_search.py index 1fe91fd7f..976ec9c57 100644 --- a/llama_stack/providers/remote/tool_runtime/tavily_search/tavily_search.py +++ b/llama_stack/providers/remote/tool_runtime/tavily_search/tavily_search.py @@ -18,7 +18,7 @@ from llama_stack.apis.tools import ( ToolParameter, ToolRuntime, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.providers.datatypes import ToolGroupsProtocolPrivate from .config import TavilySearchToolConfig diff --git a/llama_stack/providers/remote/tool_runtime/wolfram_alpha/wolfram_alpha.py b/llama_stack/providers/remote/tool_runtime/wolfram_alpha/wolfram_alpha.py index 6e1d0f61d..f12a44958 100644 --- a/llama_stack/providers/remote/tool_runtime/wolfram_alpha/wolfram_alpha.py +++ b/llama_stack/providers/remote/tool_runtime/wolfram_alpha/wolfram_alpha.py @@ -18,7 +18,7 @@ from llama_stack.apis.tools import ( ToolParameter, ToolRuntime, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.providers.datatypes import ToolGroupsProtocolPrivate from .config import WolframAlphaToolConfig diff --git a/llama_stack/providers/remote/vector_io/weaviate/weaviate.py b/llama_stack/providers/remote/vector_io/weaviate/weaviate.py index 7ae2035db..228e5fa62 100644 --- a/llama_stack/providers/remote/vector_io/weaviate/weaviate.py +++ b/llama_stack/providers/remote/vector_io/weaviate/weaviate.py @@ -18,7 +18,7 @@ from llama_stack.apis.common.errors import VectorStoreNotFoundError from llama_stack.apis.files.files import Files from llama_stack.apis.vector_dbs import VectorDB from llama_stack.apis.vector_io import Chunk, QueryChunksResponse, VectorIO -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.providers.datatypes import Api, VectorDBsProtocolPrivate from llama_stack.providers.utils.kvstore import kvstore_impl from llama_stack.providers.utils.kvstore.api import KVStore diff --git a/llama_stack/providers/utils/common/data_schema_validator.py b/llama_stack/providers/utils/common/data_schema_validator.py index 28a243863..b0305104f 100644 --- a/llama_stack/providers/utils/common/data_schema_validator.py +++ b/llama_stack/providers/utils/common/data_schema_validator.py @@ -12,7 +12,7 @@ from llama_stack.apis.common.type_system import ( CompletionInputType, StringType, ) -from llama_stack.distribution.datatypes import Api +from llama_stack.core.datatypes import Api class ColumnName(Enum): diff --git a/llama_stack/providers/utils/inference/inference_store.py b/llama_stack/providers/utils/inference/inference_store.py index 60a87494e..43006cfd5 100644 --- a/llama_stack/providers/utils/inference/inference_store.py +++ b/llama_stack/providers/utils/inference/inference_store.py @@ -10,8 +10,8 @@ from llama_stack.apis.inference import ( OpenAIMessageParam, Order, ) -from llama_stack.distribution.datatypes import AccessRule -from llama_stack.distribution.utils.config_dirs import RUNTIME_BASE_DIR +from llama_stack.core.datatypes import AccessRule +from llama_stack.core.utils.config_dirs import RUNTIME_BASE_DIR from ..sqlstore.api import ColumnDefinition, ColumnType from ..sqlstore.authorized_sqlstore import AuthorizedSqlStore diff --git a/llama_stack/providers/utils/inference/litellm_openai_mixin.py b/llama_stack/providers/utils/inference/litellm_openai_mixin.py index 6ccf2a729..abba62f87 100644 --- a/llama_stack/providers/utils/inference/litellm_openai_mixin.py +++ b/llama_stack/providers/utils/inference/litellm_openai_mixin.py @@ -38,7 +38,7 @@ from llama_stack.apis.inference import ( ToolDefinition, ToolPromptFormat, ) -from llama_stack.distribution.request_headers import NeedsRequestProviderData +from llama_stack.core.request_headers import NeedsRequestProviderData from llama_stack.log import get_logger from llama_stack.providers.utils.inference.model_registry import ModelRegistryHelper from llama_stack.providers.utils.inference.openai_compat import ( diff --git a/llama_stack/providers/utils/kvstore/config.py b/llama_stack/providers/utils/kvstore/config.py index 0219bbebe..f00cb1f8b 100644 --- a/llama_stack/providers/utils/kvstore/config.py +++ b/llama_stack/providers/utils/kvstore/config.py @@ -10,7 +10,7 @@ from typing import Annotated, Literal from pydantic import BaseModel, Field, field_validator -from llama_stack.distribution.utils.config_dirs import RUNTIME_BASE_DIR +from llama_stack.core.utils.config_dirs import RUNTIME_BASE_DIR class KVStoreType(Enum): diff --git a/llama_stack/providers/utils/responses/responses_store.py b/llama_stack/providers/utils/responses/responses_store.py index ea6db7991..04778ed1c 100644 --- a/llama_stack/providers/utils/responses/responses_store.py +++ b/llama_stack/providers/utils/responses/responses_store.py @@ -14,8 +14,8 @@ from llama_stack.apis.agents.openai_responses import ( OpenAIResponseObject, OpenAIResponseObjectWithInput, ) -from llama_stack.distribution.datatypes import AccessRule -from llama_stack.distribution.utils.config_dirs import RUNTIME_BASE_DIR +from llama_stack.core.datatypes import AccessRule +from llama_stack.core.utils.config_dirs import RUNTIME_BASE_DIR from ..sqlstore.api import ColumnDefinition, ColumnType from ..sqlstore.authorized_sqlstore import AuthorizedSqlStore diff --git a/llama_stack/providers/utils/sqlstore/authorized_sqlstore.py b/llama_stack/providers/utils/sqlstore/authorized_sqlstore.py index 864a7dbb6..ccc835768 100644 --- a/llama_stack/providers/utils/sqlstore/authorized_sqlstore.py +++ b/llama_stack/providers/utils/sqlstore/authorized_sqlstore.py @@ -7,11 +7,11 @@ from collections.abc import Mapping from typing import Any, Literal -from llama_stack.distribution.access_control.access_control import default_policy, is_action_allowed -from llama_stack.distribution.access_control.conditions import ProtectedResource -from llama_stack.distribution.access_control.datatypes import AccessRule, Action, Scope -from llama_stack.distribution.datatypes import User -from llama_stack.distribution.request_headers import get_authenticated_user +from llama_stack.core.access_control.access_control import default_policy, is_action_allowed +from llama_stack.core.access_control.conditions import ProtectedResource +from llama_stack.core.access_control.datatypes import AccessRule, Action, Scope +from llama_stack.core.datatypes import User +from llama_stack.core.request_headers import get_authenticated_user from llama_stack.log import get_logger from .api import ColumnDefinition, ColumnType, PaginatedResponse, SqlStore diff --git a/llama_stack/providers/utils/sqlstore/sqlstore.py b/llama_stack/providers/utils/sqlstore/sqlstore.py index 9f7eefcf5..fc44402ae 100644 --- a/llama_stack/providers/utils/sqlstore/sqlstore.py +++ b/llama_stack/providers/utils/sqlstore/sqlstore.py @@ -11,7 +11,7 @@ from typing import Annotated, Literal from pydantic import BaseModel, Field -from llama_stack.distribution.utils.config_dirs import RUNTIME_BASE_DIR +from llama_stack.core.utils.config_dirs import RUNTIME_BASE_DIR from .api import SqlStore diff --git a/llama_stack/providers/utils/tools/mcp.py b/llama_stack/providers/utils/tools/mcp.py index 76593a4b8..02f7aaf8a 100644 --- a/llama_stack/providers/utils/tools/mcp.py +++ b/llama_stack/providers/utils/tools/mcp.py @@ -22,7 +22,7 @@ from llama_stack.apis.tools import ( ToolInvocationResult, ToolParameter, ) -from llama_stack.distribution.datatypes import AuthenticationRequiredError +from llama_stack.core.datatypes import AuthenticationRequiredError from llama_stack.log import get_logger from llama_stack.providers.utils.tools.ttl_dict import TTLDict diff --git a/llama_stack/templates/dell/dell.py b/llama_stack/templates/dell/dell.py index 64e01535c..743b4dc86 100644 --- a/llama_stack/templates/dell/dell.py +++ b/llama_stack/templates/dell/dell.py @@ -5,7 +5,7 @@ # the root directory of this source tree. from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( BuildProvider, ModelInput, Provider, diff --git a/llama_stack/templates/meta-reference-gpu/meta_reference.py b/llama_stack/templates/meta-reference-gpu/meta_reference.py index 981c66bf5..24403a8ed 100644 --- a/llama_stack/templates/meta-reference-gpu/meta_reference.py +++ b/llama_stack/templates/meta-reference-gpu/meta_reference.py @@ -7,7 +7,7 @@ from pathlib import Path from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( BuildProvider, ModelInput, Provider, diff --git a/llama_stack/templates/nvidia/nvidia.py b/llama_stack/templates/nvidia/nvidia.py index df82cf7c0..ccf28fce5 100644 --- a/llama_stack/templates/nvidia/nvidia.py +++ b/llama_stack/templates/nvidia/nvidia.py @@ -6,7 +6,7 @@ from pathlib import Path -from llama_stack.distribution.datatypes import BuildProvider, ModelInput, Provider, ShieldInput, ToolGroupInput +from llama_stack.core.datatypes import BuildProvider, ModelInput, Provider, ShieldInput, ToolGroupInput from llama_stack.providers.remote.datasetio.nvidia import NvidiaDatasetIOConfig from llama_stack.providers.remote.eval.nvidia import NVIDIAEvalConfig from llama_stack.providers.remote.inference.nvidia import NVIDIAConfig diff --git a/llama_stack/templates/open-benchmark/open_benchmark.py b/llama_stack/templates/open-benchmark/open_benchmark.py index 0a0d9fb14..724ca0d4a 100644 --- a/llama_stack/templates/open-benchmark/open_benchmark.py +++ b/llama_stack/templates/open-benchmark/open_benchmark.py @@ -7,7 +7,7 @@ from llama_stack.apis.datasets import DatasetPurpose, URIDataSource from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( BenchmarkInput, BuildProvider, DatasetInput, diff --git a/llama_stack/templates/postgres-demo/postgres_demo.py b/llama_stack/templates/postgres-demo/postgres_demo.py index d9ded9a86..65345648a 100644 --- a/llama_stack/templates/postgres-demo/postgres_demo.py +++ b/llama_stack/templates/postgres-demo/postgres_demo.py @@ -6,7 +6,7 @@ from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( BuildProvider, ModelInput, Provider, diff --git a/llama_stack/templates/starter/starter.py b/llama_stack/templates/starter/starter.py index 3ea3c8f5e..54e200311 100644 --- a/llama_stack/templates/starter/starter.py +++ b/llama_stack/templates/starter/starter.py @@ -7,14 +7,14 @@ from typing import Any -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( BuildProvider, Provider, ProviderSpec, ShieldInput, ToolGroupInput, ) -from llama_stack.distribution.utils.dynamic import instantiate_class_type +from llama_stack.core.utils.dynamic import instantiate_class_type from llama_stack.providers.datatypes import RemoteProviderSpec from llama_stack.providers.inline.files.localfs.config import LocalfsFilesImplConfig from llama_stack.providers.inline.inference.sentence_transformers import ( diff --git a/llama_stack/templates/template.py b/llama_stack/templates/template.py index 084996cd4..4e5e6051b 100644 --- a/llama_stack/templates/template.py +++ b/llama_stack/templates/template.py @@ -14,7 +14,7 @@ from pydantic import BaseModel, Field from llama_stack.apis.datasets import DatasetPurpose from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( LLAMA_STACK_RUN_CONFIG_VERSION, Api, BenchmarkInput, @@ -27,8 +27,8 @@ from llama_stack.distribution.datatypes import ( ShieldInput, ToolGroupInput, ) -from llama_stack.distribution.distribution import get_provider_registry -from llama_stack.distribution.utils.dynamic import instantiate_class_type +from llama_stack.core.distribution import get_provider_registry +from llama_stack.core.utils.dynamic import instantiate_class_type from llama_stack.providers.utils.inference.model_registry import ProviderModelEntry from llama_stack.providers.utils.kvstore.config import SqliteKVStoreConfig from llama_stack.providers.utils.kvstore.config import get_pip_packages as get_kv_pip_packages diff --git a/llama_stack/templates/watsonx/watsonx.py b/llama_stack/templates/watsonx/watsonx.py index 5d8332c4f..570389e03 100644 --- a/llama_stack/templates/watsonx/watsonx.py +++ b/llama_stack/templates/watsonx/watsonx.py @@ -7,7 +7,7 @@ from pathlib import Path from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import BuildProvider, ModelInput, Provider, ToolGroupInput +from llama_stack.core.datatypes import BuildProvider, ModelInput, Provider, ToolGroupInput from llama_stack.providers.inline.inference.sentence_transformers import ( SentenceTransformersInferenceConfig, ) diff --git a/pyproject.toml b/pyproject.toml index be003bf92..f666000a2 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -235,17 +235,17 @@ follow_imports = "silent" exclude = [ # As we fix more and more of these, we should remove them from the list "^llama_stack/cli/download\\.py$", - "^llama_stack/distribution/build\\.py$", - "^llama_stack/distribution/client\\.py$", - "^llama_stack/distribution/request_headers\\.py$", - "^llama_stack/distribution/routers/", - "^llama_stack/distribution/routing_tables/", - "^llama_stack/distribution/server/endpoints\\.py$", - "^llama_stack/distribution/server/server\\.py$", - "^llama_stack/distribution/stack\\.py$", - "^llama_stack/distribution/store/registry\\.py$", - "^llama_stack/distribution/utils/exec\\.py$", - "^llama_stack/distribution/utils/prompt_for_config\\.py$", + "^llama_stack.core/build\\.py$", + "^llama_stack.core/client\\.py$", + "^llama_stack.core/request_headers\\.py$", + "^llama_stack.core/routers/", + "^llama_stack.core/routing_tables/", + "^llama_stack.core/server/endpoints\\.py$", + "^llama_stack.core/server/server\\.py$", + "^llama_stack.core/stack\\.py$", + "^llama_stack.core/store/registry\\.py$", + "^llama_stack.core/utils/exec\\.py$", + "^llama_stack.core/utils/prompt_for_config\\.py$", "^llama_stack/models/llama/llama3/interface\\.py$", "^llama_stack/models/llama/llama3/tokenizer\\.py$", "^llama_stack/models/llama/llama3/tool_utils\\.py$", diff --git a/scripts/provider_codegen.py b/scripts/provider_codegen.py index 6e316c539..80c5b7840 100755 --- a/scripts/provider_codegen.py +++ b/scripts/provider_codegen.py @@ -12,7 +12,7 @@ from typing import Any from rich.progress import Progress, SpinnerColumn, TextColumn -from llama_stack.distribution.distribution import get_provider_registry +from llama_stack.core.distribution import get_provider_registry REPO_ROOT = Path(__file__).parent.parent diff --git a/tests/integration/agents/test_openai_responses.py b/tests/integration/agents/test_openai_responses.py index 7ae48913b..784ab6893 100644 --- a/tests/integration/agents/test_openai_responses.py +++ b/tests/integration/agents/test_openai_responses.py @@ -6,7 +6,7 @@ import pytest from openai import BadRequestError, OpenAI -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient @pytest.fixture diff --git a/tests/integration/files/test_files.py b/tests/integration/files/test_files.py index 118a751f0..b17c7db83 100644 --- a/tests/integration/files/test_files.py +++ b/tests/integration/files/test_files.py @@ -10,8 +10,8 @@ from unittest.mock import patch import pytest from openai import OpenAI -from llama_stack.distribution.datatypes import User -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.datatypes import User +from llama_stack.core.library_client import LlamaStackAsLibraryClient def test_openai_client_basic_operations(compat_client, client_with_models): diff --git a/tests/integration/fixtures/common.py b/tests/integration/fixtures/common.py index f6b5b3026..bdbe0d66f 100644 --- a/tests/integration/fixtures/common.py +++ b/tests/integration/fixtures/common.py @@ -20,7 +20,7 @@ from llama_stack_client import LlamaStackClient from openai import OpenAI from llama_stack import LlamaStackAsLibraryClient -from llama_stack.distribution.stack import run_config_from_adhoc_config_spec +from llama_stack.core.stack import run_config_from_adhoc_config_spec from llama_stack.env import get_env_or_fail DEFAULT_PORT = 8321 diff --git a/tests/integration/inference/test_openai_completion.py b/tests/integration/inference/test_openai_completion.py index 51a208b17..14254baa7 100644 --- a/tests/integration/inference/test_openai_completion.py +++ b/tests/integration/inference/test_openai_completion.py @@ -14,7 +14,7 @@ from openai import OpenAI from reportlab.lib.pagesizes import letter from reportlab.pdfgen import canvas -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient from ..test_cases.test_case import TestCase diff --git a/tests/integration/inference/test_openai_embeddings.py b/tests/integration/inference/test_openai_embeddings.py index 1b8bd9038..2c545cc43 100644 --- a/tests/integration/inference/test_openai_embeddings.py +++ b/tests/integration/inference/test_openai_embeddings.py @@ -10,7 +10,7 @@ import struct import pytest from openai import OpenAI -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient def decode_base64_to_floats(base64_string: str) -> list[float]: diff --git a/tests/integration/providers/utils/sqlstore/test_authorized_sqlstore.py b/tests/integration/providers/utils/sqlstore/test_authorized_sqlstore.py index c32d6cd17..4002f2e1f 100644 --- a/tests/integration/providers/utils/sqlstore/test_authorized_sqlstore.py +++ b/tests/integration/providers/utils/sqlstore/test_authorized_sqlstore.py @@ -10,8 +10,8 @@ from unittest.mock import patch import pytest -from llama_stack.distribution.access_control.access_control import default_policy -from llama_stack.distribution.datatypes import User +from llama_stack.core.access_control.access_control import default_policy +from llama_stack.core.datatypes import User from llama_stack.providers.utils.sqlstore.api import ColumnType from llama_stack.providers.utils.sqlstore.authorized_sqlstore import AuthorizedSqlStore from llama_stack.providers.utils.sqlstore.sqlstore import PostgresSqlStoreConfig, SqliteSqlStoreConfig, sqlstore_impl @@ -186,7 +186,7 @@ async def test_authorized_store_attributes(mock_get_authenticated_user, authoriz @patch("llama_stack.providers.utils.sqlstore.authorized_sqlstore.get_authenticated_user") async def test_user_ownership_policy(mock_get_authenticated_user, authorized_store, request): """Test that 'user is owner' policies work correctly with record ownership""" - from llama_stack.distribution.access_control.datatypes import AccessRule, Action, Scope + from llama_stack.core.access_control.datatypes import AccessRule, Action, Scope backend_name = request.node.callspec.id diff --git a/tests/integration/tool_runtime/test_mcp.py b/tests/integration/tool_runtime/test_mcp.py index f208dcbea..91ed26684 100644 --- a/tests/integration/tool_runtime/test_mcp.py +++ b/tests/integration/tool_runtime/test_mcp.py @@ -10,7 +10,7 @@ import pytest from llama_stack_client import Agent from llama_stack import LlamaStackAsLibraryClient -from llama_stack.distribution.datatypes import AuthenticationRequiredError +from llama_stack.core.datatypes import AuthenticationRequiredError AUTH_TOKEN = "test-token" diff --git a/tests/integration/vector_io/test_openai_vector_stores.py b/tests/integration/vector_io/test_openai_vector_stores.py index a34c5b410..399f1b1cd 100644 --- a/tests/integration/vector_io/test_openai_vector_stores.py +++ b/tests/integration/vector_io/test_openai_vector_stores.py @@ -14,7 +14,7 @@ from openai import BadRequestError as OpenAIBadRequestError from openai import OpenAI from llama_stack.apis.vector_io import Chunk -from llama_stack.distribution.library_client import LlamaStackAsLibraryClient +from llama_stack.core.library_client import LlamaStackAsLibraryClient logger = logging.getLogger(__name__) diff --git a/tests/unit/cli/test_stack_config.py b/tests/unit/cli/test_stack_config.py index a41049006..daaf229e5 100644 --- a/tests/unit/cli/test_stack_config.py +++ b/tests/unit/cli/test_stack_config.py @@ -9,7 +9,7 @@ from datetime import datetime import pytest import yaml -from llama_stack.distribution.configure import ( +from llama_stack.core.configure import ( LLAMA_STACK_RUN_CONFIG_VERSION, parse_and_maybe_upgrade_config, ) diff --git a/tests/unit/distribution/routers/test_routing_tables.py b/tests/unit/distribution/routers/test_routing_tables.py index 308b5c28f..155ad0142 100644 --- a/tests/unit/distribution/routers/test_routing_tables.py +++ b/tests/unit/distribution/routers/test_routing_tables.py @@ -15,14 +15,14 @@ from llama_stack.apis.models import Model, ModelType from llama_stack.apis.shields.shields import Shield from llama_stack.apis.tools import ListToolDefsResponse, ToolDef, ToolGroup, ToolParameter from llama_stack.apis.vector_dbs import VectorDB -from llama_stack.distribution.datatypes import RegistryEntrySource -from llama_stack.distribution.routing_tables.benchmarks import BenchmarksRoutingTable -from llama_stack.distribution.routing_tables.datasets import DatasetsRoutingTable -from llama_stack.distribution.routing_tables.models import ModelsRoutingTable -from llama_stack.distribution.routing_tables.scoring_functions import ScoringFunctionsRoutingTable -from llama_stack.distribution.routing_tables.shields import ShieldsRoutingTable -from llama_stack.distribution.routing_tables.toolgroups import ToolGroupsRoutingTable -from llama_stack.distribution.routing_tables.vector_dbs import VectorDBsRoutingTable +from llama_stack.core.datatypes import RegistryEntrySource +from llama_stack.core.routing_tables.benchmarks import BenchmarksRoutingTable +from llama_stack.core.routing_tables.datasets import DatasetsRoutingTable +from llama_stack.core.routing_tables.models import ModelsRoutingTable +from llama_stack.core.routing_tables.scoring_functions import ScoringFunctionsRoutingTable +from llama_stack.core.routing_tables.shields import ShieldsRoutingTable +from llama_stack.core.routing_tables.toolgroups import ToolGroupsRoutingTable +from llama_stack.core.routing_tables.vector_dbs import VectorDBsRoutingTable class Impl: diff --git a/tests/unit/distribution/routing_tables/test_vector_dbs.py b/tests/unit/distribution/routing_tables/test_vector_dbs.py index 28887e1cf..789eda433 100644 --- a/tests/unit/distribution/routing_tables/test_vector_dbs.py +++ b/tests/unit/distribution/routing_tables/test_vector_dbs.py @@ -24,10 +24,10 @@ from llama_stack.apis.vector_io.vector_io import ( VectorStoreObject, VectorStoreSearchResponsePage, ) -from llama_stack.distribution.access_control.datatypes import AccessRule, Scope -from llama_stack.distribution.datatypes import User -from llama_stack.distribution.request_headers import request_provider_data_context -from llama_stack.distribution.routing_tables.vector_dbs import VectorDBsRoutingTable +from llama_stack.core.access_control.datatypes import AccessRule, Scope +from llama_stack.core.datatypes import User +from llama_stack.core.request_headers import request_provider_data_context +from llama_stack.core.routing_tables.vector_dbs import VectorDBsRoutingTable from tests.unit.distribution.routers.test_routing_tables import Impl, InferenceImpl, ModelsRoutingTable diff --git a/tests/unit/distribution/test_build_path.py b/tests/unit/distribution/test_build_path.py index 555cdda4a..f71dd77ec 100644 --- a/tests/unit/distribution/test_build_path.py +++ b/tests/unit/distribution/test_build_path.py @@ -9,8 +9,8 @@ from pathlib import Path from llama_stack.cli.stack._build import ( _run_stack_build_command_from_build_config, ) -from llama_stack.distribution.datatypes import BuildConfig, DistributionSpec -from llama_stack.distribution.utils.image_types import LlamaStackImageType +from llama_stack.core.datatypes import BuildConfig, DistributionSpec +from llama_stack.core.utils.image_types import LlamaStackImageType def test_container_build_passes_path(monkeypatch, tmp_path): diff --git a/tests/unit/distribution/test_context.py b/tests/unit/distribution/test_context.py index 7914be51d..43dd5a6e8 100644 --- a/tests/unit/distribution/test_context.py +++ b/tests/unit/distribution/test_context.py @@ -10,7 +10,7 @@ from contextvars import ContextVar import pytest -from llama_stack.distribution.utils.context import preserve_contexts_async_generator +from llama_stack.core.utils.context import preserve_contexts_async_generator async def test_preserve_contexts_with_exception(): diff --git a/tests/unit/distribution/test_distribution.py b/tests/unit/distribution/test_distribution.py index 04d5cde67..c72106e46 100644 --- a/tests/unit/distribution/test_distribution.py +++ b/tests/unit/distribution/test_distribution.py @@ -11,8 +11,8 @@ import pytest import yaml from pydantic import BaseModel, Field, ValidationError -from llama_stack.distribution.datatypes import Api, Provider, StackRunConfig -from llama_stack.distribution.distribution import get_provider_registry +from llama_stack.core.datatypes import Api, Provider, StackRunConfig +from llama_stack.core.distribution import get_provider_registry from llama_stack.providers.datatypes import ProviderSpec @@ -260,7 +260,7 @@ pip_packages: """Test loading an external provider from a module (success path).""" from types import SimpleNamespace - from llama_stack.distribution.datatypes import Provider, StackRunConfig + from llama_stack.core.datatypes import Provider, StackRunConfig from llama_stack.providers.datatypes import Api, ProviderSpec # Simulate a provider module with get_provider_spec @@ -299,7 +299,7 @@ pip_packages: def test_external_provider_from_module_not_found(self, mock_providers): """Test handling ModuleNotFoundError for missing provider module.""" - from llama_stack.distribution.datatypes import Provider, StackRunConfig + from llama_stack.core.datatypes import Provider, StackRunConfig import_module_side_effect = make_import_module_side_effect(raise_for_external=True) @@ -323,7 +323,7 @@ pip_packages: def test_external_provider_from_module_missing_get_provider_spec(self, mock_providers): """Test handling missing get_provider_spec in provider module (should raise ValueError).""" - from llama_stack.distribution.datatypes import Provider, StackRunConfig + from llama_stack.core.datatypes import Provider, StackRunConfig import_module_side_effect = make_import_module_side_effect(missing_get_provider_spec=True) @@ -346,7 +346,7 @@ pip_packages: def test_external_provider_from_module_building(self, mock_providers): """Test loading an external provider from a module during build (building=True, partial spec).""" - from llama_stack.distribution.datatypes import BuildConfig, BuildProvider, DistributionSpec + from llama_stack.core.datatypes import BuildConfig, BuildProvider, DistributionSpec from llama_stack.providers.datatypes import Api # No importlib patch needed, should not import module when type of `config` is BuildConfig or DistributionSpec diff --git a/tests/unit/distribution/test_library_client_initialization.py b/tests/unit/distribution/test_library_client_initialization.py index 2c394fc0e..e510d513d 100644 --- a/tests/unit/distribution/test_library_client_initialization.py +++ b/tests/unit/distribution/test_library_client_initialization.py @@ -13,7 +13,7 @@ initialize() on the library client, preventing AttributeError regressions. import pytest -from llama_stack.distribution.library_client import ( +from llama_stack.core.library_client import ( AsyncLlamaStackAsLibraryClient, LlamaStackAsLibraryClient, ) diff --git a/tests/unit/files/test_files.py b/tests/unit/files/test_files.py index c3ec25116..04f33e97d 100644 --- a/tests/unit/files/test_files.py +++ b/tests/unit/files/test_files.py @@ -9,7 +9,7 @@ import pytest from llama_stack.apis.common.responses import Order from llama_stack.apis.files import OpenAIFilePurpose -from llama_stack.distribution.access_control.access_control import default_policy +from llama_stack.core.access_control.access_control import default_policy from llama_stack.providers.inline.files.localfs import ( LocalfsFilesImpl, LocalfsFilesImplConfig, diff --git a/tests/unit/fixtures.py b/tests/unit/fixtures.py index 7174d2e78..443a1d371 100644 --- a/tests/unit/fixtures.py +++ b/tests/unit/fixtures.py @@ -6,7 +6,7 @@ import pytest -from llama_stack.distribution.store.registry import CachedDiskDistributionRegistry, DiskDistributionRegistry +from llama_stack.core.store.registry import CachedDiskDistributionRegistry, DiskDistributionRegistry from llama_stack.providers.utils.kvstore.config import SqliteKVStoreConfig from llama_stack.providers.utils.kvstore.sqlite import SqliteKVStoreImpl diff --git a/tests/unit/providers/agents/meta_reference/test_openai_responses.py b/tests/unit/providers/agents/meta_reference/test_openai_responses.py index 6485e3512..2ab5b557e 100644 --- a/tests/unit/providers/agents/meta_reference/test_openai_responses.py +++ b/tests/unit/providers/agents/meta_reference/test_openai_responses.py @@ -40,7 +40,7 @@ from llama_stack.apis.inference import ( OpenAIUserMessageParam, ) from llama_stack.apis.tools.tools import Tool, ToolGroups, ToolInvocationResult, ToolParameter, ToolRuntime -from llama_stack.distribution.access_control.access_control import default_policy +from llama_stack.core.access_control.access_control import default_policy from llama_stack.providers.inline.agents.meta_reference.openai_responses import ( OpenAIResponsesImpl, ) diff --git a/tests/unit/providers/agents/test_persistence_access_control.py b/tests/unit/providers/agents/test_persistence_access_control.py index 26001fcf1..93dd8ad95 100644 --- a/tests/unit/providers/agents/test_persistence_access_control.py +++ b/tests/unit/providers/agents/test_persistence_access_control.py @@ -12,7 +12,7 @@ import pytest from llama_stack.apis.agents import Turn from llama_stack.apis.inference import CompletionMessage, StopReason -from llama_stack.distribution.datatypes import User +from llama_stack.core.datatypes import User from llama_stack.providers.inline.agents.meta_reference.persistence import AgentPersistence, AgentSessionInfo diff --git a/tests/unit/providers/inference/test_inference_client_caching.py b/tests/unit/providers/inference/test_inference_client_caching.py index ba36a3e3d..b371cf907 100644 --- a/tests/unit/providers/inference/test_inference_client_caching.py +++ b/tests/unit/providers/inference/test_inference_client_caching.py @@ -7,7 +7,7 @@ import json from unittest.mock import MagicMock -from llama_stack.distribution.request_headers import request_provider_data_context +from llama_stack.core.request_headers import request_provider_data_context from llama_stack.providers.remote.inference.groq.config import GroqConfig from llama_stack.providers.remote.inference.groq.groq import GroqInferenceAdapter from llama_stack.providers.remote.inference.llama_openai_compat.config import LlamaCompatConfig diff --git a/tests/unit/providers/inference/test_openai_base_url_config.py b/tests/unit/providers/inference/test_openai_base_url_config.py index 453ac9089..150f6210b 100644 --- a/tests/unit/providers/inference/test_openai_base_url_config.py +++ b/tests/unit/providers/inference/test_openai_base_url_config.py @@ -7,7 +7,7 @@ import os from unittest.mock import AsyncMock, MagicMock, patch -from llama_stack.distribution.stack import replace_env_vars +from llama_stack.core.stack import replace_env_vars from llama_stack.providers.remote.inference.openai.config import OpenAIConfig from llama_stack.providers.remote.inference.openai.openai import OpenAIInferenceAdapter diff --git a/tests/unit/providers/nvidia/test_parameters.py b/tests/unit/providers/nvidia/test_parameters.py index 7e4323bd7..ad381da26 100644 --- a/tests/unit/providers/nvidia/test_parameters.py +++ b/tests/unit/providers/nvidia/test_parameters.py @@ -19,7 +19,7 @@ from llama_stack.apis.post_training.post_training import ( OptimizerType, TrainingConfig, ) -from llama_stack.distribution.library_client import convert_pydantic_to_json_value +from llama_stack.core.library_client import convert_pydantic_to_json_value from llama_stack.providers.remote.post_training.nvidia.post_training import ( NvidiaPostTrainingAdapter, NvidiaPostTrainingConfig, diff --git a/tests/unit/providers/nvidia/test_supervised_fine_tuning.py b/tests/unit/providers/nvidia/test_supervised_fine_tuning.py index bc474f3bc..91148605d 100644 --- a/tests/unit/providers/nvidia/test_supervised_fine_tuning.py +++ b/tests/unit/providers/nvidia/test_supervised_fine_tuning.py @@ -19,7 +19,7 @@ from llama_stack.apis.post_training.post_training import ( QATFinetuningConfig, TrainingConfig, ) -from llama_stack.distribution.library_client import convert_pydantic_to_json_value +from llama_stack.core.library_client import convert_pydantic_to_json_value from llama_stack.providers.remote.post_training.nvidia.post_training import ( ListNvidiaPostTrainingJobs, NvidiaPostTrainingAdapter, diff --git a/tests/unit/providers/test_configs.py b/tests/unit/providers/test_configs.py index 99081c8b0..867cfffbc 100644 --- a/tests/unit/providers/test_configs.py +++ b/tests/unit/providers/test_configs.py @@ -7,8 +7,8 @@ import pytest from pydantic import BaseModel -from llama_stack.distribution.distribution import get_provider_registry, providable_apis -from llama_stack.distribution.utils.dynamic import instantiate_class_type +from llama_stack.core.distribution import get_provider_registry, providable_apis +from llama_stack.core.utils.dynamic import instantiate_class_type class TestProviderConfigurations: diff --git a/tests/unit/registry/test_registry.py b/tests/unit/registry/test_registry.py index 87fe18d54..4ea4a20b9 100644 --- a/tests/unit/registry/test_registry.py +++ b/tests/unit/registry/test_registry.py @@ -9,7 +9,7 @@ import pytest from llama_stack.apis.inference import Model from llama_stack.apis.vector_dbs import VectorDB -from llama_stack.distribution.store.registry import ( +from llama_stack.core.store.registry import ( KEY_FORMAT, CachedDiskDistributionRegistry, DiskDistributionRegistry, diff --git a/tests/unit/registry/test_registry_acl.py b/tests/unit/registry/test_registry_acl.py index 6cfb20944..09b9a3cfb 100644 --- a/tests/unit/registry/test_registry_acl.py +++ b/tests/unit/registry/test_registry_acl.py @@ -6,8 +6,8 @@ from llama_stack.apis.models import ModelType -from llama_stack.distribution.datatypes import ModelWithOwner, User -from llama_stack.distribution.store.registry import CachedDiskDistributionRegistry +from llama_stack.core.datatypes import ModelWithOwner, User +from llama_stack.core.store.registry import CachedDiskDistributionRegistry async def test_registry_cache_with_acl(cached_disk_dist_registry): diff --git a/tests/unit/server/test_access_control.py b/tests/unit/server/test_access_control.py index d6a420c13..55449804a 100644 --- a/tests/unit/server/test_access_control.py +++ b/tests/unit/server/test_access_control.py @@ -12,9 +12,9 @@ from pydantic import TypeAdapter, ValidationError from llama_stack.apis.datatypes import Api from llama_stack.apis.models import ModelType -from llama_stack.distribution.access_control.access_control import AccessDeniedError, is_action_allowed -from llama_stack.distribution.datatypes import AccessRule, ModelWithOwner, User -from llama_stack.distribution.routing_tables.models import ModelsRoutingTable +from llama_stack.core.access_control.access_control import AccessDeniedError, is_action_allowed +from llama_stack.core.datatypes import AccessRule, ModelWithOwner, User +from llama_stack.core.routing_tables.models import ModelsRoutingTable class AsyncMock(MagicMock): @@ -40,7 +40,7 @@ async def test_setup(cached_disk_dist_registry): yield cached_disk_dist_registry, routing_table -@patch("llama_stack.distribution.routing_tables.common.get_authenticated_user") +@patch("llama_stack.core.routing_tables.common.get_authenticated_user") async def test_access_control_with_cache(mock_get_authenticated_user, test_setup): registry, routing_table = test_setup model_public = ModelWithOwner( @@ -104,7 +104,7 @@ async def test_access_control_with_cache(mock_get_authenticated_user, test_setup await routing_table.get_model("model-admin") -@patch("llama_stack.distribution.routing_tables.common.get_authenticated_user") +@patch("llama_stack.core.routing_tables.common.get_authenticated_user") async def test_access_control_and_updates(mock_get_authenticated_user, test_setup): registry, routing_table = test_setup model_public = ModelWithOwner( @@ -142,7 +142,7 @@ async def test_access_control_and_updates(mock_get_authenticated_user, test_setu assert model.identifier == "model-updates" -@patch("llama_stack.distribution.routing_tables.common.get_authenticated_user") +@patch("llama_stack.core.routing_tables.common.get_authenticated_user") async def test_access_control_empty_attributes(mock_get_authenticated_user, test_setup): registry, routing_table = test_setup model = ModelWithOwner( @@ -166,7 +166,7 @@ async def test_access_control_empty_attributes(mock_get_authenticated_user, test assert "model-empty-attrs" in model_ids -@patch("llama_stack.distribution.routing_tables.common.get_authenticated_user") +@patch("llama_stack.core.routing_tables.common.get_authenticated_user") async def test_no_user_attributes(mock_get_authenticated_user, test_setup): registry, routing_table = test_setup model_public = ModelWithOwner( @@ -196,7 +196,7 @@ async def test_no_user_attributes(mock_get_authenticated_user, test_setup): assert all_models.data[0].identifier == "model-public-2" -@patch("llama_stack.distribution.routing_tables.common.get_authenticated_user") +@patch("llama_stack.core.routing_tables.common.get_authenticated_user") async def test_automatic_access_attributes(mock_get_authenticated_user, test_setup): """Test that newly created resources inherit access attributes from their creator.""" registry, routing_table = test_setup @@ -275,7 +275,7 @@ async def test_setup_with_access_policy(cached_disk_dist_registry): yield routing_table -@patch("llama_stack.distribution.routing_tables.common.get_authenticated_user") +@patch("llama_stack.core.routing_tables.common.get_authenticated_user") async def test_access_policy(mock_get_authenticated_user, test_setup_with_access_policy): routing_table = test_setup_with_access_policy mock_get_authenticated_user.return_value = User( @@ -561,6 +561,6 @@ def test_invalid_condition(): ], ) def test_condition_reprs(condition): - from llama_stack.distribution.access_control.conditions import parse_condition + from llama_stack.core.access_control.conditions import parse_condition assert condition == str(parse_condition(condition)) diff --git a/tests/unit/server/test_auth.py b/tests/unit/server/test_auth.py index adf0140e2..37b543976 100644 --- a/tests/unit/server/test_auth.py +++ b/tests/unit/server/test_auth.py @@ -11,7 +11,7 @@ import pytest from fastapi import FastAPI from fastapi.testclient import TestClient -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( AuthenticationConfig, AuthProviderType, CustomAuthConfig, @@ -19,9 +19,9 @@ from llama_stack.distribution.datatypes import ( OAuth2JWKSConfig, OAuth2TokenAuthConfig, ) -from llama_stack.distribution.request_headers import User -from llama_stack.distribution.server.auth import AuthenticationMiddleware, _has_required_scope -from llama_stack.distribution.server.auth_providers import ( +from llama_stack.core.request_headers import User +from llama_stack.core.server.auth import AuthenticationMiddleware, _has_required_scope +from llama_stack.core.server.auth_providers import ( get_attributes_from_claims, ) @@ -150,10 +150,10 @@ def scope_middleware_with_mocks(mock_auth_endpoint): else: raise ValueError("No matching route") - import llama_stack.distribution.server.auth + import llama_stack.core.server.auth - llama_stack.distribution.server.auth.find_matching_route = mock_find_matching_route - llama_stack.distribution.server.auth.initialize_route_impls = lambda impls: {} + llama_stack.core.server.auth.find_matching_route = mock_find_matching_route + llama_stack.core.server.auth.initialize_route_impls = lambda impls: {} return middleware, mock_app diff --git a/tests/unit/server/test_auth_github.py b/tests/unit/server/test_auth_github.py index 21d2f2c6a..d87643579 100644 --- a/tests/unit/server/test_auth_github.py +++ b/tests/unit/server/test_auth_github.py @@ -11,8 +11,8 @@ import pytest from fastapi import FastAPI from fastapi.testclient import TestClient -from llama_stack.distribution.datatypes import AuthenticationConfig, AuthProviderType, GitHubTokenAuthConfig -from llama_stack.distribution.server.auth import AuthenticationMiddleware +from llama_stack.core.datatypes import AuthenticationConfig, AuthProviderType, GitHubTokenAuthConfig +from llama_stack.core.server.auth import AuthenticationMiddleware class MockResponse: @@ -78,7 +78,7 @@ def test_authenticated_endpoint_with_invalid_bearer_format(github_token_client): assert "Invalid Authorization header format" in response.json()["error"]["message"] -@patch("llama_stack.distribution.server.auth_providers.httpx.AsyncClient") +@patch("llama_stack.core.server.auth_providers.httpx.AsyncClient") def test_authenticated_endpoint_with_valid_github_token(mock_client_class, github_token_client): """Test accessing protected endpoint with valid GitHub token""" # Mock the GitHub API responses @@ -118,7 +118,7 @@ def test_authenticated_endpoint_with_valid_github_token(mock_client_class, githu assert calls[0][1]["headers"]["Authorization"] == "Bearer github_token_123" -@patch("llama_stack.distribution.server.auth_providers.httpx.AsyncClient") +@patch("llama_stack.core.server.auth_providers.httpx.AsyncClient") def test_authenticated_endpoint_with_invalid_github_token(mock_client_class, github_token_client): """Test accessing protected endpoint with invalid GitHub token""" # Mock the GitHub API to return 401 Unauthorized @@ -135,7 +135,7 @@ def test_authenticated_endpoint_with_invalid_github_token(mock_client_class, git ) -@patch("llama_stack.distribution.server.auth_providers.httpx.AsyncClient") +@patch("llama_stack.core.server.auth_providers.httpx.AsyncClient") def test_github_enterprise_support(mock_client_class): """Test GitHub Enterprise support with custom API base URL""" app = FastAPI() diff --git a/tests/unit/server/test_quota.py b/tests/unit/server/test_quota.py index 763bf8e94..85acbc66a 100644 --- a/tests/unit/server/test_quota.py +++ b/tests/unit/server/test_quota.py @@ -9,8 +9,8 @@ from fastapi import FastAPI, Request from fastapi.testclient import TestClient from starlette.middleware.base import BaseHTTPMiddleware -from llama_stack.distribution.datatypes import QuotaConfig, QuotaPeriod -from llama_stack.distribution.server.quota import QuotaMiddleware +from llama_stack.core.datatypes import QuotaConfig, QuotaPeriod +from llama_stack.core.server.quota import QuotaMiddleware from llama_stack.providers.utils.kvstore.config import SqliteKVStoreConfig diff --git a/tests/unit/server/test_replace_env_vars.py b/tests/unit/server/test_replace_env_vars.py index 55817044d..0dda682c0 100644 --- a/tests/unit/server/test_replace_env_vars.py +++ b/tests/unit/server/test_replace_env_vars.py @@ -8,7 +8,7 @@ import os import pytest -from llama_stack.distribution.stack import replace_env_vars +from llama_stack.core.stack import replace_env_vars @pytest.fixture diff --git a/tests/unit/server/test_resolver.py b/tests/unit/server/test_resolver.py index a348590b1..1ee1b2f47 100644 --- a/tests/unit/server/test_resolver.py +++ b/tests/unit/server/test_resolver.py @@ -12,14 +12,14 @@ from unittest.mock import AsyncMock, MagicMock from pydantic import BaseModel, Field from llama_stack.apis.inference import Inference -from llama_stack.distribution.datatypes import ( +from llama_stack.core.datatypes import ( Api, Provider, StackRunConfig, ) -from llama_stack.distribution.resolver import resolve_impls -from llama_stack.distribution.routers.inference import InferenceRouter -from llama_stack.distribution.routing_tables.models import ModelsRoutingTable +from llama_stack.core.resolver import resolve_impls +from llama_stack.core.routers.inference import InferenceRouter +from llama_stack.core.routing_tables.models import ModelsRoutingTable from llama_stack.providers.datatypes import InlineProviderSpec, ProviderSpec diff --git a/tests/unit/server/test_server.py b/tests/unit/server/test_server.py index d17d58b8a..803111fc7 100644 --- a/tests/unit/server/test_server.py +++ b/tests/unit/server/test_server.py @@ -10,9 +10,9 @@ from fastapi import HTTPException from openai import BadRequestError from pydantic import ValidationError -from llama_stack.distribution.access_control.access_control import AccessDeniedError -from llama_stack.distribution.datatypes import AuthenticationRequiredError -from llama_stack.distribution.server.server import translate_exception +from llama_stack.core.access_control.access_control import AccessDeniedError +from llama_stack.core.datatypes import AuthenticationRequiredError +from llama_stack.core.server.server import translate_exception class TestTranslateException: @@ -29,7 +29,7 @@ class TestTranslateException: def test_translate_access_denied_error_with_context(self): """Test that AccessDeniedError with context includes detailed information.""" - from llama_stack.distribution.datatypes import User + from llama_stack.core.datatypes import User # Create mock user and resource user = User("test-user", {"roles": ["user"], "teams": ["dev"]}) diff --git a/tests/unit/server/test_sse.py b/tests/unit/server/test_sse.py index d42857186..54afe4ee4 100644 --- a/tests/unit/server/test_sse.py +++ b/tests/unit/server/test_sse.py @@ -8,7 +8,7 @@ import asyncio from unittest.mock import AsyncMock, MagicMock from llama_stack.apis.common.responses import PaginatedResponse -from llama_stack.distribution.server.server import create_dynamic_typed_route, create_sse_event, sse_generator +from llama_stack.core.server.server import create_dynamic_typed_route, create_sse_event, sse_generator async def test_sse_generator_basic(): diff --git a/tests/unit/utils/test_authorized_sqlstore.py b/tests/unit/utils/test_authorized_sqlstore.py index 066f67a98..90eb706e4 100644 --- a/tests/unit/utils/test_authorized_sqlstore.py +++ b/tests/unit/utils/test_authorized_sqlstore.py @@ -7,9 +7,9 @@ from tempfile import TemporaryDirectory from unittest.mock import patch -from llama_stack.distribution.access_control.access_control import default_policy, is_action_allowed -from llama_stack.distribution.access_control.datatypes import Action -from llama_stack.distribution.datatypes import User +from llama_stack.core.access_control.access_control import default_policy, is_action_allowed +from llama_stack.core.access_control.datatypes import Action +from llama_stack.core.datatypes import User from llama_stack.providers.utils.sqlstore.api import ColumnType from llama_stack.providers.utils.sqlstore.authorized_sqlstore import AuthorizedSqlStore, SqlRecord from llama_stack.providers.utils.sqlstore.sqlalchemy_sqlstore import SqlAlchemySqlStoreImpl diff --git a/tests/verifications/openai_api/test_responses.py b/tests/verifications/openai_api/test_responses.py index 08bbb2252..e312de6aa 100644 --- a/tests/verifications/openai_api/test_responses.py +++ b/tests/verifications/openai_api/test_responses.py @@ -13,7 +13,7 @@ import openai import pytest from llama_stack import LlamaStackAsLibraryClient -from llama_stack.distribution.datatypes import AuthenticationRequiredError +from llama_stack.core.datatypes import AuthenticationRequiredError from tests.common.mcp import dependency_tools, make_mcp_server from tests.verifications.openai_api.fixtures.fixtures import ( case_id_generator, From cf731461322584252ceacbee6cd629aab8431cc5 Mon Sep 17 00:00:00 2001 From: Nehanth Narendrula Date: Thu, 31 Jul 2025 02:33:36 -0400 Subject: [PATCH 02/19] feat: Enable DPO training with HuggingFace inline provider (#2825) What does this PR do? This PR adds support for Direct Preference Optimization (DPO) training via the existing HuggingFace inline provider. It introduces a new DPO training recipe, config schema updates, dataset integration, and end-to-end testing to support preference-based fine-tuning with TRL. Test Plan Added integration test: tests/integration/post_training/test_post_training.py::TestPostTraining::test_preference_optimize Ran tests on both CPU and CUDA environments --------- Co-authored-by: Ubuntu Co-authored-by: Ashwin Bharambe --- .../post_training/inline_huggingface.md | 4 + .../post_training/huggingface/config.py | 6 + .../huggingface/post_training.py | 45 +- .../recipes/finetune_single_device.py | 236 +-------- .../recipes/finetune_single_device_dpo.py | 485 ++++++++++++++++++ .../inline/post_training/huggingface/utils.py | 269 ++++++++++ .../post_training/test_post_training.py | 83 ++- 7 files changed, 913 insertions(+), 215 deletions(-) create mode 100644 llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device_dpo.py create mode 100644 llama_stack/providers/inline/post_training/huggingface/utils.py diff --git a/docs/source/providers/post_training/inline_huggingface.md b/docs/source/providers/post_training/inline_huggingface.md index 82b08bf7a..0a8745e71 100644 --- a/docs/source/providers/post_training/inline_huggingface.md +++ b/docs/source/providers/post_training/inline_huggingface.md @@ -24,6 +24,10 @@ HuggingFace-based post-training provider for fine-tuning models using the Huggin | `weight_decay` | `` | No | 0.01 | | | `dataloader_num_workers` | `` | No | 4 | | | `dataloader_pin_memory` | `` | No | True | | +| `dpo_beta` | `` | No | 0.1 | | +| `use_reference_model` | `` | No | True | | +| `dpo_loss_type` | `Literal['sigmoid', 'hinge', 'ipo', 'kto_pair'` | No | sigmoid | | +| `dpo_output_dir` | `` | No | ./checkpoints/dpo | | ## Sample Configuration diff --git a/llama_stack/providers/inline/post_training/huggingface/config.py b/llama_stack/providers/inline/post_training/huggingface/config.py index 06c6d8073..dae8fcc04 100644 --- a/llama_stack/providers/inline/post_training/huggingface/config.py +++ b/llama_stack/providers/inline/post_training/huggingface/config.py @@ -67,6 +67,12 @@ class HuggingFacePostTrainingConfig(BaseModel): # Can improve data transfer speed to GPU but uses more memory dataloader_pin_memory: bool = True + # DPO-specific parameters + dpo_beta: float = 0.1 + use_reference_model: bool = True + dpo_loss_type: Literal["sigmoid", "hinge", "ipo", "kto_pair"] = "sigmoid" + dpo_output_dir: str = "./checkpoints/dpo" + @classmethod def sample_run_config(cls, __distro_dir__: str, **kwargs: Any) -> dict[str, Any]: return {"checkpoint_format": "huggingface", "distributed_backend": None, "device": "cpu"} diff --git a/llama_stack/providers/inline/post_training/huggingface/post_training.py b/llama_stack/providers/inline/post_training/huggingface/post_training.py index 0b2760792..81622e2b7 100644 --- a/llama_stack/providers/inline/post_training/huggingface/post_training.py +++ b/llama_stack/providers/inline/post_training/huggingface/post_training.py @@ -25,6 +25,9 @@ from llama_stack.providers.inline.post_training.huggingface.config import ( from llama_stack.providers.inline.post_training.huggingface.recipes.finetune_single_device import ( HFFinetuningSingleDevice, ) +from llama_stack.providers.inline.post_training.huggingface.recipes.finetune_single_device_dpo import ( + HFDPOAlignmentSingleDevice, +) from llama_stack.providers.utils.scheduler import JobArtifact, Scheduler from llama_stack.providers.utils.scheduler import JobStatus as SchedulerJobStatus from llama_stack.schema_utils import webmethod @@ -36,6 +39,7 @@ class TrainingArtifactType(Enum): _JOB_TYPE_SUPERVISED_FINE_TUNE = "supervised-fine-tune" +_JOB_TYPE_DPO_TRAINING = "dpo-training" class HuggingFacePostTrainingImpl: @@ -119,12 +123,37 @@ class HuggingFacePostTrainingImpl: hyperparam_search_config: dict[str, Any], logger_config: dict[str, Any], ) -> PostTrainingJob: - raise NotImplementedError("DPO alignment is not implemented yet") + async def handler(on_log_message_cb, on_status_change_cb, on_artifact_collected_cb): + on_log_message_cb("Starting HF DPO alignment") - async def get_training_jobs(self) -> ListPostTrainingJobsResponse: - return ListPostTrainingJobsResponse( - data=[PostTrainingJob(job_uuid=job.id) for job in self._scheduler.get_jobs()] - ) + recipe = HFDPOAlignmentSingleDevice( + job_uuid=job_uuid, + datasetio_api=self.datasetio_api, + datasets_api=self.datasets_api, + ) + + resources_allocated, checkpoints = await recipe.train( + model=finetuned_model, + output_dir=f"{self.config.dpo_output_dir}/{job_uuid}", + job_uuid=job_uuid, + dpo_config=algorithm_config, + config=training_config, + provider_config=self.config, + ) + + on_artifact_collected_cb(self._resources_stats_to_artifact(resources_allocated)) + if checkpoints: + for checkpoint in checkpoints: + artifact = self._checkpoint_to_artifact(checkpoint) + on_artifact_collected_cb(artifact) + else: + on_log_message_cb("Warning: No checkpoints were saved during DPO training") + + on_status_change_cb(SchedulerJobStatus.completed) + on_log_message_cb("HF DPO alignment completed") + + job_uuid = self._scheduler.schedule(_JOB_TYPE_DPO_TRAINING, job_uuid, handler) + return PostTrainingJob(job_uuid=job_uuid) @staticmethod def _get_artifacts_metadata_by_type(job, artifact_type): @@ -174,3 +203,9 @@ class HuggingFacePostTrainingImpl: async def get_training_job_artifacts(self, job_uuid: str) -> PostTrainingJobArtifactsResponse | None: job = self._scheduler.get_job(job_uuid) return PostTrainingJobArtifactsResponse(job_uuid=job_uuid, checkpoints=self._get_checkpoints(job)) + + @webmethod(route="/post-training/jobs", method="GET") + async def get_training_jobs(self) -> ListPostTrainingJobsResponse: + return ListPostTrainingJobsResponse( + data=[PostTrainingJob(job_uuid=job.id) for job in self._scheduler.get_jobs()] + ) diff --git a/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device.py b/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device.py index 2a024eb25..2574b995b 100644 --- a/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device.py +++ b/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device.py @@ -8,30 +8,13 @@ import gc import json import logging import multiprocessing -import os -import signal -import sys -from datetime import UTC, datetime from pathlib import Path from typing import Any -import psutil - -from llama_stack.providers.inline.post_training.common.utils import evacuate_model_from_device - -# Set tokenizer parallelism environment variable -os.environ["TOKENIZERS_PARALLELISM"] = "false" - -# Force PyTorch to use OpenBLAS instead of MKL -os.environ["MKL_THREADING_LAYER"] = "GNU" -os.environ["MKL_SERVICE_FORCE_INTEL"] = "0" -os.environ["MKL_NUM_THREADS"] = "1" - import torch from datasets import Dataset from peft import LoraConfig from transformers import ( - AutoConfig, AutoModelForCausalLM, AutoTokenizer, ) @@ -45,93 +28,25 @@ from llama_stack.apis.post_training import ( LoraFinetuningConfig, TrainingConfig, ) +from llama_stack.providers.inline.post_training.common.utils import evacuate_model_from_device from ..config import HuggingFacePostTrainingConfig +from ..utils import ( + calculate_training_steps, + create_checkpoints, + get_memory_stats, + get_save_strategy, + load_model, + load_rows_from_dataset, + setup_environment, + setup_signal_handlers, + setup_torch_device, + split_dataset, +) logger = logging.getLogger(__name__) -def get_gb(to_convert: int) -> str: - """Converts memory stats to GB and formats to 2 decimal places. - Args: - to_convert: Memory value in bytes - Returns: - str: Memory value in GB formatted to 2 decimal places - """ - return f"{(to_convert / (1024**3)):.2f}" - - -def get_memory_stats(device: torch.device) -> dict[str, Any]: - """Get memory statistics for the given device.""" - stats = { - "system_memory": { - "total": get_gb(psutil.virtual_memory().total), - "available": get_gb(psutil.virtual_memory().available), - "used": get_gb(psutil.virtual_memory().used), - "percent": psutil.virtual_memory().percent, - } - } - - if device.type == "cuda": - stats["device_memory"] = { - "allocated": get_gb(torch.cuda.memory_allocated(device)), - "reserved": get_gb(torch.cuda.memory_reserved(device)), - "max_allocated": get_gb(torch.cuda.max_memory_allocated(device)), - } - elif device.type == "mps": - # MPS doesn't provide direct memory stats, but we can track system memory - stats["device_memory"] = { - "note": "MPS memory stats not directly available", - "system_memory_used": get_gb(psutil.virtual_memory().used), - } - elif device.type == "cpu": - # For CPU, we track process memory usage - process = psutil.Process() - stats["device_memory"] = { - "process_rss": get_gb(process.memory_info().rss), - "process_vms": get_gb(process.memory_info().vms), - "process_percent": process.memory_percent(), - } - - return stats - - -def setup_torch_device(device_str: str) -> torch.device: - """Initialize and validate a PyTorch device. - This function handles device initialization and validation for different device types: - - CUDA: Validates CUDA availability and handles device selection - - MPS: Validates MPS availability for Apple Silicon - - CPU: Basic validation - - HPU: Raises error as it's not supported - Args: - device_str: String specifying the device ('cuda', 'cpu', 'mps') - Returns: - torch.device: The initialized and validated device - Raises: - RuntimeError: If device initialization fails or device is not supported - """ - try: - device = torch.device(device_str) - except RuntimeError as e: - raise RuntimeError(f"Error getting Torch Device {str(e)}") from e - - # Validate device capabilities - if device.type == "cuda": - if not torch.cuda.is_available(): - raise RuntimeError( - f"{device.type}: Torch has no CUDA/ROCm support or could not detect a compatible device." - ) - if device.index is None: - device = torch.device(device.type, torch.cuda.current_device()) - elif device.type == "mps": - if not torch.backends.mps.is_available(): - raise RuntimeError(f"{device.type}: Torch has no MPS support or could not detect a compatible device.") - elif device.type == "hpu": - raise RuntimeError(f"{device.type}: training does not support Intel Gaudi.") - - return device - - class HFFinetuningSingleDevice: def __init__( self, @@ -262,19 +177,6 @@ class HFFinetuningSingleDevice: remove_columns=ds.column_names, ) - async def _setup_data(self, dataset_id: str) -> list[dict[str, Any]]: - """Load dataset from llama stack dataset provider""" - try: - all_rows = await self.datasetio_api.iterrows( - dataset_id=dataset_id, - limit=-1, - ) - if not isinstance(all_rows.data, list): - raise RuntimeError("Expected dataset data to be a list") - return all_rows.data - except Exception as e: - raise RuntimeError(f"Failed to load dataset: {str(e)}") from e - def _run_training_sync( self, model: str, @@ -327,7 +229,7 @@ class HFFinetuningSingleDevice: # Load dataset logger.info(f"Loading dataset: {config.data_config.dataset_id}") - rows = await self._setup_data(config.data_config.dataset_id) + rows = await load_rows_from_dataset(self.datasetio_api, config.data_config.dataset_id) if not self.validate_dataset_format(rows): raise ValueError("Dataset is missing required fields: input_query, expected_answer, chat_completion_input") logger.info(f"Loaded {len(rows)} rows from dataset") @@ -369,47 +271,10 @@ class HFFinetuningSingleDevice: raise ValueError(f"Failed to create dataset: {str(e)}") from e # Split dataset - logger.info("Splitting dataset into train and validation sets") - train_val_split = ds.train_test_split(test_size=0.1, seed=42) - train_dataset = train_val_split["train"] - eval_dataset = train_val_split["test"] - logger.info(f"Split dataset into {len(train_dataset)} training and {len(eval_dataset)} validation examples") + train_dataset, eval_dataset = split_dataset(ds) return train_dataset, eval_dataset, tokenizer - def load_model( - self, - model: str, - device: torch.device, - provider_config: HuggingFacePostTrainingConfig, - ) -> AutoModelForCausalLM: - """Load and initialize the model for training. - Args: - model: The model identifier to load - device: The device to load the model onto - provider_config: Provider-specific configuration - Returns: - The loaded and initialized model - Raises: - RuntimeError: If model loading fails - """ - logger.info("Loading the base model") - try: - model_config = AutoConfig.from_pretrained(model, **provider_config.model_specific_config) - model_obj = AutoModelForCausalLM.from_pretrained( - model, - torch_dtype="auto" if device.type != "cpu" else "float32", - quantization_config=None, - config=model_config, - **provider_config.model_specific_config, - ) - # Always move model to specified device - model_obj = model_obj.to(device) - logger.info(f"Model loaded and moved to device: {model_obj.device}") - return model_obj - except Exception as e: - raise RuntimeError(f"Failed to load model: {str(e)}") from e - def setup_training_args( self, config: TrainingConfig, @@ -439,27 +304,12 @@ class HFFinetuningSingleDevice: raise ValueError("DataConfig is required for training") data_config = config.data_config - # Calculate steps - total_steps = steps_per_epoch * config.n_epochs - max_steps = min(config.max_steps_per_epoch, total_steps) - logging_steps = max(1, steps_per_epoch // 50) # Log 50 times per epoch - - logger.info("Training configuration:") - logger.info(f"- Steps per epoch: {steps_per_epoch}") - logger.info(f"- Total steps: {total_steps}") - logger.info(f"- Max steps: {max_steps}") - logger.info(f"- Logging steps: {logging_steps}") - - # Configure save strategy - save_strategy = "no" - eval_strategy = "no" - if output_dir_path: - save_strategy = "epoch" - eval_strategy = "epoch" - logger.info(f"Will save checkpoints to {output_dir_path}") + # Calculate steps and get save strategy + step_info = calculate_training_steps(steps_per_epoch, config) + save_strategy, eval_strategy = get_save_strategy(output_dir_path) return SFTConfig( - max_steps=max_steps, + max_steps=step_info["max_steps"], output_dir=str(output_dir_path) if output_dir_path is not None else None, num_train_epochs=config.n_epochs, per_device_train_batch_size=data_config.batch_size, @@ -483,7 +333,7 @@ class HFFinetuningSingleDevice: load_best_model_at_end=True if output_dir_path else False, metric_for_best_model="eval_loss", greater_is_better=False, - logging_steps=logging_steps, + logging_steps=step_info["logging_steps"], ) def save_model( @@ -523,13 +373,11 @@ class HFFinetuningSingleDevice: ) -> None: """Run the training process with signal handling.""" - def signal_handler(signum, frame): - """Handle termination signals gracefully.""" - logger.info(f"Received signal {signum}, initiating graceful shutdown") - sys.exit(0) + # Setup environment variables + setup_environment() - signal.signal(signal.SIGTERM, signal_handler) - signal.signal(signal.SIGINT, signal_handler) + # Setup signal handlers + setup_signal_handlers() # Convert config dicts back to objects logger.info("Initializing configuration objects") @@ -558,7 +406,7 @@ class HFFinetuningSingleDevice: ) # Load model - model_obj = self.load_model(model, device, provider_config_obj) + model_obj = load_model(model, device, provider_config_obj) # Initialize trainer logger.info("Initializing SFTTrainer") @@ -633,7 +481,7 @@ class HFFinetuningSingleDevice: # Train in a separate process logger.info("Starting training in separate process") try: - # Set multiprocessing start method to 'spawn' for CUDA/MPS compatibility + # Setup multiprocessing for device if device.type in ["cuda", "mps"]: multiprocessing.set_start_method("spawn", force=True) @@ -663,37 +511,7 @@ class HFFinetuningSingleDevice: checkpoints = [] if output_dir_path: - # Get all checkpoint directories and sort them numerically - checkpoint_dirs = sorted( - [d for d in output_dir_path.glob("checkpoint-*") if d.is_dir()], - key=lambda x: int(x.name.split("-")[1]), - ) - - # Add all checkpoint directories - for epoch_number, checkpoint_dir in enumerate(checkpoint_dirs, start=1): - # Get the creation time of the directory - created_time = datetime.fromtimestamp(os.path.getctime(checkpoint_dir), tz=UTC) - - checkpoint = Checkpoint( - identifier=checkpoint_dir.name, - created_at=created_time, - epoch=epoch_number, - post_training_job_id=job_uuid, - path=str(checkpoint_dir), - ) - checkpoints.append(checkpoint) - - # Add the merged model as a checkpoint - merged_model_path = output_dir_path / "merged_model" - if merged_model_path.exists(): - checkpoint = Checkpoint( - identifier=f"{model}-sft-{config.n_epochs}", - created_at=datetime.now(UTC), - epoch=config.n_epochs, - post_training_job_id=job_uuid, - path=str(merged_model_path), - ) - checkpoints.append(checkpoint) + checkpoints = create_checkpoints(output_dir_path, job_uuid, model, config, "merged_model") return memory_stats, checkpoints if checkpoints else None finally: diff --git a/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device_dpo.py b/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device_dpo.py new file mode 100644 index 000000000..a7c19faac --- /dev/null +++ b/llama_stack/providers/inline/post_training/huggingface/recipes/finetune_single_device_dpo.py @@ -0,0 +1,485 @@ +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +import gc +import logging +import multiprocessing +from pathlib import Path +from typing import Any + +import torch +from datasets import Dataset +from transformers import ( + AutoTokenizer, +) +from trl import DPOConfig, DPOTrainer + +from llama_stack.apis.datasetio import DatasetIO +from llama_stack.apis.datasets import Datasets +from llama_stack.apis.post_training import ( + Checkpoint, + DPOAlignmentConfig, + TrainingConfig, +) +from llama_stack.providers.inline.post_training.common.utils import evacuate_model_from_device + +from ..config import HuggingFacePostTrainingConfig +from ..utils import ( + calculate_training_steps, + create_checkpoints, + get_memory_stats, + get_save_strategy, + load_model, + load_rows_from_dataset, + setup_environment, + setup_signal_handlers, + setup_torch_device, + split_dataset, +) + +logger = logging.getLogger(__name__) + + +class HFDPOAlignmentSingleDevice: + def __init__( + self, + job_uuid: str, + datasetio_api: DatasetIO, + datasets_api: Datasets, + ): + self.datasetio_api = datasetio_api + self.datasets_api = datasets_api + self.job_uuid = job_uuid + + def validate_dataset_format(self, rows: list[dict]) -> None: + """Validate that the dataset has the required fields for DPO training.""" + required_fields = ["prompt", "chosen", "rejected"] + + if not rows: + logger.warning("Dataset is empty") + raise ValueError("Dataset is empty") + + for i, row in enumerate(rows): + if not isinstance(row, dict): + logger.warning(f"Row {i} is not a dictionary") + raise ValueError(f"Row {i} is not a dictionary") + + for field in required_fields: + if field not in row: + logger.warning(f"Row {i} missing required DPO field: {field}") + raise ValueError(f"Row {i} missing required DPO field: {field}") + + # Handle both string and list formats + if field == "prompt": + # Prompt should be a string + if not isinstance(row[field], str): + logger.warning(f"Row {i} field '{field}' is not a string") + raise ValueError(f"Row {i} field '{field}' is not a string") + if not row[field].strip(): + logger.warning(f"Row {i} field '{field}' is empty") + raise ValueError(f"Row {i} field '{field}' is empty") + else: + # chosen/rejected can be either strings or lists of messages + if isinstance(row[field], str): + if not row[field].strip(): + logger.warning(f"Row {i} field '{field}' is empty") + raise ValueError(f"Row {i} field '{field}' is empty") + elif isinstance(row[field], list): + if not row[field]: + logger.warning(f"Row {i} field '{field}' is empty list") + raise ValueError(f"Row {i} field '{field}' is empty list") + else: + logger.warning(f"Row {i} field '{field}' is neither string nor list") + raise ValueError(f"Row {i} field '{field}' is neither string nor list") + + logger.info(f"DPO dataset validation passed: {len(rows)} preference examples") + + def _process_dpo_format(self, row: dict) -> tuple[str | None, str | None, str | None]: + """Process a row in DPO format, handling both string and conversation list formats.""" + if all(field in row for field in ["prompt", "chosen", "rejected"]): + prompt = row["prompt"] + + # Handle chosen field - convert list to string if needed + if isinstance(row["chosen"], list): + # For conversation format, concatenate messages + chosen = "\n".join( + [msg.get("content", "") if isinstance(msg, dict) else str(msg) for msg in row["chosen"]] + ) + else: + chosen = row["chosen"] + + # Handle rejected field - convert list to string if needed + if isinstance(row["rejected"], list): + # For conversation format, concatenate messages + rejected = "\n".join( + [msg.get("content", "") if isinstance(msg, dict) else str(msg) for msg in row["rejected"]] + ) + else: + rejected = row["rejected"] + + return prompt, chosen, rejected + return None, None, None + + def _format_text_for_dpo(self, prompt: str, response: str, provider_config: HuggingFacePostTrainingConfig) -> str: + """Format prompt and response text based on model requirements.""" + if hasattr(provider_config, "chat_template") and provider_config.chat_template: + # Use the chat template, supporting both {prompt}/{response} and {input}/{output} + template = provider_config.chat_template + # Try prompt/response first (DPO style) + if "{prompt}" in template and "{response}" in template: + return template.format(prompt=prompt, response=response) + # Fall back to input/output (SFT style) + elif "{input}" in template and "{output}" in template: + return template.format(input=prompt, output=response) + else: + # If template doesn't have expected placeholders, use default + return f"{prompt}\n{response}" + return f"{prompt}\n{response}" + + def _create_dataset( + self, rows: list[dict], config: TrainingConfig, provider_config: HuggingFacePostTrainingConfig + ) -> Dataset: + """Create and preprocess the dataset for DPO.""" + dpo_examples = [] + for row in rows: + prompt, chosen, rejected = self._process_dpo_format(row) + + if prompt and chosen and rejected: + # Format the texts + chosen_formatted = self._format_text_for_dpo(prompt, chosen, provider_config) + rejected_formatted = self._format_text_for_dpo(prompt, rejected, provider_config) + + dpo_examples.append( + { + "prompt": prompt, + "chosen": chosen_formatted, + "rejected": rejected_formatted, + } + ) + + if not dpo_examples: + raise ValueError("No valid preference examples found in dataset") + + logger.info(f"Created DPO dataset with {len(dpo_examples)} preference pairs") + return Dataset.from_list(dpo_examples) + + def _preprocess_dataset( + self, ds: Dataset, tokenizer: AutoTokenizer, provider_config: HuggingFacePostTrainingConfig + ) -> Dataset: + """Preprocess the dataset with tokenizer for DPO.""" + # DPOTrainer expects raw text, so we don't tokenize here + # Just return the dataset as is + return ds + + def _run_training_sync( + self, + model: str, + provider_config: dict[str, Any], + dpo_config: dict[str, Any], + config: dict[str, Any], + output_dir_path: Path | None, + ) -> None: + """Synchronous wrapper for running DPO training process.""" + import asyncio + + logger.info("Starting DPO training process with async wrapper") + asyncio.run( + self._run_training( + model=model, + provider_config=provider_config, + dpo_config=dpo_config, + config=config, + output_dir_path=output_dir_path, + ) + ) + + async def load_dataset( + self, + model: str, + config: TrainingConfig, + provider_config: HuggingFacePostTrainingConfig, + ) -> tuple[Dataset, Dataset, AutoTokenizer]: + """Load and prepare the dataset for DPO training.""" + # Validate data config + if not config.data_config: + raise ValueError("DataConfig is required for DPO training") + + # Load dataset + logger.info(f"Loading dataset: {config.data_config.dataset_id}") + rows = await load_rows_from_dataset(self.datasetio_api, config.data_config.dataset_id) + self.validate_dataset_format(rows) + logger.info(f"Loaded {len(rows)} rows from dataset") + + # Initialize tokenizer + logger.info(f"Initializing tokenizer for model: {model}") + try: + tokenizer = AutoTokenizer.from_pretrained(model, **provider_config.model_specific_config) + + # Set pad token to eos token if not present + if not tokenizer.pad_token: + tokenizer.pad_token = tokenizer.eos_token + + # Set padding side to left for DPO + tokenizer.padding_side = "left" + + # Set truncation side to right to keep the beginning of the sequence + tokenizer.truncation_side = "right" + + # Set model max length to match provider config + tokenizer.model_max_length = provider_config.max_seq_length + + logger.info("Tokenizer initialized successfully for DPO") + except Exception as e: + raise RuntimeError(f"Failed to initialize tokenizer: {str(e)}") from e + + # Create and preprocess dataset + logger.info("Creating and preprocessing dataset for DPO") + try: + ds = self._create_dataset(rows, config, provider_config) + ds = self._preprocess_dataset(ds, tokenizer, provider_config) + logger.info(f"Dataset created with {len(ds)} examples") + except Exception as e: + raise ValueError(f"Failed to create dataset: {str(e)}") from e + + # Split dataset + train_dataset, eval_dataset = split_dataset(ds) + + return train_dataset, eval_dataset, tokenizer + + def setup_training_args( + self, + config: TrainingConfig, + provider_config: HuggingFacePostTrainingConfig, + dpo_config: DPOAlignmentConfig, + device: torch.device, + output_dir_path: Path | None, + steps_per_epoch: int, + ) -> DPOConfig: + """Setup DPO training arguments.""" + logger.info("Configuring DPO training arguments") + lr = 5e-7 # Lower learning rate for DPO + if config.optimizer_config: + lr = config.optimizer_config.lr + logger.info(f"Using custom learning rate: {lr}") + + # Validate data config + if not config.data_config: + raise ValueError("DataConfig is required for training") + data_config = config.data_config + + # Calculate steps and get save strategy + step_info = calculate_training_steps(steps_per_epoch, config) + save_strategy, eval_strategy = get_save_strategy(output_dir_path) + + logger.info("DPO training configuration:") + logger.info(f"- DPO beta: {dpo_config.beta}") + logger.info(f"- DPO loss type: {provider_config.dpo_loss_type}") + + # Calculate max prompt length as half of max sequence length + max_prompt_length = provider_config.max_seq_length // 2 + + return DPOConfig( + max_steps=step_info["max_steps"], + output_dir=str(output_dir_path) if output_dir_path is not None else None, + num_train_epochs=config.n_epochs, + per_device_train_batch_size=data_config.batch_size, + fp16=device.type == "cuda", + bf16=False, # Causes CPU issues. + eval_strategy=eval_strategy, + use_cpu=True if device.type == "cpu" and not torch.backends.mps.is_available() else False, + save_strategy=save_strategy, + report_to="none", + max_length=provider_config.max_seq_length, + max_prompt_length=max_prompt_length, + gradient_accumulation_steps=config.gradient_accumulation_steps, + gradient_checkpointing=provider_config.gradient_checkpointing, + learning_rate=lr, + warmup_ratio=provider_config.warmup_ratio, + weight_decay=provider_config.weight_decay, + remove_unused_columns=False, + dataloader_pin_memory=provider_config.dataloader_pin_memory, + dataloader_num_workers=provider_config.dataloader_num_workers, + load_best_model_at_end=True if output_dir_path else False, + metric_for_best_model="eval_loss", + greater_is_better=False, + logging_steps=step_info["logging_steps"], + save_total_limit=provider_config.save_total_limit, + # DPO specific parameters + beta=dpo_config.beta, + loss_type=provider_config.dpo_loss_type, + ) + + def save_model( + self, + trainer: DPOTrainer, + output_dir_path: Path, + ) -> None: + """Save the trained DPO model.""" + logger.info("Saving final DPO model") + + save_path = output_dir_path / "dpo_model" + logger.info(f"Saving model to {save_path}") + + # Save model and tokenizer + trainer.save_model(str(save_path)) + + async def _run_training( + self, + model: str, + provider_config: dict[str, Any], + dpo_config: dict[str, Any], + config: dict[str, Any], + output_dir_path: Path | None, + ) -> None: + """Run the DPO training process with signal handling.""" + + # Setup environment variables + setup_environment() + + # Setup signal handlers + setup_signal_handlers() + + # Convert config dicts back to objects + logger.info("Initializing configuration objects") + provider_config_obj = HuggingFacePostTrainingConfig(**provider_config) + config_obj = TrainingConfig(**config) + dpo_config_obj = DPOAlignmentConfig(**dpo_config) + + # Initialize and validate device + device = setup_torch_device(provider_config_obj.device) + logger.info(f"Using device '{device}'") + + # Load dataset and tokenizer + train_dataset, eval_dataset, tokenizer = await self.load_dataset(model, config_obj, provider_config_obj) + + # Calculate steps per epoch + if not config_obj.data_config: + raise ValueError("DataConfig is required for training") + steps_per_epoch = len(train_dataset) // config_obj.data_config.batch_size + + # Setup training arguments + training_args = self.setup_training_args( + config_obj, + provider_config_obj, + dpo_config_obj, + device, + output_dir_path, + steps_per_epoch, + ) + + # Load model and reference model + model_obj = load_model(model, device, provider_config_obj) + ref_model = None + if provider_config_obj.use_reference_model: + logger.info("Loading separate reference model for DPO") + ref_model = load_model(model, device, provider_config_obj) + else: + logger.info("Using shared reference model for DPO") + + # Initialize DPO trainer + logger.info("Initializing DPOTrainer") + trainer = DPOTrainer( + model=model_obj, + ref_model=ref_model, + args=training_args, + train_dataset=train_dataset, + eval_dataset=eval_dataset, + processing_class=tokenizer, + ) + + try: + # Train + logger.info("Starting DPO training") + trainer.train() + logger.info("DPO training completed successfully") + + # Save final model if output directory is provided + if output_dir_path: + logger.info(f"Saving model to output directory: {output_dir_path}") + self.save_model(trainer, output_dir_path) + logger.info("Model save completed") + + finally: + # Clean up resources + logger.info("Cleaning up resources") + if hasattr(trainer, "model"): + evacuate_model_from_device(trainer.model, device.type) + if ref_model: + evacuate_model_from_device(ref_model, device.type) + del trainer + del ref_model + gc.collect() + logger.info("Cleanup completed") + logger.info("DPO training process finishing successfully") + + async def train( + self, + model: str, + output_dir: str | None, + job_uuid: str, + dpo_config: DPOAlignmentConfig, + config: TrainingConfig, + provider_config: HuggingFacePostTrainingConfig, + ) -> tuple[dict[str, Any], list[Checkpoint] | None]: + """Train a model using HuggingFace's DPOTrainer""" + # Initialize and validate device + device = setup_torch_device(provider_config.device) + logger.info(f"Using device '{device}'") + + output_dir_path = None + if output_dir: + output_dir_path = Path(output_dir) + + # Track memory stats + memory_stats = { + "initial": get_memory_stats(device), + "after_training": None, + "final": None, + } + + # Validate data config + if not config.data_config: + raise ValueError("DataConfig is required for training") + + # Train in a separate process + logger.info("Starting DPO training in separate process") + try: + # Setup multiprocessing for device + if device.type in ["cuda", "mps"]: + multiprocessing.set_start_method("spawn", force=True) + + process = multiprocessing.Process( + target=self._run_training_sync, + kwargs={ + "model": model, + "provider_config": provider_config.model_dump(), + "dpo_config": dpo_config.model_dump(), + "config": config.model_dump(), + "output_dir_path": output_dir_path, + }, + ) + process.start() + + # Monitor the process + while process.is_alive(): + process.join(timeout=1) # Check every second + if not process.is_alive(): + break + + # Get the return code + if process.exitcode != 0: + raise RuntimeError(f"DPO training failed with exit code {process.exitcode}") + + memory_stats["after_training"] = get_memory_stats(device) + + checkpoints = [] + if output_dir_path: + checkpoints = create_checkpoints(output_dir_path, job_uuid, model, config, "dpo_model") + + return memory_stats, checkpoints if checkpoints else None + finally: + memory_stats["final"] = get_memory_stats(device) + gc.collect() diff --git a/llama_stack/providers/inline/post_training/huggingface/utils.py b/llama_stack/providers/inline/post_training/huggingface/utils.py new file mode 100644 index 000000000..3147c19ab --- /dev/null +++ b/llama_stack/providers/inline/post_training/huggingface/utils.py @@ -0,0 +1,269 @@ +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +import logging +import os +import signal +import sys +from datetime import UTC, datetime +from pathlib import Path +from typing import Any + +import psutil +import torch +from datasets import Dataset +from transformers import AutoConfig, AutoModelForCausalLM + +from llama_stack.apis.datasetio import DatasetIO +from llama_stack.apis.post_training import Checkpoint, TrainingConfig + +from .config import HuggingFacePostTrainingConfig + +logger = logging.getLogger(__name__) + + +def setup_environment(): + """Setup common environment variables for training.""" + os.environ["TOKENIZERS_PARALLELISM"] = "false" + os.environ["MKL_THREADING_LAYER"] = "GNU" + os.environ["MKL_SERVICE_FORCE_INTEL"] = "0" + os.environ["MKL_NUM_THREADS"] = "1" + + +def bytes_to_gb(to_convert: int) -> str: + """Converts memory stats to GB and formats to 2 decimal places. + Args: + to_convert: Memory value in bytes + Returns: + str: Memory value in GB formatted to 2 decimal places + """ + return f"{(to_convert / (1024**3)):.2f}" + + +def get_memory_stats(device: torch.device) -> dict[str, Any]: + """Get memory statistics for the given device.""" + stats = { + "system_memory": { + "total": bytes_to_gb(psutil.virtual_memory().total), + "available": bytes_to_gb(psutil.virtual_memory().available), + "used": bytes_to_gb(psutil.virtual_memory().used), + "percent": psutil.virtual_memory().percent, + } + } + + if device.type == "cuda": + stats["device_memory"] = { + "allocated": bytes_to_gb(torch.cuda.memory_allocated(device)), + "reserved": bytes_to_gb(torch.cuda.memory_reserved(device)), + "max_allocated": bytes_to_gb(torch.cuda.max_memory_allocated(device)), + } + elif device.type == "mps": + # MPS doesn't provide direct memory stats, but we can track system memory + stats["device_memory"] = { + "note": "MPS memory stats not directly available", + "system_memory_used": bytes_to_gb(psutil.virtual_memory().used), + } + elif device.type == "cpu": + # For CPU, we track process memory usage + process = psutil.Process() + stats["device_memory"] = { + "process_rss": bytes_to_gb(process.memory_info().rss), + "process_vms": bytes_to_gb(process.memory_info().vms), + "process_percent": process.memory_percent(), + } + + return stats + + +def setup_torch_device(device_str: str) -> torch.device: + """Initialize and validate a PyTorch device. + This function handles device initialization and validation for different device types: + - CUDA: Validates CUDA availability and handles device selection + - MPS: Validates MPS availability for Apple Silicon + - CPU: Basic validation + - HPU: Raises error as it's not supported + Args: + device_str: String specifying the device ('cuda', 'cpu', 'mps') + Returns: + torch.device: The initialized and validated device + Raises: + RuntimeError: If device initialization fails or device is not supported + """ + try: + device = torch.device(device_str) + except RuntimeError as e: + raise RuntimeError(f"Error getting Torch Device {str(e)}") from e + + # Validate device capabilities + if device.type == "cuda": + if not torch.cuda.is_available(): + raise RuntimeError( + f"{device.type}: Torch has no CUDA/ROCm support or could not detect a compatible device." + ) + if device.index is None: + device = torch.device(device.type, torch.cuda.current_device()) + elif device.type == "mps": + if not torch.backends.mps.is_available(): + raise RuntimeError(f"{device.type}: Torch has no MPS support or could not detect a compatible device.") + elif device.type == "hpu": + raise RuntimeError(f"{device.type}: training does not support Intel Gaudi.") + + return device + + +async def load_rows_from_dataset(datasetio_api: DatasetIO, dataset_id: str) -> list[dict[str, Any]]: + """Load dataset from llama stack dataset provider""" + try: + all_rows = await datasetio_api.iterrows( + dataset_id=dataset_id, + limit=-1, + ) + if not isinstance(all_rows.data, list): + raise RuntimeError("Expected dataset data to be a list") + return all_rows.data + except Exception as e: + raise RuntimeError(f"Failed to load dataset: {str(e)}") from e + + +def load_model( + model: str, + device: torch.device, + provider_config: HuggingFacePostTrainingConfig, +) -> AutoModelForCausalLM: + """Load and initialize the model for training. + Args: + model: The model identifier to load + device: The device to load the model onto + provider_config: Provider-specific configuration + Returns: + The loaded and initialized model + Raises: + RuntimeError: If model loading fails + """ + logger.info("Loading the base model") + try: + model_config = AutoConfig.from_pretrained(model, **provider_config.model_specific_config) + model_obj = AutoModelForCausalLM.from_pretrained( + model, + torch_dtype="auto" if device.type != "cpu" else "float32", + quantization_config=None, + config=model_config, + **provider_config.model_specific_config, + ) + # Always move model to specified device + model_obj = model_obj.to(device) + logger.info(f"Model loaded and moved to device: {model_obj.device}") + return model_obj + except Exception as e: + raise RuntimeError(f"Failed to load model: {str(e)}") from e + + +def split_dataset(ds: Dataset) -> tuple[Dataset, Dataset]: + """Split dataset into train and validation sets. + Args: + ds: Dataset to split + Returns: + tuple: (train_dataset, eval_dataset) + """ + logger.info("Splitting dataset into train and validation sets") + train_val_split = ds.train_test_split(test_size=0.1, seed=42) + train_dataset = train_val_split["train"] + eval_dataset = train_val_split["test"] + logger.info(f"Split dataset into {len(train_dataset)} training and {len(eval_dataset)} validation examples") + return train_dataset, eval_dataset + + +def setup_signal_handlers(): + """Setup signal handlers for graceful shutdown.""" + + def signal_handler(signum, frame): + logger.info(f"Received signal {signum}, initiating graceful shutdown") + sys.exit(0) + + signal.signal(signal.SIGTERM, signal_handler) + signal.signal(signal.SIGINT, signal_handler) + + +def calculate_training_steps(steps_per_epoch: int, config: TrainingConfig) -> dict[str, int]: + """Calculate training steps and logging configuration. + Args: + steps_per_epoch: Number of training steps per epoch + config: Training configuration + Returns: + dict: Dictionary with calculated step values + """ + total_steps = steps_per_epoch * config.n_epochs + max_steps = min(config.max_steps_per_epoch, total_steps) + logging_steps = max(1, steps_per_epoch // 50) # Log 50 times per epoch + + logger.info("Training configuration:") + logger.info(f"- Steps per epoch: {steps_per_epoch}") + logger.info(f"- Total steps: {total_steps}") + logger.info(f"- Max steps: {max_steps}") + logger.info(f"- Logging steps: {logging_steps}") + + return {"total_steps": total_steps, "max_steps": max_steps, "logging_steps": logging_steps} + + +def get_save_strategy(output_dir_path: Path | None) -> tuple[str, str]: + """Get save and evaluation strategy based on output directory. + Args: + output_dir_path: Optional path to save the model + Returns: + tuple: (save_strategy, eval_strategy) + """ + if output_dir_path: + logger.info(f"Will save checkpoints to {output_dir_path}") + return "epoch", "epoch" + return "no", "no" + + +def create_checkpoints( + output_dir_path: Path, job_uuid: str, model: str, config: TrainingConfig, final_model_name: str +) -> list[Checkpoint]: + """Create checkpoint objects from training output. + Args: + output_dir_path: Path to the training output directory + job_uuid: Unique identifier for the training job + model: Model identifier + config: Training configuration + final_model_name: Name of the final model directory ("merged_model" for SFT, "dpo_model" for DPO) + Returns: + List of Checkpoint objects + """ + checkpoints = [] + + # Add checkpoint directories + checkpoint_dirs = sorted( + [d for d in output_dir_path.glob("checkpoint-*") if d.is_dir()], + key=lambda x: int(x.name.split("-")[1]), + ) + + for epoch_number, checkpoint_dir in enumerate(checkpoint_dirs, start=1): + created_time = datetime.fromtimestamp(os.path.getctime(checkpoint_dir), tz=UTC) + checkpoint = Checkpoint( + identifier=checkpoint_dir.name, + created_at=created_time, + epoch=epoch_number, + post_training_job_id=job_uuid, + path=str(checkpoint_dir), + ) + checkpoints.append(checkpoint) + + # Add final model + final_model_path = output_dir_path / final_model_name + if final_model_path.exists(): + training_type = "sft" if final_model_name == "merged_model" else "dpo" + checkpoint = Checkpoint( + identifier=f"{model}-{training_type}-{config.n_epochs}", + created_at=datetime.now(UTC), + epoch=config.n_epochs, + post_training_job_id=job_uuid, + path=str(final_model_path), + ) + checkpoints.append(checkpoint) + + return checkpoints diff --git a/tests/integration/post_training/test_post_training.py b/tests/integration/post_training/test_post_training.py index 93ca4c32d..05f8717d6 100644 --- a/tests/integration/post_training/test_post_training.py +++ b/tests/integration/post_training/test_post_training.py @@ -13,6 +13,9 @@ import pytest from llama_stack.apis.post_training import ( DataConfig, + DatasetFormat, + DPOAlignmentConfig, + DPOLossType, LoraFinetuningConfig, TrainingConfig, ) @@ -51,6 +54,7 @@ sys.stdout.reconfigure(line_buffering=True) # +# SFT test class TestPostTraining: @pytest.mark.integration @pytest.mark.parametrize( @@ -90,7 +94,7 @@ class TestPostTraining: dataset_id=dataset.identifier, batch_size=1, shuffle=False, - data_format="instruct", + data_format=DatasetFormat.instruct, ) # setup training config with minimal settings @@ -132,6 +136,8 @@ class TestPostTraining: artifacts = llama_stack_client.post_training.job.artifacts(job_uuid=job_uuid) logger.info(f"Job artifacts: {artifacts}") + logger.info(f"Registered dataset with ID: {dataset.identifier}") + # TODO: Fix these tests to properly represent the Jobs API in training # # async def test_get_training_jobs(self, post_training_stack): @@ -159,3 +165,78 @@ class TestPostTraining: # assert job_artifacts.checkpoints[0].identifier == "instructlab/granite-7b-lab" # assert job_artifacts.checkpoints[0].epoch == 0 # assert "/.llama/checkpoints/Llama3.2-3B-Instruct-sft-0" in job_artifacts.checkpoints[0].path + + # DPO test + @pytest.mark.integration + @pytest.mark.parametrize( + "purpose, source", + [ + ( + "post-training/messages", + { + "type": "uri", + "uri": "huggingface://datasets/trl-internal-testing/hh-rlhf-helpful-base-trl-style?split=train[:20]", + }, + ), + ], + ) + @pytest.mark.timeout(360) + def test_preference_optimize(self, llama_stack_client, purpose, source): + logger.info("Starting DPO preference optimization test") + + # register preference dataset to train + dataset = llama_stack_client.datasets.register( + purpose=purpose, + source=source, + ) + logger.info(f"Registered preference dataset with ID: {dataset.identifier}") + + # DPO algorithm configuration + algorithm_config = DPOAlignmentConfig( + beta=0.1, + loss_type=DPOLossType.sigmoid, + ) + + data_config = DataConfig( + dataset_id=dataset.identifier, + batch_size=1, + shuffle=False, + data_format=DatasetFormat.dialog, # DPO datasets often use dialog format + ) + + # setup training config with minimal settings for DPO + training_config = TrainingConfig( + n_epochs=1, + data_config=data_config, + max_steps_per_epoch=1, # Just 2 steps for quick testing + gradient_accumulation_steps=1, + ) + + job_uuid = f"test-dpo-job-{uuid.uuid4()}" + logger.info(f"Starting DPO training job with UUID: {job_uuid}") + + # train with HuggingFace DPO implementation + _ = llama_stack_client.post_training.preference_optimize( + job_uuid=job_uuid, + finetuned_model="distilgpt2", # Much smaller model for faster CI testing + algorithm_config=algorithm_config, + training_config=training_config, + hyperparam_search_config={}, + logger_config={}, + ) + + while True: + status = llama_stack_client.post_training.job.status(job_uuid=job_uuid) + if not status: + logger.error("DPO job not found") + break + + logger.info(f"Current DPO status: {status}") + if status.status == "completed": + break + + logger.info("Waiting for DPO job to complete...") + time.sleep(10) # Increased sleep time to reduce polling frequency + + artifacts = llama_stack_client.post_training.job.artifacts(job_uuid=job_uuid) + logger.info(f"DPO job artifacts: {artifacts}") From 5c33bc1353373e7cf02cc4f45c810ee108104d8a Mon Sep 17 00:00:00 2001 From: Charlie Doern Date: Thu, 31 Jul 2025 11:26:06 -0400 Subject: [PATCH 03/19] fix: post_training ci (#2984) --- tests/integration/post_training/test_post_training.py | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/tests/integration/post_training/test_post_training.py b/tests/integration/post_training/test_post_training.py index 05f8717d6..839b9b1f2 100644 --- a/tests/integration/post_training/test_post_training.py +++ b/tests/integration/post_training/test_post_training.py @@ -194,9 +194,12 @@ class TestPostTraining: # DPO algorithm configuration algorithm_config = DPOAlignmentConfig( beta=0.1, - loss_type=DPOLossType.sigmoid, + loss_type=DPOLossType.sigmoid, # Default loss type + reward_scale=1.0, # Scaling factor for reward signal (neutral scaling) + reward_clip=5.0, # Maximum absolute value for reward clipping (prevents extreme values) + epsilon=1e-8, # Small value for numerical stability + gamma=1.0, ) - data_config = DataConfig( dataset_id=dataset.identifier, batch_size=1, From 3a574ef23cf109186738985b28b582c697eb3ba8 Mon Sep 17 00:00:00 2001 From: Nehanth Narendrula Date: Thu, 31 Jul 2025 12:11:08 -0400 Subject: [PATCH 04/19] fix: remove unused DPO parameters from schema and tests (#2988) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit # What does this PR do? I removed these DPO parameters from the schema in [this PR](https://github.com/meta-llama/llama-stack/pull/2804), but I may not have done it correctly, since they were reintroduced in [this commit](https://github.com/meta-llama/llama-stack/commit/cb7354a9cebad943c6a88b9d2ff4c12a00b36155#diff-4e9a8cb358213d6118c4b6ec2a76d0367af06441bf0717e13a775ade75e2061dR15081)—likely due to a pre-commit hook. I've made the changes again, and the pre-commit hook automatically updated the spec sheet. --- docs/_static/llama-stack-spec.html | 20 ------------------- docs/_static/llama-stack-spec.yaml | 18 ----------------- .../apis/post_training/post_training.py | 8 -------- .../post_training/test_post_training.py | 4 ---- 4 files changed, 50 deletions(-) diff --git a/docs/_static/llama-stack-spec.html b/docs/_static/llama-stack-spec.html index 6a8945bd1..f9af10165 100644 --- a/docs/_static/llama-stack-spec.html +++ b/docs/_static/llama-stack-spec.html @@ -15078,22 +15078,6 @@ "DPOAlignmentConfig": { "type": "object", "properties": { - "reward_scale": { - "type": "number", - "description": "Scaling factor for the reward signal" - }, - "reward_clip": { - "type": "number", - "description": "Maximum absolute value for reward clipping" - }, - "epsilon": { - "type": "number", - "description": "Small value added for numerical stability" - }, - "gamma": { - "type": "number", - "description": "Discount factor for future rewards" - }, "beta": { "type": "number", "description": "Temperature parameter for the DPO loss" @@ -15106,10 +15090,6 @@ }, "additionalProperties": false, "required": [ - "reward_scale", - "reward_clip", - "epsilon", - "gamma", "beta", "loss_type" ], diff --git a/docs/_static/llama-stack-spec.yaml b/docs/_static/llama-stack-spec.yaml index f1bb40dc1..d2c41b2bf 100644 --- a/docs/_static/llama-stack-spec.yaml +++ b/docs/_static/llama-stack-spec.yaml @@ -11163,20 +11163,6 @@ components: DPOAlignmentConfig: type: object properties: - reward_scale: - type: number - description: Scaling factor for the reward signal - reward_clip: - type: number - description: >- - Maximum absolute value for reward clipping - epsilon: - type: number - description: >- - Small value added for numerical stability - gamma: - type: number - description: Discount factor for future rewards beta: type: number description: Temperature parameter for the DPO loss @@ -11186,10 +11172,6 @@ components: description: The type of loss function to use for DPO additionalProperties: false required: - - reward_scale - - reward_clip - - epsilon - - gamma - beta - loss_type title: DPOAlignmentConfig diff --git a/llama_stack/apis/post_training/post_training.py b/llama_stack/apis/post_training/post_training.py index 9170cba51..c16221289 100644 --- a/llama_stack/apis/post_training/post_training.py +++ b/llama_stack/apis/post_training/post_training.py @@ -193,18 +193,10 @@ class DPOLossType(Enum): class DPOAlignmentConfig(BaseModel): """Configuration for Direct Preference Optimization (DPO) alignment. - :param reward_scale: Scaling factor for the reward signal - :param reward_clip: Maximum absolute value for reward clipping - :param epsilon: Small value added for numerical stability - :param gamma: Discount factor for future rewards :param beta: Temperature parameter for the DPO loss :param loss_type: The type of loss function to use for DPO """ - reward_scale: float - reward_clip: float - epsilon: float - gamma: float beta: float loss_type: DPOLossType = DPOLossType.sigmoid diff --git a/tests/integration/post_training/test_post_training.py b/tests/integration/post_training/test_post_training.py index 839b9b1f2..002da1160 100644 --- a/tests/integration/post_training/test_post_training.py +++ b/tests/integration/post_training/test_post_training.py @@ -195,10 +195,6 @@ class TestPostTraining: algorithm_config = DPOAlignmentConfig( beta=0.1, loss_type=DPOLossType.sigmoid, # Default loss type - reward_scale=1.0, # Scaling factor for reward signal (neutral scaling) - reward_clip=5.0, # Maximum absolute value for reward clipping (prevents extreme values) - epsilon=1e-8, # Small value for numerical stability - gamma=1.0, ) data_config = DataConfig( dataset_id=dataset.identifier, From 8a6c0fb93042d995d0ad5b091c52885b1f95fc4b Mon Sep 17 00:00:00 2001 From: Kelly Brown <86735520+kelbrown20@users.noreply.github.com> Date: Thu, 31 Jul 2025 12:21:13 -0400 Subject: [PATCH 05/19] docs: Reformat external provider documentation (#2982) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit **Description** This PR adjusts the external providers documentation to align with the new providers format. Splits up sections into the existing external providers and how to create them as well. Screenshot 2025-07-31 at 9 48 26 AM Open to feedback and adjusting titles --- .../external-providers-guide.md} | 96 ++++++++----------- .../external/external-providers-list.md | 10 ++ docs/source/providers/external/index.md | 13 +++ docs/source/providers/index.md | 2 +- 4 files changed, 65 insertions(+), 56 deletions(-) rename docs/source/providers/{external.md => external/external-providers-guide.md} (85%) create mode 100644 docs/source/providers/external/external-providers-list.md create mode 100644 docs/source/providers/external/index.md diff --git a/docs/source/providers/external.md b/docs/source/providers/external/external-providers-guide.md similarity index 85% rename from docs/source/providers/external.md rename to docs/source/providers/external/external-providers-guide.md index f906890f1..2479d406f 100644 --- a/docs/source/providers/external.md +++ b/docs/source/providers/external/external-providers-guide.md @@ -1,9 +1,4 @@ -# External Providers Guide - -Llama Stack supports external providers that live outside of the main codebase. This allows you to: -- Create and maintain your own providers independently -- Share providers with others without contributing to the main codebase -- Keep provider-specific code separate from the core Llama Stack code +# Creating External Providers ## Configuration @@ -55,17 +50,6 @@ Llama Stack supports two types of external providers: 1. **Remote Providers**: Providers that communicate with external services (e.g., cloud APIs) 2. **Inline Providers**: Providers that run locally within the Llama Stack process -## Known External Providers - -Here's a list of known external providers that you can use with Llama Stack: - -| Name | Description | API | Type | Repository | -|------|-------------|-----|------|------------| -| KubeFlow Training | Train models with KubeFlow | Post Training | Remote | [llama-stack-provider-kft](https://github.com/opendatahub-io/llama-stack-provider-kft) | -| KubeFlow Pipelines | Train models with KubeFlow Pipelines | Post Training | Inline **and** Remote | [llama-stack-provider-kfp-trainer](https://github.com/opendatahub-io/llama-stack-provider-kfp-trainer) | -| RamaLama | Inference models with RamaLama | Inference | Remote | [ramalama-stack](https://github.com/containers/ramalama-stack) | -| TrustyAI LM-Eval | Evaluate models with TrustyAI LM-Eval | Eval | Remote | [llama-stack-provider-lmeval](https://github.com/trustyai-explainability/llama-stack-provider-lmeval) | - ### Remote Provider Specification Remote providers are used when you need to communicate with external services. Here's an example for a custom Ollama provider: @@ -119,9 +103,9 @@ container_image: custom-vector-store:latest # optional - `provider_data_validator`: Optional validator for provider data - `container_image`: Optional container image to use instead of pip packages -## Required Implementation +## Required Fields -## All Providers +### All Providers All providers must contain a `get_provider_spec` function in their `provider` module. This is a standardized structure that Llama Stack expects and is necessary for getting things such as the config class. The `get_provider_spec` method returns a structure identical to the `adapter`. An example function may look like: @@ -146,7 +130,7 @@ def get_provider_spec() -> ProviderSpec: ) ``` -### Remote Providers +#### Remote Providers Remote providers must expose a `get_adapter_impl()` function in their module that takes two arguments: 1. `config`: An instance of the provider's config class @@ -162,7 +146,7 @@ async def get_adapter_impl( return OllamaInferenceAdapter(config) ``` -### Inline Providers +#### Inline Providers Inline providers must expose a `get_provider_impl()` function in their module that takes two arguments: 1. `config`: An instance of the provider's config class @@ -189,7 +173,40 @@ Version: 0.1.0 Location: /path/to/venv/lib/python3.10/site-packages ``` -## Example using `external_providers_dir`: Custom Ollama Provider +## Best Practices + +1. **Package Naming**: Use the prefix `llama-stack-provider-` for your provider packages to make them easily identifiable. + +2. **Version Management**: Keep your provider package versioned and compatible with the Llama Stack version you're using. + +3. **Dependencies**: Only include the minimum required dependencies in your provider package. + +4. **Documentation**: Include clear documentation in your provider package about: + - Installation requirements + - Configuration options + - Usage examples + - Any limitations or known issues + +5. **Testing**: Include tests in your provider package to ensure it works correctly with Llama Stack. +You can refer to the [integration tests +guide](https://github.com/meta-llama/llama-stack/blob/main/tests/integration/README.md) for more +information. Execute the test for the Provider type you are developing. + +## Troubleshooting + +If your external provider isn't being loaded: + +1. Check that `module` points to a published pip package with a top level `provider` module including `get_provider_spec`. +1. Check that the `external_providers_dir` path is correct and accessible. +2. Verify that the YAML files are properly formatted. +3. Ensure all required Python packages are installed. +4. Check the Llama Stack server logs for any error messages - turn on debug logging to get more + information using `LLAMA_STACK_LOGGING=all=debug`. +5. Verify that the provider package is installed in your Python environment if using `external_providers_dir`. + +## Examples + +### Example using `external_providers_dir`: Custom Ollama Provider Here's a complete example of creating and using a custom Ollama provider: @@ -241,7 +258,7 @@ external_providers_dir: ~/.llama/providers.d/ The provider will now be available in Llama Stack with the type `remote::custom_ollama`. -## Example using `module`: ramalama-stack +### Example using `module`: ramalama-stack [ramalama-stack](https://github.com/containers/ramalama-stack) is a recognized external provider that supports installation via module. @@ -266,35 +283,4 @@ additional_pip_packages: No other steps are required other than `llama stack build` and `llama stack run`. The build process will use `module` to install all of the provider dependencies, retrieve the spec, etc. -The provider will now be available in Llama Stack with the type `remote::ramalama`. - -## Best Practices - -1. **Package Naming**: Use the prefix `llama-stack-provider-` for your provider packages to make them easily identifiable. - -2. **Version Management**: Keep your provider package versioned and compatible with the Llama Stack version you're using. - -3. **Dependencies**: Only include the minimum required dependencies in your provider package. - -4. **Documentation**: Include clear documentation in your provider package about: - - Installation requirements - - Configuration options - - Usage examples - - Any limitations or known issues - -5. **Testing**: Include tests in your provider package to ensure it works correctly with Llama Stack. -You can refer to the [integration tests -guide](https://github.com/meta-llama/llama-stack/blob/main/tests/integration/README.md) for more -information. Execute the test for the Provider type you are developing. - -## Troubleshooting - -If your external provider isn't being loaded: - -1. Check that `module` points to a published pip package with a top level `provider` module including `get_provider_spec`. -1. Check that the `external_providers_dir` path is correct and accessible. -2. Verify that the YAML files are properly formatted. -3. Ensure all required Python packages are installed. -4. Check the Llama Stack server logs for any error messages - turn on debug logging to get more - information using `LLAMA_STACK_LOGGING=all=debug`. -5. Verify that the provider package is installed in your Python environment if using `external_providers_dir`. +The provider will now be available in Llama Stack with the type `remote::ramalama`. \ No newline at end of file diff --git a/docs/source/providers/external/external-providers-list.md b/docs/source/providers/external/external-providers-list.md new file mode 100644 index 000000000..49f49076b --- /dev/null +++ b/docs/source/providers/external/external-providers-list.md @@ -0,0 +1,10 @@ +# Known External Providers + +Here's a list of known external providers that you can use with Llama Stack: + +| Name | Description | API | Type | Repository | +|------|-------------|-----|------|------------| +| KubeFlow Training | Train models with KubeFlow | Post Training | Remote | [llama-stack-provider-kft](https://github.com/opendatahub-io/llama-stack-provider-kft) | +| KubeFlow Pipelines | Train models with KubeFlow Pipelines | Post Training | Inline **and** Remote | [llama-stack-provider-kfp-trainer](https://github.com/opendatahub-io/llama-stack-provider-kfp-trainer) | +| RamaLama | Inference models with RamaLama | Inference | Remote | [ramalama-stack](https://github.com/containers/ramalama-stack) | +| TrustyAI LM-Eval | Evaluate models with TrustyAI LM-Eval | Eval | Remote | [llama-stack-provider-lmeval](https://github.com/trustyai-explainability/llama-stack-provider-lmeval) | \ No newline at end of file diff --git a/docs/source/providers/external/index.md b/docs/source/providers/external/index.md new file mode 100644 index 000000000..989a7f5b8 --- /dev/null +++ b/docs/source/providers/external/index.md @@ -0,0 +1,13 @@ +# External Providers + +Llama Stack supports external providers that live outside of the main codebase. This allows you to: +- Create and maintain your own providers independently +- Share providers with others without contributing to the main codebase +- Keep provider-specific code separate from the core Llama Stack code + +```{toctree} +:maxdepth: 1 + +external-providers-list +external-providers-guide +``` \ No newline at end of file diff --git a/docs/source/providers/index.md b/docs/source/providers/index.md index 97971c232..3f66ecd0c 100644 --- a/docs/source/providers/index.md +++ b/docs/source/providers/index.md @@ -15,7 +15,7 @@ Importantly, Llama Stack always strives to provide at least one fully inline pro ```{toctree} :maxdepth: 1 -external +external/index openai inference/index agents/index From ea8dd581442924c80778260ecaef9b9e5663d446 Mon Sep 17 00:00:00 2001 From: Christian Zaccaria <73656840+ChristianZaccaria@users.noreply.github.com> Date: Thu, 31 Jul 2025 17:21:30 +0100 Subject: [PATCH 06/19] chore: Remove coverage badge from README.md (#2976) # What does this PR do? It looks like the coverage badge is still present in the README. This PR removes it. For more context: https://github.com/meta-llama/llama-stack/pull/2950 --- .github/workflows/README.md | 2 +- README.md | 1 - scripts/gen-ci-docs.py | 2 +- 3 files changed, 2 insertions(+), 3 deletions(-) diff --git a/.github/workflows/README.md b/.github/workflows/README.md index f38008b8a..4ac80e68b 100644 --- a/.github/workflows/README.md +++ b/.github/workflows/README.md @@ -1,6 +1,6 @@ # Llama Stack CI -Llama Stack uses GitHub Actions for Continous Integration (CI). Below is a table detailing what CI the project includes and the purpose. +Llama Stack uses GitHub Actions for Continuous Integration (CI). Below is a table detailing what CI the project includes and the purpose. | Name | File | Purpose | | ---- | ---- | ------- | diff --git a/README.md b/README.md index 7f0fed345..9148ce05d 100644 --- a/README.md +++ b/README.md @@ -6,7 +6,6 @@ [![Discord](https://img.shields.io/discord/1257833999603335178?color=6A7EC2&logo=discord&logoColor=ffffff)](https://discord.gg/llama-stack) [![Unit Tests](https://github.com/meta-llama/llama-stack/actions/workflows/unit-tests.yml/badge.svg?branch=main)](https://github.com/meta-llama/llama-stack/actions/workflows/unit-tests.yml?query=branch%3Amain) [![Integration Tests](https://github.com/meta-llama/llama-stack/actions/workflows/integration-tests.yml/badge.svg?branch=main)](https://github.com/meta-llama/llama-stack/actions/workflows/integration-tests.yml?query=branch%3Amain) -![coverage badge](./coverage.svg) [**Quick Start**](https://llama-stack.readthedocs.io/en/latest/getting_started/index.html) | [**Documentation**](https://llama-stack.readthedocs.io/en/latest/index.html) | [**Colab Notebook**](./docs/getting_started.ipynb) | [**Discord**](https://discord.gg/llama-stack) diff --git a/scripts/gen-ci-docs.py b/scripts/gen-ci-docs.py index 630cfe765..e40e3ed3a 100755 --- a/scripts/gen-ci-docs.py +++ b/scripts/gen-ci-docs.py @@ -36,7 +36,7 @@ def generate_ci_docs(): # Header section to preserve header = """# Llama Stack CI -Llama Stack uses GitHub Actions for Continous Integration (CI). Below is a table detailing what CI the project includes and the purpose. +Llama Stack uses GitHub Actions for Continuous Integration (CI). Below is a table detailing what CI the project includes and the purpose. | Name | File | Purpose | | ---- | ---- | ------- | From ffb6306fbd86ae43665c94a3e55601a5253a867b Mon Sep 17 00:00:00 2001 From: Nathan Weinberg <31703736+nathan-weinberg@users.noreply.github.com> Date: Thu, 31 Jul 2025 12:22:04 -0400 Subject: [PATCH 07/19] fix: remove redundant code from unregister_vector_db (#2983) get_vector_db() will raise an exception if a vector store won't be returned client handling is redundant Signed-off-by: Nathan Weinberg --- llama_stack/core/routing_tables/vector_dbs.py | 2 -- 1 file changed, 2 deletions(-) diff --git a/llama_stack/core/routing_tables/vector_dbs.py b/llama_stack/core/routing_tables/vector_dbs.py index 5ea4e9854..c81a27a3b 100644 --- a/llama_stack/core/routing_tables/vector_dbs.py +++ b/llama_stack/core/routing_tables/vector_dbs.py @@ -84,8 +84,6 @@ class VectorDBsRoutingTable(CommonRoutingTableImpl, VectorDBs): async def unregister_vector_db(self, vector_db_id: str) -> None: existing_vector_db = await self.get_vector_db(vector_db_id) - if existing_vector_db is None: - raise VectorStoreNotFoundError(vector_db_id) await self.unregister_object(existing_vector_db) async def openai_retrieve_vector_store( From b41d696e4ff273329455311c6f0aed141ff2c84c Mon Sep 17 00:00:00 2001 From: Nehanth Narendrula Date: Thu, 31 Jul 2025 14:22:34 -0400 Subject: [PATCH 08/19] fix: Post Training Model change in Tests in order to make it less intensive (#2991) # What does this PR do? Changed from` ibm-granite/granite-3.3-2b-instruct` to` HuggingFaceTB/SmolLM2-135M-Instruct` so it as not resource intensive in CI Idea came from - https://github.com/meta-llama/llama-stack/pull/2984#issuecomment-3140400830 --- tests/integration/post_training/test_post_training.py | 3 +-- 1 file changed, 1 insertion(+), 2 deletions(-) diff --git a/tests/integration/post_training/test_post_training.py b/tests/integration/post_training/test_post_training.py index 002da1160..f9c797593 100644 --- a/tests/integration/post_training/test_post_training.py +++ b/tests/integration/post_training/test_post_training.py @@ -70,7 +70,6 @@ class TestPostTraining: ], ) @pytest.mark.timeout(360) # 6 minutes timeout - @skip_because_resource_intensive def test_supervised_fine_tune(self, llama_stack_client, purpose, source): logger.info("Starting supervised fine-tuning test") @@ -111,7 +110,7 @@ class TestPostTraining: # train with HF trl SFTTrainer as the default _ = llama_stack_client.post_training.supervised_fine_tune( job_uuid=job_uuid, - model="ibm-granite/granite-3.3-2b-instruct", + model="HuggingFaceTB/SmolLM2-135M-Instruct", # smaller model that supports the current sft recipe algorithm_config=algorithm_config, training_config=training_config, hyperparam_search_config={}, From 709c974bd855d065c8383c128a82f1d94a5f4fa2 Mon Sep 17 00:00:00 2001 From: Charlie Doern Date: Thu, 31 Jul 2025 14:36:44 -0400 Subject: [PATCH 09/19] fix: integration tests not triggering on PR open (#2985) # What does this PR do? I realized that when a new PR is opened, the integration tests aren't triggering (or aren't always?) since the replay logic was introduced amend the concurrency logic a bit to trigger on opened PRs --------- Signed-off-by: Charlie Doern Co-authored-by: Ashwin Bharambe --- .github/workflows/integration-tests.yml | 8 +++----- 1 file changed, 3 insertions(+), 5 deletions(-) diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index 472904bb2..d30ccbe71 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -40,11 +40,9 @@ concurrency: # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. group: >- ${{ github.workflow }}-${{ github.ref }}-${{ - (github.event.action == 'opened' || - github.event.action == 'synchronize' || - (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests'))) && - (contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'replay') || - 'no-run' + ((github.event.action == 'opened' || github.event.action == 'synchronize') && 'replay') || + ((github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) && 'rerecord' || + 'no-run') }} cancel-in-progress: true From 27d866795c6849d60953e9d8b1ff92420028bdf7 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 11:50:42 -0700 Subject: [PATCH 10/19] feat(ci): add support for running vision inference tests (#2972) This PR significantly refactors the Integration Tests workflow. The main goal behind the PR was to enable recording of vision tests which were never run as part of our CI ever before. During debugging, I ended up making several other changes refactoring and hopefully increasing the robustness of the workflow. After doing the experiments, I have updated the trigger event to be `pull_request_target` so this workflow can get write permissions by default but it will run with source code from the base (main) branch in the source repository only. If you do change the workflow, you'd need to experiment using the `workflow_dispatch` triggers. This should not be news to anyone using Github Actions (except me!) It is likely to be a little rocky though while I learn more about GitHub Actions, etc. Please be patient :) --------- Co-authored-by: github-actions[bot] --- .../actions/run-and-record-tests/action.yml | 198 + .../actions/run-integration-tests/action.yml | 73 - .github/actions/setup-ollama/action.yml | 14 +- .../actions/setup-test-environment/action.yml | 48 + .github/workflows/README.md | 1 + .github/workflows/integration-tests.yml | 113 +- .../workflows/integration-vision-tests.yml | 141 + .../ollama-with-models.containerfile} | 3 +- .../ollama-with-vision-model.containerfile | 14 + .../inference/test_vision_inference.py | 31 +- tests/integration/recordings/index.sqlite | Bin 40960 -> 45056 bytes .../recordings/responses/00ba04f74a96.json | 10 +- .../recordings/responses/011f70e24ce4.json | 772 +- .../recordings/responses/04172112ffbb.json | 44 +- .../recordings/responses/0b27fd737699.json | 10 +- .../recordings/responses/10eea8c15ddc.json | 10 +- .../recordings/responses/12913f20f6ac.json | 44 +- .../recordings/responses/17253d7cc667.json | 10 +- .../recordings/responses/174458ad71b2.json | 10 +- .../recordings/responses/177ba8517262.json | 34 +- .../recordings/responses/178016edef0e.json | 10 +- .../recordings/responses/197228e26971.json | 10 +- .../recordings/responses/198ef7208389.json | 10 +- .../recordings/responses/1adfaa0e062e.json | 10 +- .../recordings/responses/1b45391880c6.json | 34 +- .../recordings/responses/1b8394f90636.json | 10 +- .../recordings/responses/1b92be674e2a.json | 10 +- .../recordings/responses/1f48f4b2ae33.json | 772 +- .../recordings/responses/211b1562d4e6.json | 10 +- .../recordings/responses/2afe3b38ca01.json | 34 +- .../recordings/responses/2d187a11704c.json | 208 +- .../recordings/responses/31407e035752.json | 108 +- .../recordings/responses/35db283fef1d.json | 12 +- .../recordings/responses/3c3f13cb7794.json | 30 +- .../recordings/responses/4014dd44c15f.json | 12 +- .../recordings/responses/40f524d1934a.json | 30 +- .../recordings/responses/43e106de6736.json | 772 +- .../recordings/responses/44fb9cf5875f.json | 10 +- .../recordings/responses/477f8946bf7d.json | 772 +- .../recordings/responses/48d2fb183a2a.json | 10 +- .../recordings/responses/4a3a4447b16b.json | 56 +- .../recordings/responses/50340cd4d253.json | 10 +- .../recordings/responses/545d86510a80.json | 34 +- .../recordings/responses/554de3cd986f.json | 46 +- .../recordings/responses/56ac6a7c6df0.json | 772 +- .../recordings/responses/61be36ad8ccd.json | 772 +- .../recordings/responses/6906a6e71988.json | 10 +- .../recordings/responses/6cc063bbd7d3.json | 48 +- .../recordings/responses/6d35c91287e2.json | 34 +- .../recordings/responses/6fbea1abca7c.json | 46 +- .../recordings/responses/70adef2c30c4.json | 10 +- .../recordings/responses/7354ec181984.json | 10 +- .../recordings/responses/75d0dd9d0fa3.json | 10 +- .../recordings/responses/76b89a84cd6f.json | 772 +- .../recordings/responses/7b4815aba6c5.json | 46 +- .../recordings/responses/7e6806cba34a.json | 34 +- .../recordings/responses/80e4404d8987.json | 28 +- .../recordings/responses/80f09f27dd61.json | 12 +- .../recordings/responses/836f51dfb3c5.json | 10 +- .../recordings/responses/83c2ffb72daa.json | 772 +- .../recordings/responses/840fbb380b73.json | 10 +- .../recordings/responses/84cab42e1f5c.json | 1176 +-- .../recordings/responses/85594a69d74a.json | 10 +- .../recordings/responses/8bba71367e87.json | 7858 ++++------------- .../recordings/responses/90fec951fdb9.json | 772 +- .../recordings/responses/97d3812bfccb.json | 10 +- .../recordings/responses/97e259c0d3e5.json | 46 +- .../recordings/responses/9b812cbcb88d.json | 10 +- .../recordings/responses/9c140a29ae09.json | 34 +- .../recordings/responses/9c4bc9c3e7ac.json | 772 +- .../recordings/responses/9e7a83d3d596.json | 12 +- .../recordings/responses/9fadf5a3d68f.json | 10 +- .../recordings/responses/a410d4840402.json | 772 +- .../recordings/responses/a59d0d7c1485.json | 10 +- .../recordings/responses/a6810c23eda8.json | 94 +- .../recordings/responses/a97477559b10.json | 772 +- .../recordings/responses/ae1c22f18ecc.json | 10 +- .../recordings/responses/ae6835cfe70e.json | 10 +- .../recordings/responses/afb33182f365.json | 12 +- .../recordings/responses/b14ff438ca99.json | 10 +- .../recordings/responses/b24590574a85.json | 442 +- .../recordings/responses/b44cc7a7afc8.json | 3076 +++---- .../recordings/responses/b4cda53cd04f.json | 12 +- .../recordings/responses/b91f1fb4aedb.json | 30 +- .../recordings/responses/bbd0637dce16.json | 466 +- .../recordings/responses/c9cba6f3ee38.json | 10 +- .../recordings/responses/d0ac68cbde69.json | 57 +- .../recordings/responses/d4c86ac355fb.json | 10 +- .../recordings/responses/dac7a32e5db9.json | 10 +- .../recordings/responses/dd226d71f844.json | 34 +- .../recordings/responses/dd9e7d5913e9.json | 12 +- .../recordings/responses/e08848bfcd28.json | 772 +- .../recordings/responses/e29300494763.json | 12 +- .../recordings/responses/e96152610712.json | 10 +- .../recordings/responses/ed9e9b34008d.json | 10 +- .../recordings/responses/eee47930e3ae.json | 46 +- .../recordings/responses/eee6a163b837.json | 772 +- .../recordings/responses/ef59cbff54d0.json | 10 +- .../recordings/responses/f477c2fe1332.json | 50 +- .../recordings/responses/fcdef245da95.json | 10 +- .../recordings/responses/fe140befeba4.json | 772 +- tests/integration/recordings/test.txt | 1 + .../recordings/vision/index.sqlite | Bin 0 -> 12288 bytes .../vision/responses/3877ecf1bc62.json | 22 + .../vision/responses/4a3a4447b16b.json | 68 + .../vision/responses/d0ac68cbde69.json | 19 + .../vision/responses/f1592dee71e5.json | 53 + .../vision/responses/ff7db0102b28.json | 2057 +++++ 108 files changed, 13985 insertions(+), 15254 deletions(-) create mode 100644 .github/actions/run-and-record-tests/action.yml delete mode 100644 .github/actions/run-integration-tests/action.yml create mode 100644 .github/actions/setup-test-environment/action.yml create mode 100644 .github/workflows/integration-vision-tests.yml rename tests/{Containerfile => containers/ollama-with-models.containerfile} (80%) create mode 100644 tests/containers/ollama-with-vision-model.containerfile create mode 100644 tests/integration/recordings/test.txt create mode 100644 tests/integration/recordings/vision/index.sqlite create mode 100644 tests/integration/recordings/vision/responses/3877ecf1bc62.json create mode 100644 tests/integration/recordings/vision/responses/4a3a4447b16b.json create mode 100644 tests/integration/recordings/vision/responses/d0ac68cbde69.json create mode 100644 tests/integration/recordings/vision/responses/f1592dee71e5.json create mode 100644 tests/integration/recordings/vision/responses/ff7db0102b28.json diff --git a/.github/actions/run-and-record-tests/action.yml b/.github/actions/run-and-record-tests/action.yml new file mode 100644 index 000000000..469049b6a --- /dev/null +++ b/.github/actions/run-and-record-tests/action.yml @@ -0,0 +1,198 @@ +name: 'Run and Record Tests' +description: 'Run integration tests and handle recording/artifact upload' + +inputs: + test-types: + description: 'JSON array of test types to run' + required: true + stack-config: + description: 'Stack configuration to use' + required: true + provider: + description: 'Provider to use for tests' + required: true + inference-mode: + description: 'Inference mode (record or replay)' + required: true + run-vision-tests: + description: 'Whether to run vision tests' + required: false + default: 'false' + +runs: + using: 'composite' + steps: + - name: Check Storage and Memory Available Before Tests + if: ${{ always() }} + shell: bash + run: | + free -h + df -h + + - name: Set environment variables + shell: bash + run: | + echo "LLAMA_STACK_CLIENT_TIMEOUT=300" >> $GITHUB_ENV + echo "LLAMA_STACK_TEST_INFERENCE_MODE=${{ inputs.inference-mode }}" >> $GITHUB_ENV + + # Configure provider-specific settings + if [ "${{ inputs.provider }}" == "ollama" ]; then + echo "OLLAMA_URL=http://0.0.0.0:11434" >> $GITHUB_ENV + echo "TEXT_MODEL=ollama/llama3.2:3b-instruct-fp16" >> $GITHUB_ENV + echo "SAFETY_MODEL=ollama/llama-guard3:1b" >> $GITHUB_ENV + else + echo "VLLM_URL=http://localhost:8000/v1" >> $GITHUB_ENV + echo "TEXT_MODEL=vllm/meta-llama/Llama-3.2-1B-Instruct" >> $GITHUB_ENV + fi + + if [ "${{ inputs.run-vision-tests }}" == "true" ]; then + echo "LLAMA_STACK_TEST_RECORDING_DIR=tests/integration/recordings/vision" >> $GITHUB_ENV + else + echo "LLAMA_STACK_TEST_RECORDING_DIR=tests/integration/recordings" >> $GITHUB_ENV + fi + + - name: Run Llama Stack Server + if: ${{ contains(inputs.stack-config, 'server:') }} + shell: bash + run: | + # Run this so pytest in a loop doesn't start-stop servers in a loop + echo "Starting Llama Stack Server" + nohup uv run llama stack run ci-tests --image-type venv > server.log 2>&1 & + + echo "Waiting for Llama Stack Server to start" + for i in {1..30}; do + if curl -s http://localhost:8321/v1/health | grep -q "OK"; then + echo "Llama Stack Server started" + exit 0 + fi + sleep 1 + done + + echo "Llama Stack Server failed to start" + cat server.log + exit 1 + + - name: Run Integration Tests + shell: bash + run: | + stack_config="${{ inputs.stack-config }}" + EXCLUDE_TESTS="builtin_tool or safety_with_image or code_interpreter or test_rag" + + # Configure provider-specific settings + if [ "${{ inputs.provider }}" == "ollama" ]; then + EXTRA_PARAMS="--safety-shield=llama-guard" + else + EXTRA_PARAMS="" + EXCLUDE_TESTS="${EXCLUDE_TESTS} or test_inference_store_tool_calls" + fi + + if [ "${{ inputs.run-vision-tests }}" == "true" ]; then + if uv run pytest -s -v tests/integration/inference/test_vision_inference.py --stack-config=${stack_config} \ + -k "not( ${EXCLUDE_TESTS} )" \ + --vision-model=ollama/llama3.2-vision:11b \ + --embedding-model=sentence-transformers/all-MiniLM-L6-v2 \ + --color=yes ${EXTRA_PARAMS} \ + --capture=tee-sys | tee pytest-${{ inputs.inference-mode }}-vision.log; then + echo "✅ Tests completed for vision" + else + echo "❌ Tests failed for vision" + exit 1 + fi + + exit 0 + fi + + # Run non-vision tests + TEST_TYPES='${{ inputs.test-types }}' + echo "Test types to run: $TEST_TYPES" + + # Collect all test files for the specified test types + TEST_FILES="" + for test_type in $(echo "$TEST_TYPES" | jq -r '.[]'); do + # if provider is vllm, exclude the following tests: (safety, post_training, tool_runtime) + if [ "${{ inputs.provider }}" == "vllm" ]; then + if [ "$test_type" == "safety" ] || [ "$test_type" == "post_training" ] || [ "$test_type" == "tool_runtime" ]; then + echo "Skipping $test_type for vllm provider" + continue + fi + fi + + if [ -d "tests/integration/$test_type" ]; then + # Find all Python test files in this directory + test_files=$(find tests/integration/$test_type -name "test_*.py" -o -name "*_test.py") + if [ -n "$test_files" ]; then + TEST_FILES="$TEST_FILES $test_files" + echo "Added test files from $test_type: $(echo $test_files | wc -w) files" + fi + else + echo "Warning: Directory tests/integration/$test_type does not exist" + fi + done + + if [ -z "$TEST_FILES" ]; then + echo "No test files found for the specified test types" + exit 1 + fi + + echo "=== Running all collected tests in a single pytest command ===" + echo "Total test files: $(echo $TEST_FILES | wc -w)" + + if uv run pytest -s -v $TEST_FILES --stack-config=${stack_config} \ + -k "not( ${EXCLUDE_TESTS} )" \ + --text-model=$TEXT_MODEL \ + --embedding-model=sentence-transformers/all-MiniLM-L6-v2 \ + --color=yes ${EXTRA_PARAMS} \ + --capture=tee-sys | tee pytest-${{ inputs.inference-mode }}-all.log; then + echo "✅ All tests completed successfully" + else + echo "❌ Tests failed" + exit 1 + fi + + - name: Check Storage and Memory Available After Tests + if: ${{ always() }} + shell: bash + run: | + free -h + df -h + + - name: Commit and push recordings + if: ${{ inputs.inference-mode == 'record' }} + shell: bash + run: | + echo "Checking for recording changes" + git status --porcelain tests/integration/recordings/ + + if [[ -n $(git status --porcelain tests/integration/recordings/) ]]; then + echo "New recordings detected, committing and pushing" + git add tests/integration/recordings/ + + if [ "${{ inputs.run-vision-tests }}" == "true" ]; then + git commit -m "Recordings update from CI (vision)" + else + git commit -m "Recordings update from CI" + fi + + git fetch origin ${{ github.event.pull_request.head.ref }} + git rebase origin/${{ github.event.pull_request.head.ref }} + echo "Rebased successfully" + git push origin HEAD:${{ github.event.pull_request.head.ref }} + echo "Pushed successfully" + else + echo "No recording changes" + fi + + - name: Write inference logs to file + if: ${{ always() }} + shell: bash + run: | + sudo docker logs ollama > ollama-${{ inputs.inference-mode }}.log || true + + - name: Upload logs + if: ${{ always() }} + uses: actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 # v4.6.2 + with: + name: ${{ inputs.inference-mode }}-logs-${{ github.run_id }}-${{ github.run_attempt || '' }}-${{ inputs.provider }} + path: | + *.log + retention-days: 1 diff --git a/.github/actions/run-integration-tests/action.yml b/.github/actions/run-integration-tests/action.yml deleted file mode 100644 index e2db846e4..000000000 --- a/.github/actions/run-integration-tests/action.yml +++ /dev/null @@ -1,73 +0,0 @@ -name: 'Run Integration Tests' -description: 'Run integration tests with configurable execution mode and provider settings' - -inputs: - test-types: - description: 'Test types to run (JSON array)' - required: true - stack-config: - description: 'Stack configuration: "ci-tests" or "server:ci-tests"' - required: true - provider: - description: 'Provider to use: "ollama" or "vllm"' - required: true - inference-mode: - description: 'Inference mode: "record" or "replay"' - required: true - -outputs: - logs-path: - description: 'Path to generated log files' - value: '*.log' - -runs: - using: 'composite' - steps: - - name: Run Integration Tests - env: - LLAMA_STACK_CLIENT_TIMEOUT: "300" - LLAMA_STACK_TEST_RECORDING_DIR: "tests/integration/recordings" - LLAMA_STACK_TEST_INFERENCE_MODE: ${{ inputs.inference-mode }} - shell: bash - run: | - stack_config="${{ inputs.stack-config }}" - EXCLUDE_TESTS="builtin_tool or safety_with_image or code_interpreter or test_rag" - - # Configure provider-specific settings - if [ "${{ inputs.provider }}" == "ollama" ]; then - export OLLAMA_URL="http://0.0.0.0:11434" - export TEXT_MODEL="ollama/llama3.2:3b-instruct-fp16" - export SAFETY_MODEL="ollama/llama-guard3:1b" - EXTRA_PARAMS="--safety-shield=llama-guard" - else - export VLLM_URL="http://localhost:8000/v1" - export TEXT_MODEL="vllm/meta-llama/Llama-3.2-1B-Instruct" - EXTRA_PARAMS="" - EXCLUDE_TESTS="${EXCLUDE_TESTS} or test_inference_store_tool_calls" - fi - - TEST_TYPES='${{ inputs.test-types }}' - echo "Test types to run: $TEST_TYPES" - - for test_type in $(echo "$TEST_TYPES" | jq -r '.[]'); do - # if provider is vllm, exclude the following tests: (safety, post_training, tool_runtime) - if [ "${{ inputs.provider }}" == "vllm" ]; then - if [ "$test_type" == "safety" ] || [ "$test_type" == "post_training" ] || [ "$test_type" == "tool_runtime" ]; then - continue - fi - fi - - echo "=== Running tests for: $test_type ===" - - if uv run pytest -s -v tests/integration/$test_type --stack-config=${stack_config} \ - -k "not( ${EXCLUDE_TESTS} )" \ - --text-model=$TEXT_MODEL \ - --embedding-model=sentence-transformers/all-MiniLM-L6-v2 \ - --color=yes ${EXTRA_PARAMS} \ - --capture=tee-sys | tee pytest-${{ inputs.inference-mode }}-$test_type.log; then - echo "✅ Tests completed for $test_type" - else - echo "❌ Tests failed for $test_type" - exit 1 - fi - done diff --git a/.github/actions/setup-ollama/action.yml b/.github/actions/setup-ollama/action.yml index bb08520f1..e57876cb0 100644 --- a/.github/actions/setup-ollama/action.yml +++ b/.github/actions/setup-ollama/action.yml @@ -1,11 +1,23 @@ name: Setup Ollama description: Start Ollama +inputs: + run-vision-tests: + description: 'Run vision tests: "true" or "false"' + required: false + default: 'false' runs: using: "composite" steps: - name: Start Ollama shell: bash run: | - docker run -d --name ollama -p 11434:11434 docker.io/leseb/ollama-with-models + if [ "${{ inputs.run-vision-tests }}" == "true" ]; then + image="ollama-with-vision-model" + else + image="ollama-with-models" + fi + + echo "Starting Ollama with image: $image" + docker run -d --name ollama -p 11434:11434 docker.io/llamastack/$image echo "Verifying Ollama status..." timeout 30 bash -c 'while ! curl -s -L http://127.0.0.1:11434; do sleep 1 && echo "."; done' diff --git a/.github/actions/setup-test-environment/action.yml b/.github/actions/setup-test-environment/action.yml new file mode 100644 index 000000000..7296e9bfd --- /dev/null +++ b/.github/actions/setup-test-environment/action.yml @@ -0,0 +1,48 @@ +name: 'Setup Test Environment' +description: 'Common setup steps for integration tests including dependencies, providers, and build' + +inputs: + python-version: + description: 'Python version to use' + required: true + client-version: + description: 'Client version (latest or published)' + required: true + provider: + description: 'Provider to setup (ollama or vllm)' + required: true + default: 'ollama' + run-vision-tests: + description: 'Whether to setup provider for vision tests' + required: false + default: 'false' + +runs: + using: 'composite' + steps: + - name: Install dependencies + uses: ./.github/actions/setup-runner + with: + python-version: ${{ inputs.python-version }} + client-version: ${{ inputs.client-version }} + + - name: Setup ollama + if: ${{ inputs.provider == 'ollama' }} + uses: ./.github/actions/setup-ollama + with: + run-vision-tests: ${{ inputs.run-vision-tests }} + + - name: Setup vllm + if: ${{ inputs.provider == 'vllm' }} + uses: ./.github/actions/setup-vllm + + - name: Build Llama Stack + shell: bash + run: | + uv run llama stack build --template ci-tests --image-type venv + + - name: Configure git for commits + shell: bash + run: | + git config --local user.email "github-actions[bot]@users.noreply.github.com" + git config --local user.name "github-actions[bot]" diff --git a/.github/workflows/README.md b/.github/workflows/README.md index 4ac80e68b..875b7f99f 100644 --- a/.github/workflows/README.md +++ b/.github/workflows/README.md @@ -10,6 +10,7 @@ Llama Stack uses GitHub Actions for Continuous Integration (CI). Below is a tabl | SqlStore Integration Tests | [integration-sql-store-tests.yml](integration-sql-store-tests.yml) | Run the integration test suite with SqlStore | | Integration Tests | [integration-tests.yml](integration-tests.yml) | Run the integration test suite from tests/integration | | Vector IO Integration Tests | [integration-vector-io-tests.yml](integration-vector-io-tests.yml) | Run the integration test suite with various VectorIO providers | +| Vision Inference Integration Tests | [integration-vision-tests.yml](integration-vision-tests.yml) | Run vision inference integration test suite from tests/integration/inference | | Pre-commit | [pre-commit.yml](pre-commit.yml) | Run pre-commit checks | | Test Llama Stack Build | [providers-build.yml](providers-build.yml) | Test llama stack build | | Python Package Build Test | [python-build-test.yml](python-build-test.yml) | Test building the llama-stack PyPI project | diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index d30ccbe71..d80fbc525 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -5,7 +5,7 @@ run-name: Run the integration test suite from tests/integration on: push: branches: [ main ] - pull_request: + pull_request_target: branches: [ main ] types: [opened, synchronize, labeled] paths: @@ -13,10 +13,10 @@ on: - 'tests/**' - 'uv.lock' - 'pyproject.toml' - - 'requirements.txt' - '.github/workflows/integration-tests.yml' # This workflow - '.github/actions/setup-ollama/action.yml' - - '.github/actions/run-integration-tests/action.yml' + - '.github/actions/setup-test-environment/action.yml' + - '.github/actions/run-and-record-tests/action.yml' schedule: # If changing the cron schedule, update the provider in the test-matrix job - cron: '0 0 * * *' # (test latest client) Daily at 12 AM UTC @@ -31,6 +31,10 @@ on: description: 'Test against a specific provider' type: string default: 'ollama' + force-inference-mode: + description: 'Force inference mode (record or replay)' + type: string + default: '' concurrency: # This creates three concurrency groups: @@ -73,10 +77,16 @@ jobs: - name: Check if re-record-tests label exists id: check-rerecord-tests run: | - if [[ "${{ contains(github.event.pull_request.labels.*.name, 're-record-tests') }}" == "true" ]]; then + if [[ "${{ inputs.force-inference-mode }}" == "record" ]]; then echo "rerecord-tests=true" >> $GITHUB_OUTPUT - else + elif [[ "${{ inputs.force-inference-mode }}" == "replay" ]]; then echo "rerecord-tests=false" >> $GITHUB_OUTPUT + else + if [[ "${{ contains(github.event.pull_request.labels.*.name, 're-record-tests') }}" == "true" ]]; then + echo "rerecord-tests=true" >> $GITHUB_OUTPUT + else + echo "rerecord-tests=false" >> $GITHUB_OUTPUT + fi fi record-tests: @@ -92,65 +102,26 @@ jobs: steps: - name: Checkout repository uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 + with: + ref: ${{ github.event.pull_request.head.ref }} + fetch-depth: 0 - - name: Install dependencies - uses: ./.github/actions/setup-runner + - name: Setup test environment + uses: ./.github/actions/setup-test-environment with: python-version: "3.12" # Use single Python version for recording client-version: "latest" + provider: ${{ inputs.test-provider || 'ollama' }} - - name: Setup ollama - if: ${{ inputs.test-provider == 'ollama' }} - uses: ./.github/actions/setup-ollama - - - name: Setup vllm - if: ${{ inputs.test-provider == 'vllm' }} - uses: ./.github/actions/setup-vllm - - - name: Build Llama Stack - run: | - uv run llama stack build --template ci-tests --image-type venv - - - name: Configure git for commits - run: | - git config --local user.email "github-actions[bot]@users.noreply.github.com" - git config --local user.name "github-actions[bot]" - - - name: Run Integration Tests for All Types (Recording Mode) - uses: ./.github/actions/run-integration-tests + - name: Run and record tests + uses: ./.github/actions/run-and-record-tests with: test-types: ${{ needs.discover-tests.outputs.test-types }} stack-config: 'server:ci-tests' # recording must be done with server since more tests are run - provider: ${{ inputs.test-provider }} + provider: ${{ inputs.test-provider || 'ollama' }} inference-mode: 'record' - - name: Commit and push recordings - run: | - if ! git diff --quiet tests/integration/recordings/; then - echo "Committing recordings" - git add tests/integration/recordings/ - git commit -m "Update recordings" - echo "Pushing all recording commits to PR" - git push origin HEAD:${{ github.head_ref }} - else - echo "No recording changes" - fi - - - name: Write inference logs to file - if: ${{ always() }} - run: | - sudo docker logs ollama > ollama-recording.log || true - - - name: Upload recording logs - if: ${{ always() }} - uses: actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 # v4.6.2 - with: - name: recording-logs-${{ github.run_id }} - path: | - *.log - retention-days: 1 - - run-tests: + run-replay-mode-tests: # Skip this job if we're in recording mode (handled by record-tests job) if: ${{ needs.discover-tests.outputs.rerecord-tests != 'true' }} needs: discover-tests @@ -169,41 +140,17 @@ jobs: - name: Checkout repository uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 - - name: Install dependencies - uses: ./.github/actions/setup-runner + - name: Setup test environment + uses: ./.github/actions/setup-test-environment with: python-version: ${{ matrix.python-version }} client-version: ${{ matrix.client-version }} + provider: ${{ matrix.provider }} - - name: Build Llama Stack - run: | - uv run llama stack build --template ci-tests --image-type venv - - - name: Check Storage and Memory Available Before Tests - if: ${{ always() }} - run: | - free -h - df -h - - - name: Run Integration Tests (Replay Mode) - uses: ./.github/actions/run-integration-tests + - name: Run and record tests + uses: ./.github/actions/run-and-record-tests with: test-types: ${{ needs.discover-tests.outputs.test-types }} stack-config: ${{ matrix.client-type == 'library' && 'ci-tests' || 'server:ci-tests' }} provider: ${{ matrix.provider }} inference-mode: 'replay' - - - name: Check Storage and Memory Available After Tests - if: ${{ always() }} - run: | - free -h - df -h - - - name: Upload test logs on failure - if: ${{ failure() }} - uses: actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 # v4.6.2 - with: - name: test-logs-${{ github.run_id }}-${{ github.run_attempt }}-${{ matrix.provider }}-${{ matrix.client-type }}-${{ matrix.python-version }}-${{ matrix.client-version }} - path: | - *.log - retention-days: 1 diff --git a/.github/workflows/integration-vision-tests.yml b/.github/workflows/integration-vision-tests.yml new file mode 100644 index 000000000..38da333c2 --- /dev/null +++ b/.github/workflows/integration-vision-tests.yml @@ -0,0 +1,141 @@ +name: Vision Inference Integration Tests + +run-name: Run vision inference integration test suite from tests/integration/inference + +on: + push: + branches: [ main ] + pull_request_target: + branches: [ main ] + types: [opened, synchronize, labeled] + paths: + - 'llama_stack/**' + - 'tests/**' + - 'uv.lock' + - 'pyproject.toml' + - '.github/workflows/integration-vision-tests.yml' # This workflow + - '.github/actions/setup-ollama/action.yml' + - '.github/actions/setup-test-environment/action.yml' + - '.github/actions/run-and-record-tests/action.yml' + workflow_dispatch: + inputs: + test-all-client-versions: + description: 'Test against both the latest and published versions' + type: boolean + default: false + force-inference-mode: + description: 'Force inference mode (record or replay)' + type: string + default: '' + +concurrency: + # This creates three concurrency groups: + # ${{ github.workflow }}-${{ github.ref }}-rerecord (for valid triggers with re-record-tests label) + # ${{ github.workflow }}-${{ github.ref }}-replay (for valid triggers without re-record-tests label) + # ${{ github.workflow }}-${{ github.ref }}-no-run (for invalid triggers that will be skipped) + # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. + group: >- + ${{ github.workflow }}-${{ github.ref }}-${{ + ((github.event.action == 'opened' || github.event.action == 'synchronize') && 'replay') || + ((github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) && 'rerecord' || + 'no-run') + }} + cancel-in-progress: true + +jobs: + discover-tests: + if: | + github.event.action == 'opened' || + github.event.action == 'synchronize' || + (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) + runs-on: ubuntu-latest + outputs: + rerecord-tests: ${{ steps.check-rerecord-tests.outputs.rerecord-tests }} + + steps: + - name: Checkout repository + uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 + + - name: Check if re-record-tests label exists + id: check-rerecord-tests + run: | + if [[ "${{ inputs.force-inference-mode }}" == "record" ]]; then + echo "rerecord-tests=true" >> $GITHUB_OUTPUT + elif [[ "${{ inputs.force-inference-mode }}" == "replay" ]]; then + echo "rerecord-tests=false" >> $GITHUB_OUTPUT + else + if [[ "${{ contains(github.event.pull_request.labels.*.name, 're-record-tests') }}" == "true" ]]; then + echo "rerecord-tests=true" >> $GITHUB_OUTPUT + else + echo "rerecord-tests=false" >> $GITHUB_OUTPUT + fi + fi + + record-tests: + # Sequential job for recording to avoid SQLite conflicts + if: ${{ needs.discover-tests.outputs.rerecord-tests == 'true' }} + needs: discover-tests + runs-on: ubuntu-latest + + permissions: + contents: write + pull-requests: write + + steps: + - name: Checkout repository + uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 + with: + ref: ${{ github.event.pull_request.head.ref }} + fetch-depth: 0 + + - name: Setup test environment + uses: ./.github/actions/setup-test-environment + with: + python-version: "3.12" # Use single Python version for recording + client-version: "latest" + provider: 'ollama' + run-vision-tests: 'true' + + - name: Run and record tests + uses: ./.github/actions/run-and-record-tests + with: + test-types: '["vision"]' + stack-config: 'server:ci-tests' # re-recording must be done in server mode + provider: 'ollama' + inference-mode: 'record' + run-vision-tests: 'true' + + run-replay-mode-tests: + # Skip this job if we're in recording mode (handled by record-tests job) + if: ${{ needs.discover-tests.outputs.rerecord-tests != 'true' }} + needs: discover-tests + runs-on: ubuntu-latest + + strategy: + fail-fast: false + matrix: + client-type: [library] + provider: [ollama] + python-version: ["3.12"] + client-version: ["latest"] + + steps: + - name: Checkout repository + uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 + + - name: Setup test environment + uses: ./.github/actions/setup-test-environment + with: + python-version: ${{ matrix.python-version }} + client-version: ${{ matrix.client-version }} + provider: ${{ matrix.provider }} + run-vision-tests: 'true' + + - name: Run and record tests + uses: ./.github/actions/run-and-record-tests + with: + test-types: '["vision"]' + stack-config: ${{ matrix.client-type == 'library' && 'ci-tests' || 'server:ci-tests' }} + provider: ${{ matrix.provider }} + inference-mode: 'replay' + run-vision-tests: 'true' diff --git a/tests/Containerfile b/tests/containers/ollama-with-models.containerfile similarity index 80% rename from tests/Containerfile rename to tests/containers/ollama-with-models.containerfile index d58216810..438fcec38 100644 --- a/tests/Containerfile +++ b/tests/containers/ollama-with-models.containerfile @@ -1,5 +1,6 @@ # Containerfile used to build our all in one ollama image to run tests in CI -# podman build --platform linux/amd64 -f Containerfile -t ollama-with-models . +# +# podman build --platform linux/amd64 -f ./ollama-with-models.containerfile -t ollama-with-models . # FROM --platform=linux/amd64 ollama/ollama:latest diff --git a/tests/containers/ollama-with-vision-model.containerfile b/tests/containers/ollama-with-vision-model.containerfile new file mode 100644 index 000000000..c1fc1a712 --- /dev/null +++ b/tests/containers/ollama-with-vision-model.containerfile @@ -0,0 +1,14 @@ +# Containerfile used to build our Ollama image with vision model to run tests in CI +# +# podman build --platform linux/amd64 -f ./ollama-with-vision-model.containerfile -t ollama-with-vision-model . +# +FROM --platform=linux/amd64 ollama/ollama:latest + +# Start ollama and pull models in a single layer +RUN ollama serve & \ + sleep 5 && \ + ollama pull llama3.2-vision:11b && \ + ollama pull all-minilm:l6-v2 + +# Set the entrypoint to start ollama serve +ENTRYPOINT ["ollama", "serve"] diff --git a/tests/integration/inference/test_vision_inference.py b/tests/integration/inference/test_vision_inference.py index d47dd3d64..7d81f74d1 100644 --- a/tests/integration/inference/test_vision_inference.py +++ b/tests/integration/inference/test_vision_inference.py @@ -25,12 +25,6 @@ def base64_image_data(image_path): return base64.b64encode(image_path.read_bytes()).decode("utf-8") -@pytest.fixture -def base64_image_url(base64_image_data, image_path): - # suffix includes the ., so we remove it - return f"data:image/{image_path.suffix[1:]};base64,{base64_image_data}" - - def test_image_chat_completion_non_streaming(client_with_models, vision_model_id): message = { "role": "user", @@ -78,7 +72,9 @@ def multi_image_data(): def test_image_chat_completion_multiple_images(client_with_models, vision_model_id, multi_image_data, stream): supported_models = ["llama-4", "gpt-4o", "llama4"] if not any(model in vision_model_id.lower() for model in supported_models): - pytest.skip(f"Skip for non-supported model: {vision_model_id}") + pytest.skip( + f"Skip since multi-image tests are only supported for {supported_models}, not for {vision_model_id}" + ) messages = [ { @@ -183,24 +179,13 @@ def test_image_chat_completion_streaming(client_with_models, vision_model_id): assert any(expected in streamed_content for expected in {"dog", "puppy", "pup"}) -@pytest.mark.parametrize("type_", ["url", "data"]) -def test_image_chat_completion_base64(client_with_models, vision_model_id, base64_image_data, base64_image_url, type_): +def test_image_chat_completion_base64(client_with_models, vision_model_id, base64_image_data): image_spec = { - "url": { - "type": "image", - "image": { - "url": { - "uri": base64_image_url, - }, - }, + "type": "image", + "image": { + "data": base64_image_data, }, - "data": { - "type": "image", - "image": { - "data": base64_image_data, - }, - }, - }[type_] + } message = { "role": "user", diff --git a/tests/integration/recordings/index.sqlite b/tests/integration/recordings/index.sqlite index 2ef7a609e247c70a3b5c49e5b0ce68fa2aece882..2e1256d210d5818a229505ec22c6936ea3b4bf1e 100644 GIT binary patch delta 3246 zcmZXWdu$Zf6~<>~A2a(Lvo@HAaTa61;KknE*;&@pEnXgm@JLK3|r(O2j#i3f<3?dtZ)4>-Z!&H20KODCJh z`}+_0_woUoe3xXLN4_h6CqI&J$$#q~mop$2GHChO(|L5|NLSKtx1Rhccvi+AN>VAo z=W{>g0^C11ceCrNt4jJ%njm&M|LJUTJa#nOuiC3^mu$0zUkDz)g+!x%dYcpWjzBsW zXBPd$-XW#)5MIVv9~;x*XhCyg%RcM&rbMQird6 zMy29nqs5sgD~IYt;#9{2_Y$*?qQ%0*+&H|XEn7#?lsQz#W!5=coef%V`)wk!VZWju&$pJf<~!T245?|A()YclvMnSNeYI)4cC^$9rCO-*K;UopDWRM+oG;pw&WvQIUksR_!5@Tev`##ln56%$C8qHk?> zJj5blVzUriXJKKLyfblmo#~yUDeQhF*f3OYjuutOI}K8;>5Xd&8*4H8hOr2(jH$T$ zjJd+%T2v)n5n5MTywND>rXc4>7F~p`@*s{5W6b*m&d;}a>6u9BLY!S?dQ~ zQ*n%6xH!Ash%4;9KiPhZ54pYfxH91c=lqLw*{0hDgcDMyd|Es%2IW#IDc*PfT-@sa z#(&ISGD$mTjS6$AuudP8qH-UF15C;8x06etRL_S$v#WX(hRC(aaMk9M>lM5os&Wb&G!`)rWY4f-QurEW}W>n}Q45CgHf%4VFb(CcR8hf0b`?ZwDupfB&%$Q>o`p_~EvC?oS{7bIa4&@p z+-6rbh)koKPBm_NVqVHAO!5DnTh!Y;e{jORRWy(ZSU#Q98`zuJSNt20fq< zJs_sX@fid?xX(fW$&xC+0LrlsPhY3;l3n9h`RHmAw}O)!{8#*Qeee1L-o2iio)zwo z-Nmkh(gSIOI3UKHN1b*@qy36KX?w@!725gl_>u3^LKYM6hMqv{8mo@Fp*|MR=4HDfd1)`q z({picnT16o#AF;CYhe-Q^}+j;l~E_vJy?}*VbtOCbQkUnG1~Skh&j+W#?%p0h}qCs zZDrKKbsk3p*>0#(XQocCbBN}eEUu}nYrcffhB4Otx8c!vi!QU1)uU&E$*5y8-4;xr zXfoO%S-s6TV`Y&Du}xSx$=s=8#_qMz+})m;oo& z+zqZP(r=_0;+vw&+3NV#QEUI5eXcDi^t!5S9|{83#6Qn{)7xBTd%jD-7su)|N1}ul z&TPz{E;X|W&!%PfubnA;yFW1^A2ah`?@FQdKq9j&hx}1oIM6B`6B zQLG#*x~;l6>3n$g6)RJSc`@nLk<5knMllW*uUU1G4v$VtW)y3~nu8;CFXHAXwg)Y* zkJRnPpE|89OydjTuOt4CfW90Bzce3fMdk>sFT+Px#*VxCQapLYqNCR^pfACVBjysU z(AW`IUtqR}rrjxgkT%;x1zVz$J|F*1o3A}FmDoJA=~hNZTs;L>mz62RDpB{WmC<2D YpNl)g7#)K2AL51eZG#69mF^M#KZXrsjsO4v delta 2535 zcmcJReNYs27{_<__IB@WcM)Hp#wlcUBthBRy*s!Kbv414RJ58b#}r3Jz_&YuLsaDU z&L*)p;8DlW5=9Z;P`v&~Q&DT0GRKtZ4__LzrY6SGvNT6Cr=GnPS>^9;hW&A$@Avom zJilize2og9q-vk#$%h0%IPfq2sY~-EA{h~SmlN(06kkK|HNqD$a@sR|tcG7jCelKe zZgdEL3wMOygf7?0*noty0thmbjNBge@PvX-coLcthwUpKO4aM#S-K4gpGy|E^r z+kjZt29&FicaZw@ix~suOJ7(LzvWvC` z$6Xu6hM4QjY-@)#PblNN_*vX}ZX7wA=pbI=ylug~II2yngn^T*z%u5||0C!<2O!Cs zEh58fwcVSQCB=)$5s~0pFXo2Db!Hz1I3mKmUMv&V)>l0WfldN*Af%6$?J%cO zh5h5lbSI-%5hX1#Lc)d#rXYB-kP$F!qM7MIsD_ghYr6Y-6c3KAt${ZtnVF39QSi@r z6Z85xkqE)sLpC14iU^HXRYOJ;Rtoe78Od_RyIXQx8<;Z zG4!1)3E&RDp`$P^KL~B^&YwU9PXk|3sKsnF~_}vCxe9T%7tL zc{6rDb-h*P*r8bCI|#;Vt)y?#b1fZ0kpGE)mHU$Ox|&@{&UKFajuQJ7`()d3+X!|W zQ)l6*3i3x{5q7*y+-?ijTWHN0DS@Lc;YSx0Q&z$%8Bsvo|8&;*;aD%04HFKSFQA^C zS&yq=XCcO?wYv)HS&3CJ>W6cA#$|IOP()03VntoB4A&A(gRaJK_l1@f*@J9o{6Y`M zs(US@Li_gYp0&pHA;9c-x=-rkK2_x^JZLiak#WrDVd5Uso@JeJuyKc(Nyr>9xXH}i z$Oss_^C=doh6}yeYAD!c_KA4b%i(u3!=mVnl){>3lP}H!!0k3OTuWprZ0*GsLwZZN z`_}8R(_j3F)=q+xTw7g?GvK)F7-w&_C);X;68?LB6!*0)nQdepOqsQ_EqK@(yXiYk zTFRr9^}Z5buO71E@hOt8JX!{;5|*-$zCuu0M`P^9z{NeL*mMn?m^ z&&)i?MnS_94Bybvkq{bZ^2zu%j^=k1l~=V^RD^M|=-%F?KNfH2EK6Mda}S diff --git a/tests/integration/recordings/responses/00ba04f74a96.json b/tests/integration/recordings/responses/00ba04f74a96.json index 9ec4e3ede..c9eb76585 100644 --- a/tests/integration/recordings/responses/00ba04f74a96.json +++ b/tests/integration/recordings/responses/00ba04f74a96.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:56.173703Z", + "created_at": "2025-07-31T17:59:03.590710793Z", "done": true, "done_reason": "stop", - "total_duration": 136508333, - "load_duration": 65819417, + "total_duration": 1043588135, + "load_duration": 41601119, "prompt_eval_count": 216, - "prompt_eval_duration": 58491125, + "prompt_eval_duration": 957017411, "eval_count": 2, - "eval_duration": 11513166, + "eval_duration": 44409088, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/011f70e24ce4.json b/tests/integration/recordings/responses/011f70e24ce4.json index 316deab63..f89cb1ff2 100644 --- a/tests/integration/recordings/responses/011f70e24ce4.json +++ b/tests/integration/recordings/responses/011f70e24ce4.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 36244625, - "load_duration": 29784250, + "total_duration": 14017069, + "load_duration": 6084798, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.062304743, - 0.04315718, - -0.056847535, - 0.03486019, - -0.045148205, - -0.1325256, - 0.021795923, - 0.039035086, - -0.048403695, - -0.03187157, - -0.03934502, - 0.006355416, - 0.07870429, - -0.004275144, - 0.023635335, - -0.02171452, - -0.055756103, - -0.009452624, - 0.03968397, - -0.11446917, - -0.011574315, - 0.06161675, - -0.026243819, - 0.024376081, - 0.029439807, - -0.0035745306, - -0.0014413354, - -0.0031348146, - 0.0137771955, - -0.00021878166, - -0.0148119675, - 0.08438267, - 0.06679146, - 0.042289164, - 0.0077238376, - 0.073178865, - -0.008341517, - -0.094652176, - -0.09245101, - 0.0075944075, - -0.07389992, - 0.015481098, - -0.04405396, - -0.04497366, - -0.041315924, - 0.06968346, - -0.027464444, - 0.014380017, - -0.036109854, - -0.006690219, - -0.080297194, - -5.8296577e-05, - -0.03897778, - -0.049029846, - 0.017797105, - -0.0064906515, - 0.05977029, - -0.0031445406, - -0.024804324, - -0.114971094, - -0.047434244, - 0.018489277, - -0.009801151, - 0.09573786, - -0.009445709, - -0.035714474, - -0.031265706, - -0.0032087746, - 0.07714283, - -0.076175354, - -0.11878057, - -0.06322687, - -0.0045974515, - 0.06524851, - 0.045755487, - -0.13797933, - 0.045973603, - -0.03356543, - -0.013575197, - 0.004536992, - 0.01706251, - -0.0016689816, - -0.051292486, - 0.10251468, - 0.015364908, - -0.05339754, - 0.046751976, - 0.11428272, - -0.0060051866, - 0.010296865, - -0.03160346, - -0.051935352, - 0.02092994, - 0.008887596, - -0.069010794, - 0.08132733, - 0.012102074, - -0.06409327, - -0.036342084, - 0.046690084, - 0.011248327, - -0.050334014, - 0.073782355, - -0.02119414, - 0.0324611, - -0.026148362, - 0.06814877, - -0.03795885, - 0.030811384, - -0.037118603, - -0.036956605, - -0.02943471, - -0.0328876, - -0.00579801, - 0.04255975, - 0.05469473, - -0.01927437, - 0.12277417, - 0.0037985598, - 0.032079652, - 0.023717156, - 0.019211154, - 0.019987307, - -0.012261412, - -0.032464176, - -0.004472998, - -0.03568547, - -6.953471e-33, - -0.02200053, - -0.06861985, - -0.035355665, - 0.008892092, - 0.07110619, - -0.02524488, - 0.091491714, - -0.009333656, - -0.059515916, - -0.03471947, - 0.04331791, - 0.033350475, - 0.02423151, - 0.08795865, - 0.020580785, - -0.00087637454, - -0.012995603, - 0.088356934, - 0.04568453, - 0.025818799, - 0.054319557, - 0.09676607, - 0.02314351, - 0.024316499, - 0.014192086, - -0.01867069, - -0.024500258, - -0.032566376, - 0.025218401, - 0.016804473, - -0.07628905, - 0.012665322, - -0.021314982, - 0.006895667, - 0.030793479, - -0.00033363912, - 0.0005291749, - -0.08589274, - 0.040542576, - 0.0062958263, - -0.009977536, - 0.0016065374, - 0.012649728, - -0.036491103, - -0.023085777, - 0.012404348, - -0.0051287347, - 0.020217113, - -0.08761001, - 0.0451902, - -0.0012827619, - -0.06574815, - 0.07477121, - 0.08403992, - -0.01390955, - 0.05589554, - 0.019330526, - -0.019641383, - -0.016001293, - -0.02915193, - 0.037374426, - 0.068089314, - 0.069200926, - -0.007668733, - 0.021160824, - 0.040417258, - 0.035068225, - 0.082075246, - 0.08809441, - 0.05050193, - -0.059343174, - 0.04576526, - -0.025118835, - 0.03583576, - -0.028081506, - 0.019838363, - 0.033905286, - -0.07977674, - 0.023003135, - 0.062460173, - -0.034886148, - -0.05390937, - -0.016114287, - -0.0057315156, - -0.03051132, - -0.02269694, - -0.010376983, - 0.06762264, - -0.010560655, - -0.09605588, - -0.07854035, - -0.08528194, - 0.029969428, - -0.0059528793, - -0.039581347, - 2.9781768e-33, - 0.011482255, - 0.010417832, - -0.0698601, - 0.019292813, - -0.08453582, - -0.08570265, - 0.06624837, - 0.063025005, - 0.050434116, - 0.033736084, - -0.0058885855, - -0.069622226, - 0.12551048, - 0.021380005, - 0.07413853, - 0.0342258, - -0.045818888, - 0.014834041, - -0.012672501, - 0.0036430089, - -0.08024709, - 0.06730083, - -0.056032285, - -0.086702436, - -0.027874194, - -0.03391202, - -0.03872441, - -0.07792124, - -0.017794719, - 0.061800934, - 0.014696384, - 0.019996569, - -0.08146178, - 0.052340467, - 0.06287676, - -0.0015751559, - 0.040512506, - -0.027605608, - -0.009630798, - -0.017303543, - 0.11392578, - 0.044186074, - 0.035317622, - 0.12113664, - 0.018812222, - 0.049269576, - -0.036081262, - 0.07789768, - -0.0296637, - -0.07068735, - -0.006731622, - 0.0060941395, - 0.042274125, - -0.039680813, - -0.048600707, - -0.03980193, - 0.032409266, - 0.03371183, - -0.092499994, - -0.049876206, - -0.06597403, - -0.042388365, - 0.031259395, - 0.011791109, - -0.04424881, - 0.04685171, - -0.12302249, - -0.034650978, - -0.01387166, - -0.13122807, - 0.1448325, - 0.0056148693, - -0.0031096544, - 0.022904772, - -0.07642485, - 0.016454488, - -0.019540928, - -0.024970472, - -0.068574235, - 0.07073104, - 0.026643677, - -0.035163663, - -0.0015607082, - 0.029314166, - -0.08943546, - -0.022545528, - -0.031130569, - 0.053781237, - 0.007896568, - 0.023091432, - -0.0043701245, - 0.05380369, - 0.01729408, - 0.05636822, - -0.05328019, - -1.3478804e-08, - -0.039678477, - 0.013365443, - 0.036817312, - 0.009736139, - 0.004703614, - 0.06661744, - 0.02291141, - -0.047423527, - -0.04049001, - 0.0068159057, - 0.008662143, - -0.006292634, - -0.045681197, - -0.06387613, - -0.013174571, - 0.11696965, - 0.016895585, - -0.0013498863, - 0.023227682, - 0.022274282, - 0.07852807, - -0.04508963, - -0.009177306, - 0.06640095, - -0.06651727, - -0.015498115, - 0.054094598, - 0.07642527, - 0.0082470365, - -0.12409585, - 0.01265297, - -0.017635401, - -0.020622984, - 0.03250185, - -0.012997484, - 0.022324847, - 0.010529934, - -0.0883164, - 0.021471445, - -0.0029947716, - -0.03183814, - 0.0718419, - 0.010377949, - 0.0035974192, - 0.048932698, - 0.07039089, - -0.03657371, - -0.035186097, - -0.03655875, - -0.07017832, - -0.030322824, - 0.028595895, - -0.019070871, - -0.0025186248, - 0.021279149, - 0.07436103, - -0.114249244, - -0.027311146, - -0.0107884705, - 0.010422842, - -0.022787437, - 0.11515081, - 0.18532182, - -0.026544156 + -0.062299512, + 0.04314291, + -0.056856677, + 0.03487498, + -0.045130543, + -0.13253723, + 0.021801258, + 0.03905167, + -0.048422147, + -0.031866066, + -0.039334282, + 0.0063861525, + 0.078711785, + -0.004295658, + 0.023596749, + -0.021716505, + -0.05573506, + -0.009471944, + 0.039706443, + -0.114432074, + -0.011571138, + 0.061599534, + -0.026234824, + 0.02437703, + 0.029446855, + -0.0035651308, + -0.00145838, + -0.00313903, + 0.0137839755, + -0.00021519467, + -0.014771578, + 0.08437898, + 0.06679487, + 0.042340428, + 0.0076946374, + 0.07313361, + -0.008328885, + -0.09465153, + -0.09245484, + 0.0076101488, + -0.07390885, + 0.015470385, + -0.044050634, + -0.044988655, + -0.041298985, + 0.06967625, + -0.027475385, + 0.01439177, + -0.03610871, + -0.0066690356, + -0.08027576, + -6.320903e-05, + -0.038967375, + -0.04901159, + 0.01780741, + -0.0064625116, + 0.05977013, + -0.003139111, + -0.024790227, + -0.11497569, + -0.04741595, + 0.018494949, + -0.009821916, + 0.09573474, + -0.009432823, + -0.03572252, + -0.031270232, + -0.0032188955, + 0.07713915, + -0.07621238, + -0.11879392, + -0.063214734, + -0.004622067, + 0.06525516, + 0.045760594, + -0.13793096, + 0.045978762, + -0.033560168, + -0.013592423, + 0.0045015467, + 0.01705248, + -0.0016773397, + -0.05126322, + 0.102517396, + 0.015358336, + -0.05337354, + 0.046742212, + 0.11427399, + -0.005986359, + 0.010281509, + -0.031590212, + -0.05193758, + 0.02094042, + 0.00889564, + -0.06902529, + 0.08132795, + 0.012084552, + -0.06408848, + -0.03637125, + 0.04667938, + 0.011233042, + -0.050319683, + 0.073782675, + -0.021215191, + 0.03245006, + -0.026153775, + 0.06814923, + -0.03795168, + 0.030797591, + -0.037129108, + -0.03695134, + -0.029432079, + -0.032888234, + -0.00580058, + 0.04259698, + 0.05470057, + -0.019268109, + 0.12275155, + 0.003795531, + 0.03207379, + 0.02372011, + 0.019182375, + 0.01998619, + -0.012273767, + -0.03248627, + -0.0044953367, + -0.035685856, + -6.953945e-33, + -0.02199191, + -0.0686648, + -0.0353737, + 0.00889737, + 0.07112167, + -0.025211865, + 0.0914874, + -0.009342371, + -0.05954011, + -0.03471374, + 0.043332614, + 0.0333655, + 0.024237446, + 0.08791945, + 0.020623982, + -0.00088081614, + -0.013014688, + 0.088370614, + 0.04570386, + 0.025825853, + 0.05431844, + 0.09674628, + 0.023137445, + 0.024317676, + 0.014196965, + -0.018658916, + -0.02449057, + -0.03254813, + 0.025230253, + 0.0167997, + -0.07629053, + 0.012663858, + -0.02127982, + 0.006900138, + 0.03077926, + -0.00032187518, + 0.0005111945, + -0.085893854, + 0.040517006, + 0.006310925, + -0.009996223, + 0.0015871905, + 0.012663539, + -0.036496088, + -0.02311059, + 0.012365358, + -0.0051299105, + 0.020204524, + -0.08760432, + 0.045186196, + -0.0012780412, + -0.06578143, + 0.07478501, + 0.08405124, + -0.013907717, + 0.055900548, + 0.01933963, + -0.019657157, + -0.016009875, + -0.029160723, + 0.03739787, + 0.06809498, + 0.06920713, + -0.007672135, + 0.021142934, + 0.04040559, + 0.035094846, + 0.08207594, + 0.088103354, + 0.050499115, + -0.05933218, + 0.045776226, + -0.025103334, + 0.03583547, + -0.028066712, + 0.019852906, + 0.033922214, + -0.07975417, + 0.02300144, + 0.062443927, + -0.03490803, + -0.053939816, + -0.01613488, + -0.0057205497, + -0.030501934, + -0.02271051, + -0.010379288, + 0.06760881, + -0.010573027, + -0.09605811, + -0.07852684, + -0.085278705, + 0.029953092, + -0.005949969, + -0.03959023, + 2.979382e-33, + 0.011482047, + 0.010405214, + -0.06986261, + 0.019275747, + -0.08455298, + -0.08570306, + 0.066268414, + 0.06303412, + 0.05044079, + 0.033729207, + -0.005918433, + -0.06963068, + 0.12552938, + 0.021379305, + 0.07415631, + 0.034211684, + -0.045811858, + 0.014828219, + -0.012704339, + 0.0036554744, + -0.080252334, + 0.06730209, + -0.05603338, + -0.08669251, + -0.02789593, + -0.033893805, + -0.03873136, + -0.07794548, + -0.017803997, + 0.061792277, + 0.014711371, + 0.020018095, + -0.08146497, + 0.052354332, + 0.06289804, + -0.0015964498, + 0.040503405, + -0.027576957, + -0.009646813, + -0.017321808, + 0.113927364, + 0.04419595, + 0.035337232, + 0.12111621, + 0.018830014, + 0.049245883, + -0.036052346, + 0.07788832, + -0.02968157, + -0.070657946, + -0.006732323, + 0.0060839457, + 0.042294417, + -0.03963716, + -0.048594773, + -0.039805196, + 0.03239508, + 0.033688314, + -0.092505686, + -0.049885467, + -0.0659565, + -0.04236759, + 0.031238468, + 0.011814915, + -0.044232145, + 0.046881076, + -0.12301668, + -0.03465581, + -0.01388215, + -0.13120441, + 0.14485523, + 0.0056016897, + -0.0030743086, + 0.022897985, + -0.076423965, + 0.016426744, + -0.019541634, + -0.02496784, + -0.06859387, + 0.070740156, + 0.026620118, + -0.0351797, + -0.0015670933, + 0.029303383, + -0.08942909, + -0.022550073, + -0.031130616, + 0.05381134, + 0.007876352, + 0.023096293, + -0.0043927482, + 0.05381174, + 0.017291587, + 0.056370165, + -0.053297367, + -1.3478304e-08, + -0.039681002, + 0.01336931, + 0.03682005, + 0.009732852, + 0.004675352, + 0.06660335, + 0.022932611, + -0.04741615, + -0.04049429, + 0.006841735, + 0.008672197, + -0.0062891566, + -0.045680486, + -0.06389349, + -0.013189537, + 0.11696302, + 0.016887287, + -0.0013747291, + 0.023227474, + 0.02228575, + 0.07854934, + -0.045100793, + -0.009169939, + 0.066385396, + -0.06650943, + -0.015503365, + 0.054116882, + 0.07644889, + 0.008241338, + -0.124083355, + 0.012669299, + -0.017633973, + -0.020603409, + 0.03251493, + -0.013004719, + 0.022333013, + 0.010550418, + -0.08830502, + 0.021466808, + -0.0029931213, + -0.031842466, + 0.071854234, + 0.010362922, + 0.0036116007, + 0.04894235, + 0.070390284, + -0.0365594, + -0.035181943, + -0.03654571, + -0.07017962, + -0.030360749, + 0.028622892, + -0.019087547, + -0.0025200765, + 0.02127114, + 0.07437197, + -0.114239074, + -0.027314458, + -0.010757821, + 0.01041863, + -0.022775937, + 0.1151369, + 0.18533714, + -0.026517315 ] ] } diff --git a/tests/integration/recordings/responses/04172112ffbb.json b/tests/integration/recordings/responses/04172112ffbb.json index 739c74ffc..bf94b0697 100644 --- a/tests/integration/recordings/responses/04172112ffbb.json +++ b/tests/integration/recordings/responses/04172112ffbb.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.453648Z", + "created_at": "2025-07-31T17:59:18.033900164Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.498568Z", + "created_at": "2025-07-31T17:59:18.213371151Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.542162Z", + "created_at": "2025-07-31T17:59:18.387513976Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.583708Z", + "created_at": "2025-07-31T17:59:18.564344287Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.624941Z", + "created_at": "2025-07-31T17:59:18.746579415Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.666128Z", + "created_at": "2025-07-31T17:59:18.923276047Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.707429Z", + "created_at": "2025-07-31T17:59:19.099961963Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.748343Z", + "created_at": "2025-07-31T17:59:19.275621884Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.789526Z", + "created_at": "2025-07-31T17:59:19.452204196Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.830744Z", + "created_at": "2025-07-31T17:59:19.626937514Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,7 +201,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.871855Z", + "created_at": "2025-07-31T17:59:19.805566767Z", "done": false, "done_reason": null, "total_duration": null, @@ -219,7 +219,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.913174Z", + "created_at": "2025-07-31T17:59:19.985987477Z", "done": false, "done_reason": null, "total_duration": null, @@ -237,7 +237,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.954413Z", + "created_at": "2025-07-31T17:59:20.166458601Z", "done": false, "done_reason": null, "total_duration": null, @@ -255,7 +255,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.995741Z", + "created_at": "2025-07-31T17:59:20.343346795Z", "done": false, "done_reason": null, "total_duration": null, @@ -273,7 +273,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:44.036825Z", + "created_at": "2025-07-31T17:59:20.525008091Z", "done": false, "done_reason": null, "total_duration": null, @@ -291,7 +291,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:44.078477Z", + "created_at": "2025-07-31T17:59:20.709087695Z", "done": false, "done_reason": null, "total_duration": null, @@ -309,7 +309,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:44.120217Z", + "created_at": "2025-07-31T17:59:20.887074305Z", "done": false, "done_reason": null, "total_duration": null, @@ -327,15 +327,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:44.162092Z", + "created_at": "2025-07-31T17:59:21.065244925Z", "done": true, "done_reason": "stop", - "total_duration": 875451166, - "load_duration": 90922958, + "total_duration": 4373531496, + "load_duration": 44438132, "prompt_eval_count": 56, - "prompt_eval_duration": 69995084, + "prompt_eval_duration": 1296273199, "eval_count": 18, - "eval_duration": 711865666, + "eval_duration": 3032321735, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/0b27fd737699.json b/tests/integration/recordings/responses/0b27fd737699.json index a1b19c3e4..5ea133b73 100644 --- a/tests/integration/recordings/responses/0b27fd737699.json +++ b/tests/integration/recordings/responses/0b27fd737699.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:06.573878Z", + "created_at": "2025-07-31T17:43:37.656005919Z", "done": true, "done_reason": "stop", - "total_duration": 316204542, - "load_duration": 107614333, + "total_duration": 2903834654, + "load_duration": 43364090, "prompt_eval_count": 223, - "prompt_eval_duration": 194351792, + "prompt_eval_duration": 2812583330, "eval_count": 2, - "eval_duration": 12661500, + "eval_duration": 47252843, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/10eea8c15ddc.json b/tests/integration/recordings/responses/10eea8c15ddc.json index ca55e8e5e..286ea6ed5 100644 --- a/tests/integration/recordings/responses/10eea8c15ddc.json +++ b/tests/integration/recordings/responses/10eea8c15ddc.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:54.937275Z", + "created_at": "2025-07-31T17:58:57.232902036Z", "done": true, "done_reason": "stop", - "total_duration": 215909000, - "load_duration": 100935042, + "total_duration": 1112852961, + "load_duration": 40973130, "prompt_eval_count": 212, - "prompt_eval_duration": 68865625, + "prompt_eval_duration": 898310684, "eval_count": 5, - "eval_duration": 45417959, + "eval_duration": 173032826, "response": "unsafe\nS8", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/12913f20f6ac.json b/tests/integration/recordings/responses/12913f20f6ac.json index 9934da93c..83c1a08c4 100644 --- a/tests/integration/recordings/responses/12913f20f6ac.json +++ b/tests/integration/recordings/responses/12913f20f6ac.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -21,7 +21,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -36,7 +36,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984661, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -47,7 +47,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -62,7 +62,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984662, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -73,11 +73,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { - "content": " word", + "content": " name", "function_call": null, "refusal": null, "role": "assistant", @@ -88,7 +88,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984662, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -99,7 +99,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -114,7 +114,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984662, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -125,7 +125,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -140,7 +140,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984662, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -151,7 +151,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -166,7 +166,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984662, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -177,7 +177,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -192,7 +192,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984663, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -203,7 +203,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -218,7 +218,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984663, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -229,7 +229,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -244,7 +244,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984663, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -255,7 +255,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-471", + "id": "chatcmpl-389", "choices": [ { "delta": { @@ -270,7 +270,7 @@ "logprobs": null } ], - "created": 1753814881, + "created": 1753984663, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/17253d7cc667.json b/tests/integration/recordings/responses/17253d7cc667.json index fc792ca6a..28d20a76a 100644 --- a/tests/integration/recordings/responses/17253d7cc667.json +++ b/tests/integration/recordings/responses/17253d7cc667.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:55.608024Z", + "created_at": "2025-07-31T17:59:00.488769233Z", "done": true, "done_reason": "stop", - "total_duration": 154273583, - "load_duration": 77349416, + "total_duration": 983619113, + "load_duration": 41180204, "prompt_eval_count": 212, - "prompt_eval_duration": 62367583, + "prompt_eval_duration": 898774731, "eval_count": 2, - "eval_duration": 13720542, + "eval_duration": 43157634, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/174458ad71b2.json b/tests/integration/recordings/responses/174458ad71b2.json index 8407188df..109545580 100644 --- a/tests/integration/recordings/responses/174458ad71b2.json +++ b/tests/integration/recordings/responses/174458ad71b2.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:56.012655Z", + "created_at": "2025-07-31T17:59:02.537234214Z", "done": true, "done_reason": "stop", - "total_duration": 161642875, - "load_duration": 91197291, + "total_duration": 1015701193, + "load_duration": 47423661, "prompt_eval_count": 213, - "prompt_eval_duration": 58076125, + "prompt_eval_duration": 923910712, "eval_count": 2, - "eval_duration": 11680042, + "eval_duration": 43825530, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/177ba8517262.json b/tests/integration/recordings/responses/177ba8517262.json index 4be92ecb1..5ea63194a 100644 --- a/tests/integration/recordings/responses/177ba8517262.json +++ b/tests/integration/recordings/responses/177ba8517262.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -40,7 +40,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -55,7 +55,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984048, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -66,7 +66,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -81,7 +81,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984048, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -92,7 +92,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -107,7 +107,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984049, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -118,7 +118,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -133,7 +133,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984049, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -144,7 +144,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -159,7 +159,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984049, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -170,7 +170,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -185,7 +185,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984049, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -196,7 +196,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -211,7 +211,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984049, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -222,7 +222,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-751", + "id": "chatcmpl-943", "choices": [ { "delta": { @@ -237,7 +237,7 @@ "logprobs": null } ], - "created": 1753819484, + "created": 1753984050, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/178016edef0e.json b/tests/integration/recordings/responses/178016edef0e.json index f7f17a85d..eeb807239 100644 --- a/tests/integration/recordings/responses/178016edef0e.json +++ b/tests/integration/recordings/responses/178016edef0e.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:55.17221Z", + "created_at": "2025-07-31T17:58:58.335054803Z", "done": true, "done_reason": "stop", - "total_duration": 202771625, - "load_duration": 106436417, + "total_duration": 1092208348, + "load_duration": 45235490, "prompt_eval_count": 210, - "prompt_eval_duration": 51609667, + "prompt_eval_duration": 872577650, "eval_count": 5, - "eval_duration": 44108208, + "eval_duration": 173862384, "response": "unsafe\nS12", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/197228e26971.json b/tests/integration/recordings/responses/197228e26971.json index 1a06c1200..2045c1e60 100644 --- a/tests/integration/recordings/responses/197228e26971.json +++ b/tests/integration/recordings/responses/197228e26971.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:55.416769Z", + "created_at": "2025-07-31T17:58:59.493522835Z", "done": true, "done_reason": "stop", - "total_duration": 210498750, - "load_duration": 99757250, + "total_duration": 1148787184, + "load_duration": 42525776, "prompt_eval_count": 213, - "prompt_eval_duration": 64811541, + "prompt_eval_duration": 933024346, "eval_count": 5, - "eval_duration": 44342917, + "eval_duration": 172702102, "response": "unsafe\nS2", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/198ef7208389.json b/tests/integration/recordings/responses/198ef7208389.json index 39315c150..1264a1d26 100644 --- a/tests/integration/recordings/responses/198ef7208389.json +++ b/tests/integration/recordings/responses/198ef7208389.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:54.471261Z", + "created_at": "2025-07-31T17:58:55.036336937Z", "done": true, "done_reason": "stop", - "total_duration": 452251417, - "load_duration": 50378583, + "total_duration": 7465058538, + "load_duration": 4385384246, "prompt_eval_count": 212, - "prompt_eval_duration": 357221125, + "prompt_eval_duration": 2899257826, "eval_count": 5, - "eval_duration": 43951958, + "eval_duration": 179706529, "response": "unsafe\nS1", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/1adfaa0e062e.json b/tests/integration/recordings/responses/1adfaa0e062e.json index 19df3ff84..295f101e6 100644 --- a/tests/integration/recordings/responses/1adfaa0e062e.json +++ b/tests/integration/recordings/responses/1adfaa0e062e.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:55.817669Z", + "created_at": "2025-07-31T17:59:01.511604644Z", "done": true, "done_reason": "stop", - "total_duration": 184528709, - "load_duration": 110146792, + "total_duration": 1013469330, + "load_duration": 43788351, "prompt_eval_count": 213, - "prompt_eval_duration": 62390792, + "prompt_eval_duration": 924367637, "eval_count": 2, - "eval_duration": 11431333, + "eval_duration": 44766715, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/1b45391880c6.json b/tests/integration/recordings/responses/1b45391880c6.json index 3b38bd4a6..747d7233d 100644 --- a/tests/integration/recordings/responses/1b45391880c6.json +++ b/tests/integration/recordings/responses/1b45391880c6.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -24,7 +24,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -39,7 +39,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984042, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -50,7 +50,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -65,7 +65,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984042, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -76,7 +76,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -91,7 +91,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984042, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -102,7 +102,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -117,7 +117,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984042, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -128,7 +128,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -143,7 +143,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984042, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -154,7 +154,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -169,7 +169,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984043, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -180,7 +180,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -195,7 +195,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984043, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -206,7 +206,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-434", + "id": "chatcmpl-639", "choices": [ { "delta": { @@ -221,7 +221,7 @@ "logprobs": null } ], - "created": 1753819483, + "created": 1753984043, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/1b8394f90636.json b/tests/integration/recordings/responses/1b8394f90636.json index 551f99d0f..8cf8a08f8 100644 --- a/tests/integration/recordings/responses/1b8394f90636.json +++ b/tests/integration/recordings/responses/1b8394f90636.json @@ -22,15 +22,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.383192Z", + "created_at": "2025-07-31T17:53:05.348866852Z", "done": true, "done_reason": "stop", - "total_duration": 2393598000, - "load_duration": 90501917, + "total_duration": 8992037446, + "load_duration": 46010833, "prompt_eval_count": 18, - "prompt_eval_duration": 545025792, + "prompt_eval_duration": 1464208894, "eval_count": 43, - "eval_duration": 1756031208, + "eval_duration": 7481270003, "response": " _______.\n\nThe best answer is blue. The traditional nursery rhyme goes like this:\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you! (Or something similar.)", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/1b92be674e2a.json b/tests/integration/recordings/responses/1b92be674e2a.json index 38e2633c4..2ed061949 100644 --- a/tests/integration/recordings/responses/1b92be674e2a.json +++ b/tests/integration/recordings/responses/1b92be674e2a.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:09.951749Z", + "created_at": "2025-07-31T17:50:06.140190726Z", "done": true, "done_reason": "stop", - "total_duration": 1110064084, - "load_duration": 86978750, + "total_duration": 5213341378, + "load_duration": 43943569, "prompt_eval_count": 23, - "prompt_eval_duration": 71337125, + "prompt_eval_duration": 1049424427, "eval_count": 24, - "eval_duration": 951124708, + "eval_duration": 4119422888, "response": "Mark Zuckerberg is the founder, chairman and CEO of Meta, which he originally founded as Facebook in 2004.", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/1f48f4b2ae33.json b/tests/integration/recordings/responses/1f48f4b2ae33.json index 5bef80e0d..faf498e21 100644 --- a/tests/integration/recordings/responses/1f48f4b2ae33.json +++ b/tests/integration/recordings/responses/1f48f4b2ae33.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 33233334, - "load_duration": 22084667, + "total_duration": 14447665, + "load_duration": 7154333, "prompt_eval_count": 2, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.024330627, - 0.016706778, - 0.03767714, - -0.009157433, - -0.03053444, - -0.017140865, - 0.07427198, - 0.04569162, - -0.0093771415, - 0.009883054, - -0.0056895353, - 0.00766826, - 0.039537337, - 0.015226259, - -0.08318956, - 0.019439543, - -0.022046668, - -0.033254836, - -0.18105465, - -0.1302509, - -0.002267121, - 0.013451511, - -0.024325471, - -0.0370128, - 0.002008361, - 0.085667126, - 0.0047639436, - -0.0033431135, - -0.006082333, - -0.115755625, - 0.06682907, - -0.018777594, - 0.08786826, - -0.0074177794, - -0.09357302, - 0.06146397, - -0.0811061, - 0.012222829, - 0.039710645, - -0.0026197857, - -0.04657112, - -0.08183902, - 0.039596144, - 0.015451171, - 0.043706182, - 0.103643835, - -0.058421474, - 0.036699373, - -0.05269955, - 0.040590122, - -0.1257893, - 0.0065005445, - -0.035836272, - -0.010050958, - -0.023851683, - 0.04597228, - 0.0146055985, - 0.01941457, - 0.028465142, - -0.055030942, - 0.024210218, - -0.052867528, - 0.015230754, - -0.004392124, - 0.092372015, - 0.033849876, - -0.047372803, - 0.032044917, - 0.0013220925, - -0.051211506, - 0.025862314, - 0.08155329, - 0.04092597, - 0.019154714, - 0.056453936, - -0.05275891, - 0.030533383, - -0.016634358, - 0.078772455, - -0.05426298, - -0.042149365, - -0.045443613, - -0.052689914, - 0.112255, - 0.01989106, - -0.042375352, - -0.0116811395, - 0.024315955, - 0.019157894, - -0.016550401, - -0.010308833, - -0.0854528, - 0.023834353, - -0.042181354, - -0.02503507, - 0.062114812, - -0.0045557567, - -0.15369567, - 0.0011066995, - 0.19423287, - -0.033851102, - 0.026153002, - -0.020320926, - 0.0012884212, - -0.0010269387, - -0.024112608, - 0.01749549, - -0.009808729, - 0.070379406, - -0.13769858, - -0.11118059, - -0.017364793, - 0.06603104, - -0.051888943, - 0.0019609837, - 0.014606661, - 0.060775448, - 0.09628018, - 0.013551948, - 0.019343184, - -0.00010513823, - -0.026652295, - -0.009341821, - 0.070832476, - -0.0034617381, - -0.06241276, - -0.044611085, - -8.796703e-34, - -0.11188401, - -0.042566102, - 0.027425224, - 0.06574075, - 0.0028303477, - -0.044104453, - 0.0052388306, - -0.036899917, - -0.015583542, - 0.020654282, - -0.059225976, - 0.007236481, - -0.028716046, - 0.040467374, - 0.13387094, - 0.0067958245, - -0.016369572, - 0.082198456, - -0.02261006, - -0.036412977, - 0.065244555, - 0.021011828, - -0.00547238, - -0.038433444, - 0.0014620472, - 0.0073671998, - 0.016773432, - -0.062663004, - 0.035388518, - -0.014395802, - 0.027888596, - 0.08375459, - -0.027772011, - -0.0036210902, - 0.039035592, - -0.026879633, - -0.018737212, - 0.019059159, - 0.06522145, - 0.007041419, - 0.0047491803, - -0.0030224104, - 0.040062234, - 0.028016087, - -0.004660967, - 0.012264516, - 0.08708115, - -0.007017102, - -0.037498116, - 0.011326796, - 0.015419678, - 0.013775384, - 0.017958459, - -0.009817914, - 0.090115435, - 0.05170552, - -0.034259032, - 0.0043903063, - -0.018848868, - -0.03148135, - 0.08216297, - 0.01687526, - -0.022163706, - 0.06844145, - 0.01581626, - 0.020322636, - 0.006385708, - 0.01646202, - 0.12718281, - 0.014996439, - -0.010813829, - 0.0017669294, - 0.03166719, - -0.044353943, - -0.05225622, - 0.022843977, - 0.050988894, - -0.018916972, - 0.0027931023, - -0.033645585, - -0.13571607, - -0.02701516, - -0.03567225, - -0.033537835, - 0.04786428, - -0.005438142, - 0.021346746, - -0.040034916, - 0.019374574, - 0.012011435, - -0.043362334, - 0.00054703583, - 0.03487962, - 0.017960638, - -0.06250195, - 8.224181e-34, - -0.094501406, - 0.013776652, - -0.025351115, - 0.098992504, - 0.04550355, - -0.020534594, - -0.02969489, - -0.05920057, - 0.042453784, - 0.0844487, - -0.043211546, - -0.0077362475, - 0.04935481, - 0.04203367, - -0.036539596, - 0.014424799, - 0.04035699, - -0.05897147, - 0.010022975, - 0.059877153, - -0.02790866, - 0.034927685, - -0.08759751, - -0.060616292, - -0.0048867413, - 0.08776904, - -0.0053599314, - -0.021816812, - -0.048162397, - 0.046919808, - 0.008398897, - -0.05172891, - -0.020422194, - 0.08581075, - -0.022597928, - 0.034425054, - -0.014506652, - 0.0031332595, - -0.04651879, - 0.030281473, - 0.039713893, - 0.029692288, - -0.093102165, - 0.05152783, - 0.0078089847, - -0.057008673, - -0.0417926, - 0.08987065, - -0.008134044, - -0.040822867, - -0.053487618, - -0.03437895, - -0.04525393, - -0.09715309, - -0.05819444, - 0.060935497, - -0.009079973, - 0.0069185137, - 0.012345735, - 0.06203646, - -0.006023849, - -0.08642951, - 0.058728326, - 0.053304967, - -0.053526226, - 0.039521404, - -0.044984024, - 0.07279109, - -0.039616205, - -0.05134445, - 0.103348814, - 0.021767734, - 0.00016650943, - 0.009423315, - 0.022016354, - -0.006902842, - -0.12888299, - -0.009864121, - -0.03639677, - -0.042481665, - 0.00442071, - -0.04766024, - 0.0065179234, - 0.102602765, - -0.05316684, - 0.07328582, - 0.015810942, - -0.029149026, - 0.025130942, - -0.06305578, - -0.04346251, - 0.06719971, - 0.014921193, - -0.0010985582, - -0.0986947, - -1.468275e-08, - 0.00461101, - -0.06715222, - 0.0764481, - -0.019802472, - 0.06737911, - 0.044783674, - -0.050963383, - -0.0077186986, - -0.029319696, - 0.028867694, - 0.018877203, - -0.02427935, - 0.044120654, - 0.044162665, - 0.034328103, - 0.04651797, - 0.021580769, - -0.0017484649, - -0.002995664, - 0.014355778, - 0.12525897, - 0.03431847, - -0.014617607, - 0.039184693, - -0.0023036075, - -0.014352938, - 0.010101757, - 0.024309622, - -0.041730713, - 0.088324144, - -0.031459358, - 0.03007363, - -0.0029376259, - 0.0049478672, - 0.09588392, - 0.09396657, - 0.014125666, - -0.077148244, - -0.039246853, - -0.010649013, - -0.008556113, - 0.06409407, - -0.03303714, - -0.030499754, - 0.09458461, - -0.008954661, - -0.029921891, - -0.13298501, - 0.059934624, - -0.011668433, - 0.007173723, - 0.035627667, - 0.0041028494, - 0.05619811, - 0.07656151, - -0.010067124, - 0.056783147, - 0.023536064, - -0.06377051, - 0.08934554, - 0.04375695, - 0.04337245, - 0.046287097, - -0.07039029 + -0.02437173, + 0.016691571, + 0.037651334, + -0.009169466, + -0.030578919, + -0.017076846, + 0.07430663, + 0.04566769, + -0.009368396, + 0.009935814, + -0.005688737, + 0.0076862546, + 0.039611656, + 0.015205378, + -0.083217494, + 0.019419, + -0.02204396, + -0.033271633, + -0.1810318, + -0.13026708, + -0.0022674492, + 0.013488196, + -0.02434116, + -0.036990467, + 0.0020005303, + 0.085704565, + 0.0048125703, + -0.0033972764, + -0.006103051, + -0.115693145, + 0.06680104, + -0.018714054, + 0.087860815, + -0.0074218297, + -0.093622826, + 0.06146732, + -0.08101325, + 0.012246703, + 0.039720677, + -0.0026448935, + -0.046548385, + -0.08182065, + 0.039587125, + 0.015448616, + 0.043755636, + 0.10366629, + -0.058401912, + 0.036637913, + -0.05266015, + 0.040520575, + -0.12584542, + 0.006513187, + -0.035867475, + -0.010049964, + -0.023868648, + 0.045952503, + 0.014643884, + 0.019400682, + 0.028450979, + -0.055104982, + 0.024237337, + -0.052891206, + 0.01526735, + -0.0043554287, + 0.092411734, + 0.033868838, + -0.0473778, + 0.032004114, + 0.0013329537, + -0.051189225, + 0.025842959, + 0.08156662, + 0.040880162, + 0.019199997, + 0.056540668, + -0.052779485, + 0.030562375, + -0.016645174, + 0.07881059, + -0.05429127, + -0.042108275, + -0.045501526, + -0.052706387, + 0.11225399, + 0.019902289, + -0.042404417, + -0.011690239, + 0.024314694, + 0.019212393, + -0.01657069, + -0.010302843, + -0.08546401, + 0.02384196, + -0.042174995, + -0.024951732, + 0.062075637, + -0.00458379, + -0.15365618, + 0.0011485998, + 0.19421324, + -0.033859447, + 0.02611495, + -0.020310923, + 0.0013013423, + -0.0009998817, + -0.024108203, + 0.017511548, + -0.009832005, + 0.07044699, + -0.1376917, + -0.11118457, + -0.017314779, + 0.06600386, + -0.051878963, + 0.0019530356, + 0.014586777, + 0.06080839, + 0.096305825, + 0.0135452775, + 0.019365564, + -9.473925e-05, + -0.026673997, + -0.009385724, + 0.07080032, + -0.0033958114, + -0.062400278, + -0.044617876, + -8.786998e-34, + -0.11190001, + -0.042532727, + 0.027410066, + 0.06570358, + 0.0028389343, + -0.044089977, + 0.005261214, + -0.036915902, + -0.015572142, + 0.020601038, + -0.059248205, + 0.0072750626, + -0.028684014, + 0.040509213, + 0.13384926, + 0.006766541, + -0.016410895, + 0.08215301, + -0.02261861, + -0.03641547, + 0.0652159, + 0.020951675, + -0.005514451, + -0.03837839, + 0.0014661213, + 0.007356805, + 0.016814455, + -0.062671445, + 0.035449203, + -0.014394421, + 0.027855018, + 0.083778515, + -0.027821619, + -0.003602467, + 0.039032556, + -0.02683506, + -0.01879125, + 0.01901679, + 0.06520433, + 0.007066841, + 0.0047632074, + -0.002972422, + 0.04009995, + 0.027956821, + -0.004595677, + 0.01224324, + 0.08707175, + -0.0070247534, + -0.037466988, + 0.0112514375, + 0.015385426, + 0.013791659, + 0.017975507, + -0.009874813, + 0.09012836, + 0.05173974, + -0.03426752, + 0.0043836883, + -0.018890336, + -0.03143595, + 0.0821047, + 0.016943024, + -0.02216519, + 0.06846694, + 0.015813861, + 0.020375654, + 0.0063640494, + 0.01645771, + 0.12721963, + 0.0150219, + -0.010827533, + 0.0017831607, + 0.031596202, + -0.04437783, + -0.0522816, + 0.02283393, + 0.050929666, + -0.01897314, + 0.002736589, + -0.03365577, + -0.13567695, + -0.027060354, + -0.035655867, + -0.033519205, + 0.047887404, + -0.005414933, + 0.02131625, + -0.04000849, + 0.019388696, + 0.011998282, + -0.04336669, + 0.00050136494, + 0.03487659, + 0.017963642, + -0.06246313, + 8.231736e-34, + -0.09450524, + 0.013722238, + -0.025376102, + 0.099012874, + 0.045497514, + -0.020499378, + -0.029740887, + -0.059197847, + 0.042443916, + 0.08437303, + -0.043213997, + -0.007738174, + 0.049371954, + 0.04206579, + -0.036542624, + 0.014377386, + 0.040342458, + -0.058944605, + 0.010021014, + 0.05985318, + -0.027902877, + 0.0349437, + -0.08764893, + -0.060625143, + -0.004807651, + 0.08776686, + -0.005401222, + -0.021765916, + -0.048159987, + 0.046951044, + 0.008384747, + -0.051710356, + -0.020393599, + 0.085794024, + -0.022611415, + 0.03439592, + -0.0144272465, + 0.0031382157, + -0.046493594, + 0.03027418, + 0.039738458, + 0.029673891, + -0.093155324, + 0.051494524, + 0.007791395, + -0.057023305, + -0.041827053, + 0.089955375, + -0.008166286, + -0.040813755, + -0.053475816, + -0.034331154, + -0.045241453, + -0.09715105, + -0.058199886, + 0.060881007, + -0.009054726, + 0.006942832, + 0.012339512, + 0.06204418, + -0.006036043, + -0.0864186, + 0.058729477, + 0.053356454, + -0.05354962, + 0.039538804, + -0.044991873, + 0.07283141, + -0.03960586, + -0.051347718, + 0.103338495, + 0.02179528, + 0.00014486129, + 0.009510344, + 0.021997727, + -0.0068747676, + -0.1288963, + -0.009832364, + -0.036413576, + -0.04248718, + 0.004492611, + -0.047635976, + 0.006537413, + 0.1025696, + -0.053211726, + 0.07332653, + 0.015861318, + -0.02916268, + 0.025154423, + -0.06311103, + -0.043543685, + 0.06714647, + 0.014881924, + -0.0010914755, + -0.09870542, + -1.4681843e-08, + 0.004633685, + -0.067102544, + 0.07647304, + -0.01981857, + 0.06737649, + 0.04482623, + -0.050963704, + -0.0077299844, + -0.029333303, + 0.028893374, + 0.018828921, + -0.024264988, + 0.044066, + 0.04414379, + 0.034373876, + 0.046520673, + 0.021618845, + -0.0017267675, + -0.0029906677, + 0.014380984, + 0.12527594, + 0.03429198, + -0.014653963, + 0.039171875, + -0.002297837, + -0.014404986, + 0.010117208, + 0.024292482, + -0.04174585, + 0.08831709, + -0.03145136, + 0.030084575, + -0.0029161053, + 0.00487737, + 0.09588144, + 0.09388587, + 0.014207209, + -0.07716958, + -0.039264996, + -0.010718448, + -0.008490537, + 0.064107336, + -0.03299578, + -0.03049028, + 0.09460791, + -0.008975077, + -0.029871479, + -0.13294572, + 0.059894353, + -0.011694143, + 0.0071492735, + 0.035602562, + 0.0040614423, + 0.056197774, + 0.07654246, + -0.010018939, + 0.056764524, + 0.023490718, + -0.0637896, + 0.0893437, + 0.043716535, + 0.04345191, + 0.046286818, + -0.070387095 ] ] } diff --git a/tests/integration/recordings/responses/211b1562d4e6.json b/tests/integration/recordings/responses/211b1562d4e6.json index 05aefe656..6074d46a2 100644 --- a/tests/integration/recordings/responses/211b1562d4e6.json +++ b/tests/integration/recordings/responses/211b1562d4e6.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.108049Z", + "created_at": "2025-07-31T17:53:32.034301682Z", "done": true, "done_reason": "stop", - "total_duration": 334746667, - "load_duration": 55090709, + "total_duration": 2744857982, + "load_duration": 42715044, "prompt_eval_count": 23, - "prompt_eval_duration": 74557791, + "prompt_eval_duration": 1796728611, "eval_count": 6, - "eval_duration": 204410292, + "eval_duration": 904888348, "response": "Humans live on Earth.", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/2afe3b38ca01.json b/tests/integration/recordings/responses/2afe3b38ca01.json index ba6382ca6..13cbdd47a 100644 --- a/tests/integration/recordings/responses/2afe3b38ca01.json +++ b/tests/integration/recordings/responses/2afe3b38ca01.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.08526Z", + "created_at": "2025-07-31T17:44:24.112976431Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.127897Z", + "created_at": "2025-07-31T17:44:24.296397308Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.169908Z", + "created_at": "2025-07-31T17:44:24.481740915Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.212259Z", + "created_at": "2025-07-31T17:44:24.69085854Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.255764Z", + "created_at": "2025-07-31T17:44:24.912926392Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.297701Z", + "created_at": "2025-07-31T17:44:25.132377421Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.33936Z", + "created_at": "2025-07-31T17:44:25.350113938Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.381619Z", + "created_at": "2025-07-31T17:44:25.570117503Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.423841Z", + "created_at": "2025-07-31T17:44:25.781633125Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.466161Z", + "created_at": "2025-07-31T17:44:25.967547363Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.508442Z", + "created_at": "2025-07-31T17:44:26.153858447Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.552476Z", + "created_at": "2025-07-31T17:44:26.339696219Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,15 +238,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:10.596175Z", + "created_at": "2025-07-31T17:44:26.527991344Z", "done": true, "done_reason": "stop", - "total_duration": 656770208, - "load_duration": 67886042, + "total_duration": 4477782788, + "load_duration": 41272054, "prompt_eval_count": 399, - "prompt_eval_duration": 74761708, + "prompt_eval_duration": 2019228112, "eval_count": 13, - "eval_duration": 513356958, + "eval_duration": 2416529563, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/2d187a11704c.json b/tests/integration/recordings/responses/2d187a11704c.json index c2604364f..cfc07c895 100644 --- a/tests/integration/recordings/responses/2d187a11704c.json +++ b/tests/integration/recordings/responses/2d187a11704c.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.104971Z", + "created_at": "2025-07-31T17:45:10.501556237Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.150643Z", + "created_at": "2025-07-31T17:45:10.681603728Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.193729Z", + "created_at": "2025-07-31T17:45:10.862378525Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.235751Z", + "created_at": "2025-07-31T17:45:11.048405016Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.277303Z", + "created_at": "2025-07-31T17:45:11.232090592Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.3194Z", + "created_at": "2025-07-31T17:45:11.411960511Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.36125Z", + "created_at": "2025-07-31T17:45:11.594645161Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.40507Z", + "created_at": "2025-07-31T17:45:11.770984586Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.450288Z", + "created_at": "2025-07-31T17:45:11.959640962Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.494537Z", + "created_at": "2025-07-31T17:45:12.15150461Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.536615Z", + "created_at": "2025-07-31T17:45:12.329179165Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.578045Z", + "created_at": "2025-07-31T17:45:12.508407538Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.619451Z", + "created_at": "2025-07-31T17:45:12.687572823Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.660899Z", + "created_at": "2025-07-31T17:45:12.863560188Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.702694Z", + "created_at": "2025-07-31T17:45:13.044545528Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.744157Z", + "created_at": "2025-07-31T17:45:13.221961172Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.78564Z", + "created_at": "2025-07-31T17:45:13.401324607Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.826981Z", + "created_at": "2025-07-31T17:45:13.582305037Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,7 +346,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.868479Z", + "created_at": "2025-07-31T17:45:13.76376504Z", "done": false, "done_reason": null, "total_duration": null, @@ -364,7 +364,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.909719Z", + "created_at": "2025-07-31T17:45:13.944303696Z", "done": false, "done_reason": null, "total_duration": null, @@ -382,7 +382,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.950947Z", + "created_at": "2025-07-31T17:45:14.125163201Z", "done": false, "done_reason": null, "total_duration": null, @@ -400,7 +400,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:17.992338Z", + "created_at": "2025-07-31T17:45:14.305006544Z", "done": false, "done_reason": null, "total_duration": null, @@ -418,7 +418,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.034644Z", + "created_at": "2025-07-31T17:45:14.484481513Z", "done": false, "done_reason": null, "total_duration": null, @@ -436,7 +436,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.076808Z", + "created_at": "2025-07-31T17:45:14.667407015Z", "done": false, "done_reason": null, "total_duration": null, @@ -454,7 +454,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.118703Z", + "created_at": "2025-07-31T17:45:14.845425227Z", "done": false, "done_reason": null, "total_duration": null, @@ -472,7 +472,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.160749Z", + "created_at": "2025-07-31T17:45:15.028728032Z", "done": false, "done_reason": null, "total_duration": null, @@ -490,7 +490,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.203112Z", + "created_at": "2025-07-31T17:45:15.209162691Z", "done": false, "done_reason": null, "total_duration": null, @@ -508,7 +508,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.245415Z", + "created_at": "2025-07-31T17:45:15.386536468Z", "done": false, "done_reason": null, "total_duration": null, @@ -526,7 +526,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.287409Z", + "created_at": "2025-07-31T17:45:15.567411759Z", "done": false, "done_reason": null, "total_duration": null, @@ -544,7 +544,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.329242Z", + "created_at": "2025-07-31T17:45:15.746245449Z", "done": false, "done_reason": null, "total_duration": null, @@ -562,7 +562,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.370759Z", + "created_at": "2025-07-31T17:45:15.924864832Z", "done": false, "done_reason": null, "total_duration": null, @@ -580,7 +580,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.412875Z", + "created_at": "2025-07-31T17:45:16.108171685Z", "done": false, "done_reason": null, "total_duration": null, @@ -598,7 +598,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.454857Z", + "created_at": "2025-07-31T17:45:16.290545444Z", "done": false, "done_reason": null, "total_duration": null, @@ -616,7 +616,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.497182Z", + "created_at": "2025-07-31T17:45:16.471500704Z", "done": false, "done_reason": null, "total_duration": null, @@ -634,7 +634,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.540949Z", + "created_at": "2025-07-31T17:45:16.65171538Z", "done": false, "done_reason": null, "total_duration": null, @@ -652,7 +652,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.583594Z", + "created_at": "2025-07-31T17:45:16.835027449Z", "done": false, "done_reason": null, "total_duration": null, @@ -670,7 +670,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.626557Z", + "created_at": "2025-07-31T17:45:17.01515185Z", "done": false, "done_reason": null, "total_duration": null, @@ -688,7 +688,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.669737Z", + "created_at": "2025-07-31T17:45:17.195610049Z", "done": false, "done_reason": null, "total_duration": null, @@ -706,7 +706,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.712956Z", + "created_at": "2025-07-31T17:45:17.374675826Z", "done": false, "done_reason": null, "total_duration": null, @@ -724,7 +724,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.759178Z", + "created_at": "2025-07-31T17:45:17.554298267Z", "done": false, "done_reason": null, "total_duration": null, @@ -742,7 +742,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.80112Z", + "created_at": "2025-07-31T17:45:17.737527818Z", "done": false, "done_reason": null, "total_duration": null, @@ -760,7 +760,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.844475Z", + "created_at": "2025-07-31T17:45:17.918877822Z", "done": false, "done_reason": null, "total_duration": null, @@ -778,7 +778,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.90258Z", + "created_at": "2025-07-31T17:45:18.105911231Z", "done": false, "done_reason": null, "total_duration": null, @@ -796,7 +796,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.946507Z", + "created_at": "2025-07-31T17:45:18.283923436Z", "done": false, "done_reason": null, "total_duration": null, @@ -814,7 +814,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:18.987529Z", + "created_at": "2025-07-31T17:45:18.459719394Z", "done": false, "done_reason": null, "total_duration": null, @@ -832,7 +832,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.028229Z", + "created_at": "2025-07-31T17:45:18.640906341Z", "done": false, "done_reason": null, "total_duration": null, @@ -850,7 +850,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.070662Z", + "created_at": "2025-07-31T17:45:18.826961086Z", "done": false, "done_reason": null, "total_duration": null, @@ -868,7 +868,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.112716Z", + "created_at": "2025-07-31T17:45:19.003320248Z", "done": false, "done_reason": null, "total_duration": null, @@ -886,7 +886,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.154631Z", + "created_at": "2025-07-31T17:45:19.182881734Z", "done": false, "done_reason": null, "total_duration": null, @@ -904,7 +904,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.196493Z", + "created_at": "2025-07-31T17:45:19.359495992Z", "done": false, "done_reason": null, "total_duration": null, @@ -922,7 +922,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.238168Z", + "created_at": "2025-07-31T17:45:19.537161501Z", "done": false, "done_reason": null, "total_duration": null, @@ -940,7 +940,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.279638Z", + "created_at": "2025-07-31T17:45:19.717797841Z", "done": false, "done_reason": null, "total_duration": null, @@ -958,7 +958,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.321162Z", + "created_at": "2025-07-31T17:45:19.897791992Z", "done": false, "done_reason": null, "total_duration": null, @@ -976,7 +976,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.363369Z", + "created_at": "2025-07-31T17:45:20.078554441Z", "done": false, "done_reason": null, "total_duration": null, @@ -994,7 +994,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.406499Z", + "created_at": "2025-07-31T17:45:20.256647775Z", "done": false, "done_reason": null, "total_duration": null, @@ -1012,7 +1012,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.448237Z", + "created_at": "2025-07-31T17:45:20.435851722Z", "done": false, "done_reason": null, "total_duration": null, @@ -1030,7 +1030,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.490347Z", + "created_at": "2025-07-31T17:45:20.614887157Z", "done": false, "done_reason": null, "total_duration": null, @@ -1048,7 +1048,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.532079Z", + "created_at": "2025-07-31T17:45:20.792750108Z", "done": false, "done_reason": null, "total_duration": null, @@ -1066,7 +1066,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.574944Z", + "created_at": "2025-07-31T17:45:20.97316853Z", "done": false, "done_reason": null, "total_duration": null, @@ -1084,7 +1084,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.616889Z", + "created_at": "2025-07-31T17:45:21.151001292Z", "done": false, "done_reason": null, "total_duration": null, @@ -1102,7 +1102,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.659068Z", + "created_at": "2025-07-31T17:45:21.329346225Z", "done": false, "done_reason": null, "total_duration": null, @@ -1120,7 +1120,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.701004Z", + "created_at": "2025-07-31T17:45:21.504478824Z", "done": false, "done_reason": null, "total_duration": null, @@ -1138,7 +1138,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.743887Z", + "created_at": "2025-07-31T17:45:21.684098768Z", "done": false, "done_reason": null, "total_duration": null, @@ -1156,7 +1156,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.785584Z", + "created_at": "2025-07-31T17:45:21.859984956Z", "done": false, "done_reason": null, "total_duration": null, @@ -1174,7 +1174,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.827467Z", + "created_at": "2025-07-31T17:45:22.039452807Z", "done": false, "done_reason": null, "total_duration": null, @@ -1192,7 +1192,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.86933Z", + "created_at": "2025-07-31T17:45:22.215880487Z", "done": false, "done_reason": null, "total_duration": null, @@ -1210,7 +1210,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.911647Z", + "created_at": "2025-07-31T17:45:22.394139356Z", "done": false, "done_reason": null, "total_duration": null, @@ -1228,7 +1228,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.954177Z", + "created_at": "2025-07-31T17:45:22.570525681Z", "done": false, "done_reason": null, "total_duration": null, @@ -1246,7 +1246,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:19.997803Z", + "created_at": "2025-07-31T17:45:22.749431991Z", "done": false, "done_reason": null, "total_duration": null, @@ -1264,7 +1264,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.041521Z", + "created_at": "2025-07-31T17:45:22.925115527Z", "done": false, "done_reason": null, "total_duration": null, @@ -1282,7 +1282,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.084075Z", + "created_at": "2025-07-31T17:45:23.101143142Z", "done": false, "done_reason": null, "total_duration": null, @@ -1300,7 +1300,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.125946Z", + "created_at": "2025-07-31T17:45:23.27672205Z", "done": false, "done_reason": null, "total_duration": null, @@ -1318,7 +1318,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.167976Z", + "created_at": "2025-07-31T17:45:23.454962118Z", "done": false, "done_reason": null, "total_duration": null, @@ -1336,7 +1336,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.209931Z", + "created_at": "2025-07-31T17:45:23.632398162Z", "done": false, "done_reason": null, "total_duration": null, @@ -1354,7 +1354,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.251411Z", + "created_at": "2025-07-31T17:45:23.809693128Z", "done": false, "done_reason": null, "total_duration": null, @@ -1372,7 +1372,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.292702Z", + "created_at": "2025-07-31T17:45:23.986707624Z", "done": false, "done_reason": null, "total_duration": null, @@ -1390,7 +1390,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.333936Z", + "created_at": "2025-07-31T17:45:24.166677843Z", "done": false, "done_reason": null, "total_duration": null, @@ -1408,7 +1408,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.375411Z", + "created_at": "2025-07-31T17:45:24.346964308Z", "done": false, "done_reason": null, "total_duration": null, @@ -1426,7 +1426,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.417414Z", + "created_at": "2025-07-31T17:45:24.5291978Z", "done": false, "done_reason": null, "total_duration": null, @@ -1444,7 +1444,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.458433Z", + "created_at": "2025-07-31T17:45:24.706691694Z", "done": false, "done_reason": null, "total_duration": null, @@ -1462,7 +1462,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.500158Z", + "created_at": "2025-07-31T17:45:24.88460415Z", "done": false, "done_reason": null, "total_duration": null, @@ -1480,7 +1480,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.541472Z", + "created_at": "2025-07-31T17:45:25.065539114Z", "done": false, "done_reason": null, "total_duration": null, @@ -1498,7 +1498,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.582783Z", + "created_at": "2025-07-31T17:45:25.243710072Z", "done": false, "done_reason": null, "total_duration": null, @@ -1516,7 +1516,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.623884Z", + "created_at": "2025-07-31T17:45:25.422103246Z", "done": false, "done_reason": null, "total_duration": null, @@ -1534,7 +1534,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.664779Z", + "created_at": "2025-07-31T17:45:25.600249999Z", "done": false, "done_reason": null, "total_duration": null, @@ -1552,7 +1552,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.706141Z", + "created_at": "2025-07-31T17:45:25.780448009Z", "done": false, "done_reason": null, "total_duration": null, @@ -1570,7 +1570,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.747329Z", + "created_at": "2025-07-31T17:45:25.958982616Z", "done": false, "done_reason": null, "total_duration": null, @@ -1588,7 +1588,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.788243Z", + "created_at": "2025-07-31T17:45:26.138676515Z", "done": false, "done_reason": null, "total_duration": null, @@ -1606,7 +1606,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.829298Z", + "created_at": "2025-07-31T17:45:26.314631321Z", "done": false, "done_reason": null, "total_duration": null, @@ -1624,7 +1624,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.870474Z", + "created_at": "2025-07-31T17:45:26.494578612Z", "done": false, "done_reason": null, "total_duration": null, @@ -1642,7 +1642,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.911553Z", + "created_at": "2025-07-31T17:45:26.673360876Z", "done": false, "done_reason": null, "total_duration": null, @@ -1660,7 +1660,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.95263Z", + "created_at": "2025-07-31T17:45:26.854225873Z", "done": false, "done_reason": null, "total_duration": null, @@ -1678,7 +1678,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:20.996106Z", + "created_at": "2025-07-31T17:45:27.02961726Z", "done": false, "done_reason": null, "total_duration": null, @@ -1696,7 +1696,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.040033Z", + "created_at": "2025-07-31T17:45:27.208399033Z", "done": false, "done_reason": null, "total_duration": null, @@ -1714,7 +1714,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.08252Z", + "created_at": "2025-07-31T17:45:27.386827403Z", "done": false, "done_reason": null, "total_duration": null, @@ -1732,7 +1732,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.126039Z", + "created_at": "2025-07-31T17:45:27.567038363Z", "done": false, "done_reason": null, "total_duration": null, @@ -1750,7 +1750,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.16902Z", + "created_at": "2025-07-31T17:45:27.746885334Z", "done": false, "done_reason": null, "total_duration": null, @@ -1768,7 +1768,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.211499Z", + "created_at": "2025-07-31T17:45:27.922875436Z", "done": false, "done_reason": null, "total_duration": null, @@ -1786,7 +1786,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.254364Z", + "created_at": "2025-07-31T17:45:28.101430863Z", "done": false, "done_reason": null, "total_duration": null, @@ -1804,15 +1804,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:21.29714Z", + "created_at": "2025-07-31T17:45:28.278640455Z", "done": true, "done_reason": "stop", - "total_duration": 4385243208, - "load_duration": 115325375, + "total_duration": 19952737777, + "load_duration": 43350760, "prompt_eval_count": 36, - "prompt_eval_duration": 74814291, + "prompt_eval_duration": 2130287337, "eval_count": 100, - "eval_duration": 4194301000, + "eval_duration": 17778591810, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/31407e035752.json b/tests/integration/recordings/responses/31407e035752.json index 078757420..fc52c7d6c 100644 --- a/tests/integration/recordings/responses/31407e035752.json +++ b/tests/integration/recordings/responses/31407e035752.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -21,7 +21,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -36,7 +36,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984691, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -47,7 +47,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -62,7 +62,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984692, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -73,7 +73,33 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", + "choices": [ + { + "delta": { + "content": " city", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984692, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -88,7 +114,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984692, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -99,7 +125,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -114,7 +140,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984692, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -125,7 +151,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -140,7 +166,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984692, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -151,7 +177,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -166,7 +192,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984693, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -177,7 +203,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -192,7 +218,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984693, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -203,7 +229,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -218,7 +244,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984693, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -229,7 +255,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -244,7 +270,7 @@ "logprobs": null } ], - "created": 1753814884, + "created": 1753984693, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -255,7 +281,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -270,7 +296,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984693, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -281,7 +307,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -296,7 +322,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984693, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -307,7 +333,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -322,7 +348,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984694, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -333,7 +359,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -348,7 +374,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984694, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -359,7 +385,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -374,7 +400,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984694, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -385,7 +411,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -400,7 +426,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984694, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -411,7 +437,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -426,7 +452,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984694, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -437,7 +463,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -452,7 +478,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984694, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -463,7 +489,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -478,7 +504,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984695, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -489,7 +515,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -504,7 +530,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984695, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -515,7 +541,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-850", + "id": "chatcmpl-847", "choices": [ { "delta": { @@ -530,7 +556,7 @@ "logprobs": null } ], - "created": 1753814885, + "created": 1753984695, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/35db283fef1d.json b/tests/integration/recordings/responses/35db283fef1d.json index 719606d3c..a813e0389 100644 --- a/tests/integration/recordings/responses/35db283fef1d.json +++ b/tests/integration/recordings/responses/35db283fef1d.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -38,7 +38,7 @@ "body": { "__type__": "openai.types.chat.chat_completion.ChatCompletion", "__data__": { - "id": "chatcmpl-331", + "id": "chatcmpl-264", "choices": [ { "finish_reason": "tool_calls", @@ -53,7 +53,7 @@ "function_call": null, "tool_calls": [ { - "id": "call_za2swdo9", + "id": "call_99dd5wna", "function": { "arguments": "{\"city\":\"Tokyo\"}", "name": "get_weather" @@ -65,15 +65,15 @@ } } ], - "created": 1753814888, + "created": 1753984717, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion", "service_tier": null, "system_fingerprint": "fp_ollama", "usage": { - "completion_tokens": 18, + "completion_tokens": 15, "prompt_tokens": 177, - "total_tokens": 195, + "total_tokens": 192, "completion_tokens_details": null, "prompt_tokens_details": null } diff --git a/tests/integration/recordings/responses/3c3f13cb7794.json b/tests/integration/recordings/responses/3c3f13cb7794.json index 856c8c47a..3ee40ec59 100644 --- a/tests/integration/recordings/responses/3c3f13cb7794.json +++ b/tests/integration/recordings/responses/3c3f13cb7794.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.322498Z", + "created_at": "2025-07-31T17:53:33.397554906Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.366077Z", + "created_at": "2025-07-31T17:53:33.586274792Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.408909Z", + "created_at": "2025-07-31T17:53:33.765410961Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.451051Z", + "created_at": "2025-07-31T17:53:33.943915332Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.492622Z", + "created_at": "2025-07-31T17:53:34.129095862Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.534265Z", + "created_at": "2025-07-31T17:53:34.311777808Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.576141Z", + "created_at": "2025-07-31T17:53:34.494210854Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.617693Z", + "created_at": "2025-07-31T17:53:34.678960603Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.658779Z", + "created_at": "2025-07-31T17:53:34.8612541Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.699936Z", + "created_at": "2025-07-31T17:53:35.046305051Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,15 +201,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:29.74208Z", + "created_at": "2025-07-31T17:53:35.234671072Z", "done": true, "done_reason": "stop", - "total_duration": 570982833, - "load_duration": 78768458, + "total_duration": 3173172688, + "load_duration": 41807088, "prompt_eval_count": 26, - "prompt_eval_duration": 69632083, + "prompt_eval_duration": 1292827697, "eval_count": 11, - "eval_duration": 421479000, + "eval_duration": 1837992731, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/4014dd44c15f.json b/tests/integration/recordings/responses/4014dd44c15f.json index 1c2ff7200..5ff0c9e52 100644 --- a/tests/integration/recordings/responses/4014dd44c15f.json +++ b/tests/integration/recordings/responses/4014dd44c15f.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -39,7 +39,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-448", + "id": "chatcmpl-347", "choices": [ { "delta": { @@ -50,7 +50,7 @@ "tool_calls": [ { "index": 0, - "id": "call_esyvjxp3", + "id": "call_9732h2cb", "function": { "arguments": "{\"city\":\"Tokyo\"}", "name": "get_weather" @@ -64,7 +64,7 @@ "logprobs": null } ], - "created": 1753814883, + "created": 1753984686, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -75,7 +75,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-448", + "id": "chatcmpl-347", "choices": [ { "delta": { @@ -90,7 +90,7 @@ "logprobs": null } ], - "created": 1753814883, + "created": 1753984686, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/40f524d1934a.json b/tests/integration/recordings/responses/40f524d1934a.json index 68c0470c7..834fb6a74 100644 --- a/tests/integration/recordings/responses/40f524d1934a.json +++ b/tests/integration/recordings/responses/40f524d1934a.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.070599Z", + "created_at": "2025-07-31T17:54:04.749672905Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.112828Z", + "created_at": "2025-07-31T17:54:04.935716874Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.154976Z", + "created_at": "2025-07-31T17:54:05.124992167Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.197203Z", + "created_at": "2025-07-31T17:54:05.313306567Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.239672Z", + "created_at": "2025-07-31T17:54:05.496860421Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.281331Z", + "created_at": "2025-07-31T17:54:05.682903521Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.323134Z", + "created_at": "2025-07-31T17:54:05.868324667Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.364766Z", + "created_at": "2025-07-31T17:54:06.061112692Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.406481Z", + "created_at": "2025-07-31T17:54:06.268369491Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.448383Z", + "created_at": "2025-07-31T17:54:06.47768613Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,15 +201,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.490154Z", + "created_at": "2025-07-31T17:54:06.671958182Z", "done": true, "done_reason": "stop", - "total_duration": 531176667, - "load_duration": 65048792, + "total_duration": 2151822386, + "load_duration": 44157323, "prompt_eval_count": 324, - "prompt_eval_duration": 44536417, + "prompt_eval_duration": 183972482, "eval_count": 11, - "eval_duration": 420819750, + "eval_duration": 1923143597, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/43e106de6736.json b/tests/integration/recordings/responses/43e106de6736.json index b565f0dc9..72e055ee3 100644 --- a/tests/integration/recordings/responses/43e106de6736.json +++ b/tests/integration/recordings/responses/43e106de6736.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 58258750, - "load_duration": 25899542, + "total_duration": 15536662, + "load_duration": 7128104, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.028407024, - 0.08176727, - -0.07856116, - 0.027924549, - 0.05008439, - -0.035268802, - -0.0040619136, - 0.029315198, - -0.05775003, - 0.013769637, - 0.14610882, - -0.012019041, - -0.024392882, - -0.05509032, - -0.02661779, - -0.013253934, - -0.109151706, - -0.037233494, - -0.0036058167, - 0.04766495, - 0.06212885, - 0.0070259646, - -0.015513743, - -0.008010851, - 0.037648663, - 0.01587603, - -0.041856695, - 0.09732178, - -0.025641596, - -0.11368298, - 0.03550726, - 0.07043342, - 0.016779423, - 0.02220752, - 0.123395406, - 0.0077137193, - 0.12550895, - 0.008077936, - -0.026158499, - 0.0028612812, - 0.018155744, - -0.04666325, - 0.041025575, - 0.0013476727, - 0.0019516364, - 0.008663665, - 0.016689047, - 0.02200178, - 0.0020768014, - -0.032861207, - -0.086455174, - 0.008047145, - -0.07434091, - -0.016292974, - 0.06051878, - 0.005966867, - 0.0160179, - 0.021412006, - 0.009540338, - 0.03177335, - 0.023032434, - 0.03437097, - -0.04224765, - 0.024748176, - 0.116213955, - -0.024936162, - -0.03895259, - -0.024991278, - -0.020854436, - -0.08835937, - -0.15073228, - 0.020921277, - -0.022518696, - 0.0023868105, - 0.0057663955, - -0.0015790414, - -0.11985628, - -0.0029912454, - 0.0550998, - -0.11830636, - -0.058846988, - -0.15046737, - 0.018624697, - -0.0093440395, - -0.028901154, - 0.08400474, - 0.0437436, - -0.0006745939, - -0.052540295, - 0.00024754918, - 0.040431518, - 0.0066545215, - 0.02609114, - 0.051891107, - 0.012606882, - 0.061448827, - 0.013889043, - 0.038454182, - 0.048222367, - 0.104106456, - -0.026478294, - -0.021488149, - -0.020865437, - 0.05061779, - -0.05171592, - -0.07573864, - 0.057483904, - -0.049993664, - 0.06528295, - -0.02875688, - 0.038766492, - -0.062760465, - -0.0144796055, - -0.063462086, - 0.06642258, - -0.014848135, - -0.03523116, - 0.0774014, - -0.039893247, - 0.032182425, - 0.10171478, - -0.022525396, - -0.059299074, - 0.00038746602, - -0.05779858, - -0.07034273, - 0.06375495, - -4.088634e-33, - -0.021801252, - -0.07985834, - -0.013881648, - 0.14923096, - 0.02520313, - -0.042283125, - -0.0067697223, - 0.054634638, - -0.09223034, - 0.0081036305, - -0.03861765, - -0.117698364, - 0.012977803, - 0.034548674, - -0.01703291, - 0.011910173, - 0.012945288, - 0.04277919, - -0.017591223, - -0.0184066, - 0.06513148, - 0.04050013, - -0.02252127, - -0.060939074, - -0.018603502, - 0.011679816, - 0.01410369, - -0.06763908, - 0.08543174, - 0.030138582, - 0.010859261, - -0.054844614, - -0.024129191, - 0.048327282, - 0.00750549, - 0.013356204, - 0.024558878, - -0.005942624, - -0.045620095, - -0.00484637, - 0.004418298, - -0.0023806267, - 0.013590539, - -0.016870445, - 0.06959721, - -0.07736302, - 0.02058481, - 0.0048155314, - 0.055696823, - 0.0131223425, - -0.011748222, - 0.040935397, - 0.007458848, - 0.042072233, - 0.010358565, - 0.019406458, - 0.011092792, - 0.017259602, - 0.018278012, - 0.077335365, - 0.019612921, - 0.05268688, - -0.05863009, - 0.039751627, - -0.050250556, - -0.048913844, - -0.05265637, - -0.09227304, - 0.0755598, - 0.08097828, - -0.022257954, - -0.042141132, - 0.056546185, - 0.023585746, - 0.0015263582, - -0.049815144, - 0.002336895, - 0.028626408, - -0.06897293, - -0.04780049, - -0.048637427, - -0.076585636, - -0.03285766, - -0.046012525, - -0.0573021, - -0.080889866, - -0.008056378, - -0.0936112, - 0.051229417, - -0.058302302, - -0.0005942833, - 0.02222621, - -0.046907477, - -0.08964737, - 0.1195762, - 2.0452953e-33, - 0.012159685, - 0.086426094, - -0.023217503, - 0.002771192, - -0.0010614472, - 0.03487195, - 0.07328719, - -0.049876485, - -0.041938163, - 0.13486409, - -0.00690217, - 0.006254477, - 0.059122436, - -0.028893106, - 0.09141587, - -0.018487127, - 0.0077112317, - -0.044207573, - -0.0251735, - -0.014999972, - -0.035417248, - 0.12413253, - 0.13118097, - 0.081015825, - -0.03327241, - 0.003976432, - 0.026454262, - 0.026598025, - 0.017349144, - -0.0036153824, - 0.035460044, - 0.05956128, - -0.124593176, - 0.021954069, - 0.025635097, - -0.11063109, - 0.096061416, - -0.06731725, - -0.011819293, - 0.042329434, - 0.03790837, - 0.10582649, - 0.0073426333, - 0.06629678, - 0.022922922, - 0.0494007, - 0.14639522, - -0.0067070075, - 0.004380622, - -0.029196544, - -0.009010303, - -0.08637028, - 0.03588363, - 0.0029887543, - -0.029351206, - 0.07019312, - 0.014898416, - 0.028345235, - -0.040354595, - 0.01916304, - 0.015590835, - 0.028637327, - -0.019529723, - -0.018309733, - -0.0054176697, - -0.093132764, - -0.06116049, - 0.038816936, - 0.02793884, - 0.034137025, - -0.027511358, - 0.010699668, - -0.05521562, - -0.07380209, - 0.021521263, - -0.015450832, - -0.024988633, - -0.004755674, - 0.030465573, - -0.024057997, - 0.0341225, - -0.0103128245, - -0.012666524, - 0.03628323, - -0.0044518244, - -0.014977736, - 0.02790076, - 0.0978009, - -0.026436698, - -0.005187212, - -0.019124882, - 0.06205225, - 0.052137945, - 0.037870288, - 0.012578256, - -1.705626e-08, - -0.05000592, - -0.08913878, - -0.0035273295, - -0.01577607, - -0.021846429, - 0.07184407, - -0.050185654, - -0.010643527, - -0.030602882, - -0.01577121, - 0.013220822, - -0.0025653532, - -0.04210823, - 0.009286525, - -0.041129403, - -0.029615805, - 0.002200794, - -0.032989334, - -0.05041253, - -0.021504797, - -0.0068345494, - 0.0084738685, - 0.03568697, - 0.0252117, - -0.016504692, - 0.04915123, - 0.018349955, - 0.049084183, - -0.058165494, - -0.015055481, - 0.045743454, - 0.049920842, - 0.020444298, - -0.052004594, - -0.033592116, - 0.061816722, - 0.111411005, - 0.07770497, - 0.022457859, - 0.0025742552, - -0.043929543, - 0.008576763, - -0.036182683, - 0.029673496, - -0.017278075, - -0.09458994, - -0.057882637, - -0.06579892, - -0.06124832, - -0.10455079, - -0.02925637, - 0.0013624659, - 0.0060532107, - 0.04077331, - -0.036694046, - 0.016800206, - 0.005279432, - 0.030968234, - -0.05446385, - 0.0048696757, - 0.070877954, - 0.06684445, - 0.017715273, - -0.029237686 + -0.02839711, + 0.0818053, + -0.07853445, + 0.02792148, + 0.05005452, + -0.035238173, + -0.0040396755, + 0.02928838, + -0.057782255, + 0.013747614, + 0.14607728, + -0.012043185, + -0.024383053, + -0.055092573, + -0.026610607, + -0.01324528, + -0.109175414, + -0.037209943, + -0.0035725583, + 0.04765195, + 0.06211419, + 0.00703526, + -0.015518899, + -0.007973487, + 0.03763324, + 0.01586704, + -0.041856498, + 0.097324215, + -0.02564764, + -0.11369229, + 0.035487138, + 0.07041544, + 0.016785262, + 0.022201158, + 0.1234195, + 0.007680676, + 0.12553541, + 0.0081102215, + -0.026146678, + 0.0028899247, + 0.018154126, + -0.046665825, + 0.041037504, + 0.0013452142, + 0.0019477131, + 0.008671534, + 0.016716687, + 0.02204051, + 0.0020750419, + -0.032865297, + -0.08644402, + 0.008038449, + -0.07436438, + -0.016300498, + 0.060510594, + 0.0059645884, + 0.015995186, + 0.021407088, + 0.009546037, + 0.03173758, + 0.023011131, + 0.03439496, + -0.042227626, + 0.024753809, + 0.11620387, + -0.024936425, + -0.03898177, + -0.024962299, + -0.020868327, + -0.08833928, + -0.15071589, + 0.020941459, + -0.022525651, + 0.0023695363, + 0.0057225176, + -0.0015978776, + -0.11984311, + -0.0029637238, + 0.05510895, + -0.11829667, + -0.058854777, + -0.1504783, + 0.018591402, + -0.009350579, + -0.02891901, + 0.083976336, + 0.043746613, + -0.0006955484, + -0.05254747, + 0.00023166445, + 0.04039829, + 0.006650695, + 0.02611124, + 0.05187556, + 0.012637232, + 0.061457768, + 0.013881842, + 0.038474612, + 0.04822178, + 0.10411109, + -0.026456181, + -0.021487249, + -0.020877272, + 0.050628837, + -0.051682167, + -0.07575808, + 0.05747169, + -0.04998164, + 0.06526268, + -0.028748322, + 0.038778387, + -0.062783346, + -0.014459063, + -0.06346632, + 0.06643585, + -0.014839471, + -0.03520943, + 0.07738897, + -0.03990594, + 0.03218616, + 0.10172238, + -0.02251418, + -0.059295975, + 0.00040212218, + -0.057794202, + -0.070333555, + 0.06377695, + -4.0873922e-33, + -0.0217928, + -0.079860926, + -0.013875922, + 0.14925155, + 0.025234098, + -0.042267527, + -0.006789101, + 0.054648004, + -0.09224933, + 0.008109618, + -0.038605478, + -0.117707536, + 0.012982382, + 0.034528743, + -0.017045766, + 0.01192032, + 0.012973965, + 0.042740148, + -0.017594555, + -0.018439855, + 0.06514173, + 0.040521882, + -0.022523073, + -0.060915224, + -0.018601585, + 0.011646964, + 0.0141018815, + -0.0676442, + 0.085437365, + 0.030129185, + 0.010850847, + -0.054872133, + -0.024110869, + 0.04832469, + 0.0074957223, + 0.013342751, + 0.024545655, + -0.00593543, + -0.04560701, + -0.0048439344, + 0.004394637, + -0.0023842545, + 0.013562894, + -0.016870767, + 0.06960542, + -0.077338316, + 0.020594154, + 0.004850868, + 0.055702493, + 0.013107641, + -0.011738689, + 0.04095329, + 0.0074854614, + 0.04204865, + 0.010375211, + 0.019378148, + 0.011061705, + 0.01726371, + 0.018246066, + 0.07732507, + 0.019622408, + 0.052688163, + -0.058638565, + 0.039727792, + -0.050275218, + -0.04894181, + -0.05262661, + -0.09227883, + 0.07558117, + 0.08100475, + -0.022263734, + -0.04214191, + 0.056570332, + 0.02357359, + 0.0015351619, + -0.049823847, + 0.0023157697, + 0.028624237, + -0.06897604, + -0.047824685, + -0.04863061, + -0.07660466, + -0.03283358, + -0.045931168, + -0.05727989, + -0.08089162, + -0.008027813, + -0.09357923, + 0.05126201, + -0.058291912, + -0.00058476225, + 0.022253899, + -0.04685808, + -0.08969063, + 0.11958076, + 2.0447206e-33, + 0.012184043, + 0.08640385, + -0.023207484, + 0.0027744523, + -0.0010493582, + 0.034863044, + 0.07328646, + -0.049892753, + -0.041898787, + 0.13484605, + -0.00690132, + 0.0062357984, + 0.0591438, + -0.028874595, + 0.09140647, + -0.018482381, + 0.0077092745, + -0.044212285, + -0.025144871, + -0.014995891, + -0.03540694, + 0.12411378, + 0.13117358, + 0.081000485, + -0.033294227, + 0.0039907615, + 0.026457148, + 0.026615122, + 0.017333155, + -0.0036460846, + 0.035482634, + 0.059582442, + -0.12458558, + 0.021935958, + 0.025609804, + -0.11062111, + 0.096059345, + -0.06729404, + -0.011844103, + 0.042349346, + 0.03789521, + 0.10581876, + 0.007365172, + 0.066275194, + 0.02294345, + 0.049393825, + 0.14640132, + -0.0067232805, + 0.004346095, + -0.029184747, + -0.009045802, + -0.086417, + 0.03588149, + 0.003007588, + -0.029339395, + 0.070202544, + 0.014933954, + 0.02831331, + -0.04035844, + 0.019160643, + 0.015603886, + 0.028645555, + -0.01953373, + -0.018291809, + -0.005431855, + -0.09320857, + -0.06113579, + 0.038820617, + 0.027979009, + 0.034132123, + -0.027506083, + 0.010690486, + -0.0551807, + -0.07381125, + 0.02152818, + -0.015417321, + -0.024984676, + -0.0047469, + 0.030462446, + -0.024068687, + 0.034130465, + -0.010350399, + -0.012667777, + 0.03628245, + -0.004432098, + -0.014948573, + 0.027915701, + 0.0978373, + -0.026430307, + -0.005174212, + -0.019117763, + 0.062028185, + 0.052109554, + 0.0378246, + 0.012581808, + -1.7055598e-08, + -0.050023284, + -0.08912732, + -0.0035682702, + -0.015776077, + -0.021857934, + 0.07185828, + -0.050184846, + -0.010655182, + -0.030601466, + -0.015778068, + 0.01321684, + -0.0025456804, + -0.042094428, + 0.009284693, + -0.041169193, + -0.029597968, + 0.0022024116, + -0.03303234, + -0.05039899, + -0.021473281, + -0.0068473304, + 0.008506351, + 0.035692476, + 0.025189023, + -0.016516164, + 0.049185548, + 0.018324668, + 0.049055174, + -0.05820532, + -0.015019503, + 0.04573769, + 0.049916334, + 0.02044857, + -0.05203969, + -0.0335851, + 0.061823603, + 0.11141345, + 0.077694215, + 0.0224589, + 0.0025537123, + -0.043906957, + 0.008579427, + -0.03620856, + 0.029681833, + -0.017270379, + -0.094624266, + -0.05785328, + -0.06581307, + -0.06124199, + -0.10454261, + -0.029261446, + 0.0013341395, + 0.0060936743, + 0.040794034, + -0.036677115, + 0.016793394, + 0.0052748835, + 0.03099207, + -0.054484233, + 0.0048635365, + 0.07086335, + 0.066848375, + 0.017699955, + -0.029221617 ] ] } diff --git a/tests/integration/recordings/responses/44fb9cf5875f.json b/tests/integration/recordings/responses/44fb9cf5875f.json index 462bd3a0c..c7b0333f2 100644 --- a/tests/integration/recordings/responses/44fb9cf5875f.json +++ b/tests/integration/recordings/responses/44fb9cf5875f.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:49.398099Z", + "created_at": "2025-07-31T17:59:42.166585642Z", "done": true, "done_reason": "stop", - "total_duration": 2251794875, - "load_duration": 108848667, + "total_duration": 9490295253, + "load_duration": 42349084, "prompt_eval_count": 20, - "prompt_eval_duration": 82008917, + "prompt_eval_duration": 545470166, "eval_count": 51, - "eval_duration": 2060141416, + "eval_duration": 8901928284, "response": "It seems like you're trying to test the system, but I'm not sure what specific functionality or feature you'd like to test. Could you please provide more context or clarify what you're looking for? I'll do my best to assist you!", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/477f8946bf7d.json b/tests/integration/recordings/responses/477f8946bf7d.json index d4a7f4de7..5840b0c40 100644 --- a/tests/integration/recordings/responses/477f8946bf7d.json +++ b/tests/integration/recordings/responses/477f8946bf7d.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 32336042, - "load_duration": 24852000, + "total_duration": 22476443, + "load_duration": 7010939, "prompt_eval_count": 21, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.07642644, - 0.0213101, - -0.03612849, - -0.0012144424, - -0.048599217, - -0.13194773, - -0.084226094, - 0.059389386, - -0.0617182, - -0.009323243, - -0.08099486, - 0.055514984, - 0.052610602, - 0.026061919, - 0.063071534, - -0.062316332, - -0.065115415, - -0.022351492, - 0.017378356, - -0.11605584, - -0.036349725, - 0.0404155, - -0.0325302, - -0.01770141, - 0.05722761, - 0.012393438, - -0.018529164, - -0.030017126, - 0.002365914, - 0.0066701965, - -0.08862459, - 0.0779319, - 0.03702611, - 0.029523117, - -0.01977821, - 0.05424799, - -0.00074063655, - -0.08949148, - -0.05312112, - -0.012703181, - -0.08622611, - 0.07689996, - -0.038602136, - -0.011616902, - -0.03234132, - -0.0073969415, - -0.024779495, - -0.067999884, - -0.03039565, - -0.025974417, - -0.09690519, - 0.009931951, - -0.05362519, - -0.09107193, - -0.009222061, - -0.008804084, - 0.048185978, - -0.003329437, - -0.0058579347, - -0.13306528, - -0.09721703, - 0.013474277, - 0.047286008, - 0.06279936, - -0.01582815, - -0.03771013, - -0.01651892, - 0.029905442, - 0.09326656, - -0.06746783, - -0.13385954, - -0.020873511, - -0.02586237, - 0.11623731, - 0.030632136, - -0.10494776, - 0.03905967, - -0.010701787, - -0.0014734551, - 0.020711906, - 0.0017687598, - 0.027797814, - -0.078500465, - 0.10791581, - 0.02910256, - -0.05398749, - 0.030513834, - 0.07001416, - -0.034323946, - 0.00986597, - 0.034644563, - -0.04232179, - 0.065106474, - 0.026648693, - -0.032122962, - 0.07616709, - 0.020026332, - -0.030642457, - -0.07188906, - 0.027189687, - -0.018678213, - -0.05416582, - 0.07488992, - 0.017753933, - 0.03386007, - 0.02414506, - 0.09077034, - -0.052096054, - 0.040722203, - -0.018450806, - -0.012474094, - -0.06403705, - -0.023205942, - -0.061878704, - 0.053436812, - 0.047876816, - -0.010608645, - 0.07852118, - 0.03579911, - 0.027097313, - 0.022424318, - -0.004912598, - -0.02455264, - 0.003700777, - 0.00039888592, - -0.008842094, - 0.009365857, - 2.05052e-34, - -0.03236592, - -0.024301885, - 0.027186498, - 0.021633558, - 0.06519107, - -0.019539308, - 0.05306087, - 0.007985293, - -0.03927361, - -0.020062907, - 0.008070545, - 0.02382429, - 0.015006528, - 0.1128094, - 0.06113956, - -0.011911169, - 0.016901307, - 0.045509744, - 0.0013988831, - 0.00907712, - 0.01314859, - -0.012022324, - 0.027043821, - 0.0071581583, - 0.022573117, - -0.013721936, - -0.004378743, - -0.0007087661, - 0.033585846, - 0.011227843, - -0.05136015, - -0.0739591, - -0.03094639, - 0.01957863, - -0.010360539, - -0.0029881562, - -0.00480912, - -0.10446798, - 0.034694213, - -0.02424012, - -0.047155295, - 0.035451673, - 0.037169226, - -0.016986743, - 0.0056092087, - 0.05057555, - -0.008601115, - 0.0060349177, - -0.12273999, - 0.036871877, - -0.022267655, - -0.009739047, - 0.075974636, - 0.08902226, - 0.01647873, - 0.044345584, - 0.06792565, - 0.06456903, - -0.050189856, - -0.0016995457, - -0.00090498856, - 0.09925942, - 0.09253569, - -0.011321612, - 0.050309792, - 0.07697773, - 0.0100068, - 0.101032645, - 0.03268899, - 0.06433435, - -0.044524822, - 0.03860177, - -0.019314477, - 0.037440598, - -0.0017394378, - 0.011816814, - 0.011359969, - -0.1040215, - 0.06984421, - 0.01910163, - -0.028409261, - -0.013704911, - 0.048502754, - -0.015429918, - -0.03423058, - -0.055616368, - 0.005001686, - 0.026054256, - -0.0007700968, - -0.0041726283, - -0.0640977, - -0.05985385, - 0.0813829, - 0.014288322, - -0.038147252, - -2.1576616e-33, - -0.027279941, - -0.034765568, - -0.02465107, - 0.026859807, - -0.090699576, - -0.045698144, - 0.013666582, - 0.002109106, - 0.054007426, - 0.032838397, - -0.029939773, - -0.058843046, - 0.09825693, - 0.03251322, - 0.109977886, - 0.020682266, - -0.0958973, - 0.0005566991, - 0.0018037638, - 0.017544486, - -0.06843023, - 0.06435102, - -0.050149646, - -0.048880838, - -0.027535524, - -0.014993001, - -0.1210176, - -0.04412877, - -0.011025324, - 0.058610573, - -0.007498303, - 0.038722932, - -0.07025986, - 0.030281536, - 0.055707317, - -0.001162887, - 0.01707519, - -0.042081844, - -0.016578361, - -0.025714336, - 0.117893435, - 0.04196084, - 0.064787276, - 0.046081997, - 0.014950138, - 0.030026693, - -0.039077066, - 0.087156676, - -0.012328571, - -0.035646956, - -0.048145168, - 0.041394625, - 0.038984135, - -0.025188481, - -0.028836856, - -0.02917782, - 0.029690607, - 0.051454436, - -0.08629761, - -0.06921346, - -0.07273269, - -0.05952071, - 0.0050034616, - 0.025693603, - -0.022103382, - 0.024972659, - -0.09724792, - 0.0062089814, - -0.04963219, - -0.13054384, - 0.124669954, - -0.01361085, - -0.022798477, - 0.039057832, - -0.07550591, - 0.049364913, - 0.0007779102, - 0.004692535, - -0.040757872, - 0.06355995, - 0.110190175, - 0.02015945, - -0.048807338, - 0.05842704, - -0.066375315, - 0.026938869, - -0.062775925, - -0.014049011, - 0.023343485, - 0.02358394, - -0.002172394, - 0.07766165, - 0.031056313, - 0.020171564, - -0.020073414, - -2.4317085e-08, - 0.020261949, - -0.008623839, - 0.0621209, - -0.008334477, - 0.02526615, - 0.08902315, - -0.007958188, - -0.018911751, - -0.035572145, - 0.06189234, - -0.017249323, - -0.030186126, - -0.10225455, - -0.06522741, - -0.004033112, - 0.10897627, - -0.02168822, - -0.053784374, - 0.011841631, - 0.052263785, - 0.058334205, - 0.0052479547, - -0.06017166, - 0.08723854, - -0.08275336, - -0.040676847, - 0.065786876, - 0.028317772, - -0.012168614, - -0.07196286, - 0.014588226, - -0.03231537, - 0.0028357722, - 0.03868031, - 0.055439528, - -0.015238348, - 0.05482384, - -0.025080629, - -0.033771332, - 0.0030752022, - -0.037511814, - 0.015122315, - 0.02292684, - 0.012024873, - 0.03559873, - 0.006865039, - -0.04049267, - -0.049685854, - -0.05455341, - -0.073071465, - -0.024902396, - -0.002133957, - -0.013212662, - -0.06657236, - 0.023245512, - 0.046919, - -0.13278763, - -0.011092663, - -0.023939205, - 0.043182902, - 0.024406029, - 0.06922961, - 0.15658055, - 0.017658537 + -0.07644922, + 0.021318993, + -0.036126964, + -0.0012044739, + -0.048612095, + -0.13192746, + -0.08423256, + 0.059381723, + -0.061683927, + -0.009348091, + -0.081030406, + 0.05550002, + 0.052616827, + 0.026079413, + 0.063077666, + -0.062315546, + -0.06512819, + -0.02229665, + 0.017397532, + -0.1160329, + -0.036332168, + 0.040402204, + -0.032573264, + -0.017708758, + 0.05723005, + 0.012369807, + -0.018542344, + -0.030028714, + 0.0023615656, + 0.006659271, + -0.0885923, + 0.07790947, + 0.037012804, + 0.029490992, + -0.019760288, + 0.05421732, + -0.00073781994, + -0.08950901, + -0.053158604, + -0.012716266, + -0.08623249, + 0.07690697, + -0.038633663, + -0.011597453, + -0.032314006, + -0.0074278843, + -0.024758225, + -0.06797268, + -0.03035838, + -0.025995128, + -0.096888065, + 0.0099435905, + -0.053624775, + -0.09104344, + -0.009231492, + -0.008822432, + 0.04818155, + -0.0033450315, + -0.0058557615, + -0.13309938, + -0.09719051, + 0.013506018, + 0.04729355, + 0.06281491, + -0.01586599, + -0.037687704, + -0.016521314, + 0.029923148, + 0.093276426, + -0.067442276, + -0.13386938, + -0.020885147, + -0.025864335, + 0.116227925, + 0.030623658, + -0.10494704, + 0.03906256, + -0.010738701, + -0.0014873091, + 0.020708071, + 0.0017483904, + 0.027790338, + -0.07846251, + 0.10790454, + 0.029114574, + -0.053953465, + 0.030514322, + 0.07002214, + -0.0343377, + 0.009869935, + 0.034672886, + -0.042333603, + 0.06509199, + 0.02666166, + -0.032117628, + 0.07613336, + 0.020031841, + -0.030653432, + -0.07187661, + 0.027188664, + -0.018698178, + -0.054159895, + 0.074888855, + 0.017748112, + 0.03388562, + 0.024155568, + 0.09078823, + -0.052107602, + 0.04071798, + -0.01846267, + -0.0124565, + -0.06405017, + -0.023211012, + -0.06188541, + 0.05343985, + 0.047868032, + -0.010622221, + 0.07852332, + 0.035839524, + 0.027102223, + 0.02240619, + -0.004891384, + -0.02456285, + 0.0037151189, + 0.00039547117, + -0.008838611, + 0.009371476, + 2.0515453e-34, + -0.032390445, + -0.024334554, + 0.027150098, + 0.021630002, + 0.06519911, + -0.019550668, + 0.053052407, + 0.007951343, + -0.039268915, + -0.020086676, + 0.0080776885, + 0.02382864, + 0.015012353, + 0.11279827, + 0.06113922, + -0.011914555, + 0.016920203, + 0.045502547, + 0.001394539, + 0.009074133, + 0.013133291, + -0.012016043, + 0.027050933, + 0.0071878177, + 0.022549521, + -0.013711725, + -0.004366378, + -0.0007136731, + 0.033571508, + 0.01122357, + -0.051396187, + -0.07395165, + -0.030959165, + 0.019595258, + -0.010384256, + -0.0029798083, + -0.004823488, + -0.10445505, + 0.03467776, + -0.024233725, + -0.047162082, + 0.035441577, + 0.03716666, + -0.01702174, + 0.0056008953, + 0.050594125, + -0.008599615, + 0.0060342806, + -0.12273874, + 0.036802854, + -0.022243306, + -0.009694798, + 0.07591922, + 0.08904486, + 0.016491221, + 0.044297636, + 0.06791793, + 0.06454211, + -0.05018115, + -0.0016970917, + -0.0009100337, + 0.09926223, + 0.09258295, + -0.011353339, + 0.05032501, + 0.07698045, + 0.009997087, + 0.10103169, + 0.032655906, + 0.06433115, + -0.04454715, + 0.03860544, + -0.019333873, + 0.037454415, + -0.001721842, + 0.011826793, + 0.011386428, + -0.10405232, + 0.069838874, + 0.01912115, + -0.028386243, + -0.013710603, + 0.048529655, + -0.015396224, + -0.03423858, + -0.055645425, + 0.0049964655, + 0.026062267, + -0.0007718523, + -0.0042009777, + -0.06409095, + -0.059850696, + 0.08137787, + 0.014278817, + -0.038195916, + -2.1589785e-33, + -0.027295526, + -0.034773894, + -0.024641098, + 0.026864044, + -0.090734534, + -0.045691974, + 0.013699863, + 0.0021261072, + 0.05404863, + 0.03285422, + -0.029929286, + -0.05883433, + 0.09826083, + 0.032517377, + 0.10999013, + 0.020698903, + -0.09591734, + 0.0005467174, + 0.0018373779, + 0.017558018, + -0.06844123, + 0.06432574, + -0.050150894, + -0.048873555, + -0.027538775, + -0.014966375, + -0.12098801, + -0.044132344, + -0.011028691, + 0.058583282, + -0.007502001, + 0.038751014, + -0.07027614, + 0.030262535, + 0.055714924, + -0.0011363372, + 0.017083727, + -0.04206832, + -0.016568454, + -0.025682067, + 0.11789456, + 0.04198409, + 0.06481419, + 0.04607849, + 0.014978292, + 0.03001545, + -0.03910612, + 0.08715018, + -0.012336109, + -0.03564661, + -0.04812303, + 0.04141488, + 0.03897653, + -0.025203561, + -0.028823132, + -0.029183073, + 0.029703744, + 0.051458877, + -0.086284295, + -0.06920673, + -0.07273957, + -0.059528224, + 0.0049837893, + 0.025650585, + -0.022120077, + 0.024956776, + -0.0972337, + 0.0061748885, + -0.04960218, + -0.1305334, + 0.12471198, + -0.013604223, + -0.022810707, + 0.03906276, + -0.075510286, + 0.049388453, + 0.0008171022, + 0.004682814, + -0.04076038, + 0.06357199, + 0.1101723, + 0.02017848, + -0.04873689, + 0.0584356, + -0.06637572, + 0.026938135, + -0.06277571, + -0.014051585, + 0.023363862, + 0.023567248, + -0.0021611133, + 0.07768197, + 0.031047512, + 0.020165777, + -0.02006235, + -2.4314515e-08, + 0.020272322, + -0.008597304, + 0.06210691, + -0.008328929, + 0.025253547, + 0.089005895, + -0.007974264, + -0.018915428, + -0.035587803, + 0.0618582, + -0.017240847, + -0.030206975, + -0.10226169, + -0.065235354, + -0.0040415884, + 0.109014235, + -0.021687664, + -0.053811464, + 0.011844342, + 0.052247472, + 0.0583252, + 0.0052674375, + -0.060206596, + 0.08722171, + -0.082785375, + -0.040664576, + 0.06578738, + 0.0282874, + -0.012157491, + -0.07194093, + 0.014612263, + -0.032293286, + 0.002835932, + 0.038650285, + 0.05545503, + -0.015265302, + 0.054820932, + -0.025081055, + -0.03375923, + 0.0030857057, + -0.037500594, + 0.0151155675, + 0.022939838, + 0.012013316, + 0.035608154, + 0.006845111, + -0.040476773, + -0.049682803, + -0.05456417, + -0.07305824, + -0.02487007, + -0.0021548867, + -0.013222908, + -0.066566885, + 0.023217667, + 0.04692784, + -0.13282707, + -0.011092963, + -0.023976086, + 0.04316705, + 0.02437864, + 0.06919968, + 0.15656404, + 0.017655756 ] ] } diff --git a/tests/integration/recordings/responses/48d2fb183a2a.json b/tests/integration/recordings/responses/48d2fb183a2a.json index c54bf5add..5b72b9745 100644 --- a/tests/integration/recordings/responses/48d2fb183a2a.json +++ b/tests/integration/recordings/responses/48d2fb183a2a.json @@ -67,15 +67,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:48.260787Z", + "created_at": "2025-07-31T17:55:30.590990089Z", "done": true, "done_reason": "stop", - "total_duration": 3136253292, - "load_duration": 81917125, + "total_duration": 32260966927, + "load_duration": 43969344, "prompt_eval_count": 259, - "prompt_eval_duration": 540110750, + "prompt_eval_duration": 21518828528, "eval_count": 60, - "eval_duration": 2513196708, + "eval_duration": 10697338625, "response": "{\n \"first_name\": \"Michael\",\n \"last_name\": \"Jordan\",\n \"year_of_birth\": 1963,\n \"nba_stats\": {\n \"year_for_draft\": 1984,\n \"num_seasons_in_nba\": 15\n }\n}", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/4a3a4447b16b.json b/tests/integration/recordings/responses/4a3a4447b16b.json index f1e91d7db..45d4cfe51 100644 --- a/tests/integration/recordings/responses/4a3a4447b16b.json +++ b/tests/integration/recordings/responses/4a3a4447b16b.json @@ -14,7 +14,7 @@ "models": [ { "model": "nomic-embed-text:latest", - "modified_at": "2025-07-29T16:46:26.304701-07:00", + "modified_at": "2025-07-31T17:43:12.217040Z", "digest": "0a109f422b47e3a30ba2b10eca18548e944e8a23073ee3f3e947efcf3c45e59f", "size": 274302450, "details": { @@ -30,7 +30,7 @@ }, { "model": "llama-guard3:1b", - "modified_at": "2025-07-25T14:39:44.978630-07:00", + "modified_at": "2025-07-31T04:44:58Z", "digest": "494147e06bf99e10dbe67b63a07ac81c162f18ef3341aa3390007ac828571b3b", "size": 1600181919, "details": { @@ -46,7 +46,7 @@ }, { "model": "all-minilm:l6-v2", - "modified_at": "2025-07-24T15:15:11.129290-07:00", + "modified_at": "2025-07-31T04:42:15Z", "digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef", "size": 45960996, "details": { @@ -60,57 +60,9 @@ "quantization_level": "F16" } }, - { - "model": "llama3.2:1b", - "modified_at": "2025-07-17T22:02:24.953208-07:00", - "digest": "baf6a787fdffd633537aa2eb51cfd54cb93ff08e28040095462bb63daf552878", - "size": 1321098329, - "details": { - "parent_model": "", - "format": "gguf", - "family": "llama", - "families": [ - "llama" - ], - "parameter_size": "1.2B", - "quantization_level": "Q8_0" - } - }, - { - "model": "all-minilm:latest", - "modified_at": "2025-06-03T16:50:10.946583-07:00", - "digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef", - "size": 45960996, - "details": { - "parent_model": "", - "format": "gguf", - "family": "bert", - "families": [ - "bert" - ], - "parameter_size": "23M", - "quantization_level": "F16" - } - }, - { - "model": "llama3.2:3b", - "modified_at": "2025-05-01T11:15:23.797447-07:00", - "digest": "a80c4f17acd55265feec403c7aef86be0c25983ab279d83f3bcd3abbcb5b8b72", - "size": 2019393189, - "details": { - "parent_model": "", - "format": "gguf", - "family": "llama", - "families": [ - "llama" - ], - "parameter_size": "3.2B", - "quantization_level": "Q4_K_M" - } - }, { "model": "llama3.2:3b-instruct-fp16", - "modified_at": "2025-04-30T15:33:48.939665-07:00", + "modified_at": "2025-07-31T04:42:05Z", "digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d", "size": 6433703586, "details": { diff --git a/tests/integration/recordings/responses/50340cd4d253.json b/tests/integration/recordings/responses/50340cd4d253.json index dc55c4a28..b6210ee79 100644 --- a/tests/integration/recordings/responses/50340cd4d253.json +++ b/tests/integration/recordings/responses/50340cd4d253.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:23.326652Z", + "created_at": "2025-07-31T17:45:36.538134237Z", "done": true, "done_reason": "stop", - "total_duration": 208897666, - "load_duration": 52693583, + "total_duration": 2796712434, + "load_duration": 44874416, "prompt_eval_count": 216, - "prompt_eval_duration": 144231750, + "prompt_eval_duration": 2709689823, "eval_count": 2, - "eval_duration": 11141958, + "eval_duration": 41625474, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/545d86510a80.json b/tests/integration/recordings/responses/545d86510a80.json index 11e698aac..cbc2ad98b 100644 --- a/tests/integration/recordings/responses/545d86510a80.json +++ b/tests/integration/recordings/responses/545d86510a80.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.454595Z", + "created_at": "2025-07-31T17:48:08.291262469Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.499184Z", + "created_at": "2025-07-31T17:48:08.482518171Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.547038Z", + "created_at": "2025-07-31T17:48:08.67196398Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.593861Z", + "created_at": "2025-07-31T17:48:08.854871756Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.641073Z", + "created_at": "2025-07-31T17:48:09.039123774Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.688437Z", + "created_at": "2025-07-31T17:48:09.22769491Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.736972Z", + "created_at": "2025-07-31T17:48:09.41528678Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.781427Z", + "created_at": "2025-07-31T17:48:09.616301386Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.827155Z", + "created_at": "2025-07-31T17:48:09.800460611Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.8731Z", + "created_at": "2025-07-31T17:48:09.986683207Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.916794Z", + "created_at": "2025-07-31T17:48:10.17267686Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:46.965672Z", + "created_at": "2025-07-31T17:48:10.358862363Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,15 +238,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:47.013995Z", + "created_at": "2025-07-31T17:48:10.546042909Z", "done": true, "done_reason": "stop", - "total_duration": 711746792, - "load_duration": 79927208, + "total_duration": 4336248519, + "load_duration": 45632644, "prompt_eval_count": 408, - "prompt_eval_duration": 69131500, + "prompt_eval_duration": 2034392149, "eval_count": 13, - "eval_duration": 561561375, + "eval_duration": 2255671980, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/554de3cd986f.json b/tests/integration/recordings/responses/554de3cd986f.json index 0d818a584..644a689ff 100644 --- a/tests/integration/recordings/responses/554de3cd986f.json +++ b/tests/integration/recordings/responses/554de3cd986f.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.730544Z", + "created_at": "2025-07-31T17:44:35.603264527Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.776362Z", + "created_at": "2025-07-31T17:44:35.78602068Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.818895Z", + "created_at": "2025-07-31T17:44:35.96866699Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.861401Z", + "created_at": "2025-07-31T17:44:36.150414646Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.904334Z", + "created_at": "2025-07-31T17:44:36.333587619Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.946942Z", + "created_at": "2025-07-31T17:44:36.518108567Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:11.988807Z", + "created_at": "2025-07-31T17:44:36.701519577Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.030657Z", + "created_at": "2025-07-31T17:44:36.887968019Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.072804Z", + "created_at": "2025-07-31T17:44:37.070912948Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.115051Z", + "created_at": "2025-07-31T17:44:37.255400593Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.157568Z", + "created_at": "2025-07-31T17:44:37.435774155Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.200298Z", + "created_at": "2025-07-31T17:44:37.614737276Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.243723Z", + "created_at": "2025-07-31T17:44:37.79509715Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.286287Z", + "created_at": "2025-07-31T17:44:37.976454651Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.329284Z", + "created_at": "2025-07-31T17:44:38.157671144Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.371308Z", + "created_at": "2025-07-31T17:44:38.337351473Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.413286Z", + "created_at": "2025-07-31T17:44:38.517835532Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.455335Z", + "created_at": "2025-07-31T17:44:38.700974215Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,15 +346,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:12.497365Z", + "created_at": "2025-07-31T17:44:38.884515807Z", "done": true, "done_reason": "stop", - "total_duration": 1007895792, - "load_duration": 128942250, + "total_duration": 6189734480, + "load_duration": 41239564, "prompt_eval_count": 376, - "prompt_eval_duration": 109187542, + "prompt_eval_duration": 2865253062, "eval_count": 19, - "eval_duration": 768957125, + "eval_duration": 3282706083, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/56ac6a7c6df0.json b/tests/integration/recordings/responses/56ac6a7c6df0.json index e56ede21a..06fe4a11a 100644 --- a/tests/integration/recordings/responses/56ac6a7c6df0.json +++ b/tests/integration/recordings/responses/56ac6a7c6df0.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 54720667, - "load_duration": 26019750, + "total_duration": 13596464, + "load_duration": 5559519, "prompt_eval_count": 5, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.03427073, - 0.090051405, - -0.11458989, - 0.0021456745, - 0.059038658, - -0.027524853, - -0.020602634, - 0.03373726, - -0.038729247, - 0.026002944, - 0.11481002, - 0.027119067, - -0.015927644, - -0.021832926, - -0.046713773, - -0.0463825, - -0.074167565, - -0.0528447, - -0.028117927, - 0.06325688, - 0.029135453, - 0.047131006, - -0.052675154, - -0.005349263, - 0.030659368, - 0.017706472, - -0.01687267, - 0.08681507, - -0.014155131, - -0.0838676, - 0.020020565, - 0.07115838, - 0.08365558, - 0.030919788, - 0.11829893, - 0.028751066, - 0.069536895, - -0.017295403, - -0.005784813, - 0.005809313, - 0.0012009157, - -0.0653044, - 0.0373506, - 0.018565746, - -0.0034945607, - -0.0011305016, - -0.029752811, - -0.021266408, - 0.0058016903, - -0.035597492, - -0.03722647, - 0.012373253, - -0.066935256, - -0.023148224, - 0.056864377, - 0.0014741909, - 0.014408296, - -0.017165763, - 0.009236472, - 0.06087921, - 0.024628488, - 0.03699286, - -0.050610077, - 0.05173448, - 0.10159555, - 0.008507267, - -0.04803921, - -0.013024803, - 0.03110457, - -0.16593884, - -0.1410075, - 0.009813814, - -0.025974236, - 0.05233053, - -0.0078903325, - 0.00788491, - -0.08471812, - -0.044507448, - 0.054161046, - -0.0704361, - -0.05769206, - -0.100796975, - 0.02182441, - 0.022125391, - 0.0071617346, - 0.13063926, - 0.080232956, - -0.004421626, - -0.018768508, - 0.0076132733, - -0.03163366, - 0.031986494, - -0.022168567, - 0.03073627, - -0.023798423, - 0.06954045, - 0.016659362, - 0.009536805, - 0.027459558, - 0.102133445, - 0.021457382, - -0.021377807, - 0.015131543, - 0.039423607, - -0.09434147, - -0.11544392, - 0.09468138, - -0.011155598, - 0.07266597, - -0.03601087, - -0.011743829, - -0.06654009, - -0.03470551, - -0.10300434, - 0.03020924, - -0.06319472, - -0.0908424, - 0.04116676, - -0.033686537, - 0.045706224, - 0.07134009, - -0.031778418, - -0.059655976, - -0.017215038, - -0.03229557, - -0.058579948, - 0.06733934, - -5.023814e-33, - -0.0058283503, - -0.0719842, - -0.009296622, - 0.09659216, - 0.03709538, - -0.03478395, - -0.004713233, - 0.016686605, - -0.09859812, - 0.00547005, - -0.014113569, - -0.0840751, - 0.0027168505, - 0.04445616, - -0.012728728, - 0.034566686, - -0.0006014651, - 0.06319148, - -0.026799418, - -0.013500979, - 0.024169419, - 0.015417236, - -0.04135526, - -0.055208974, - -0.06455241, - 0.03148543, - -0.0073052812, - -0.03945437, - 0.059831504, - 0.026674163, - 0.01396753, - -0.038841277, - -0.048514687, - 0.01756627, - 0.020964677, - 0.035239976, - 0.0115498835, - -0.00846713, - -0.044673763, - 0.014640657, - 5.2045852e-05, - -0.04694704, - 0.02703366, - 0.006635295, - 0.064396136, - -0.044757996, - -0.026173549, - -0.016282372, - 0.05521396, - 0.014104745, - -0.008479494, - 0.04204778, - 0.05049772, - 0.021629427, - 0.011260506, - 0.04858872, - 0.017662494, - -0.005005865, - 0.0019118759, - 0.06333162, - 0.035875723, - 0.03504778, - -0.06642375, - 0.008791644, - -0.027326671, - -0.05987137, - -0.0272001, - -0.08728625, - 0.112434424, - 0.05879801, - -0.041698616, - -0.06924583, - 0.06434144, - 0.01583225, - -0.027750073, - -0.037574448, - -0.011715211, - 0.0694801, - -0.07104981, - -0.039085716, - -0.043068763, - -0.11208956, - -0.030723054, - -0.063793585, - -0.03527373, - -0.06119042, - -0.01526633, - -0.10094421, - 0.047486804, - -0.08320468, - -0.0029513796, - 0.0131224785, - -0.056690685, - -0.057956036, - 0.06140136, - 2.7669969e-33, - 0.0036719525, - 0.06695694, - -0.05591421, - 0.025166295, - 0.014735592, - 0.03381445, - 0.09345791, - -0.01053347, - -0.046693947, - 0.14254177, - -0.015430197, - 0.0066938214, - 0.07679359, - -0.045779705, - 0.07989786, - 0.0036165903, - 0.023604553, - -0.06533708, - -0.04253485, - -0.025912313, - -0.0748119, - 0.10020777, - 0.12578633, - 0.06409652, - -0.016682886, - 0.01406972, - 0.025274348, - 0.0017218525, - -0.013340701, - 0.01172295, - 0.03772902, - 0.040607873, - -0.120578945, - 0.024344057, - 0.03439985, - -0.10167353, - 0.11863072, - -0.03571693, - -0.0126576, - 0.022622129, - 0.039235484, - 0.10625315, - 0.0106492825, - 0.076503076, - 0.02088746, - 0.06468519, - 0.08582322, - -0.032148413, - 0.04359905, - 0.011070053, - 0.023209164, - -0.06709916, - 0.055355705, - -0.008128262, - -0.026921155, - 0.076995976, - -0.011614669, - 0.044967294, - -0.02459807, - 0.020910041, - -0.0016746842, - 0.02905443, - -0.03898753, - -0.01360213, - -0.019878393, - -0.057056017, - -0.014543598, - 0.010161744, - 0.016893594, - 0.011981163, - 0.019902436, - 0.019194229, - -0.06551642, - -0.050247267, - 0.050837662, - -0.075614415, - -0.018767305, - -0.012229684, - 0.0019464786, - -0.0035209567, - 0.0699799, - -0.02925182, - -0.008455151, - 0.04742619, - -0.0004527954, - -0.014011262, - -0.0035493495, - 0.08439228, - -0.001586065, - 0.0016962147, - -0.023180604, - 0.059889086, - 0.019616995, - 0.05435093, - 0.012301163, - -1.5289881e-08, - -0.038103975, - -0.084179275, - -0.013605872, - -0.03277629, - -0.020995136, - 0.08924277, - 0.005438667, - -0.07047066, - -0.03966912, - -0.018226335, - 0.05716885, - -0.026391266, - -0.09881308, - 0.017511, - -0.01952465, - -0.06237397, - -0.019553065, - -0.0112019945, - -0.030052405, - 0.010624359, - -0.005598304, - 0.05326868, - 0.044162616, - 0.025812192, - 0.0059228353, - 0.059632093, - 0.06885661, - 0.08894283, - -0.06225795, - -0.038893122, - 0.028817136, - 0.08772772, - 0.017759481, - -0.050048865, - -0.0009810333, - 0.1297453, - 0.083138496, - 0.08161095, - 0.011747931, - 0.006871316, - -0.07277484, - -0.0020051182, - -0.018357608, - 0.008882652, - -0.03823878, - -0.09057624, - -0.06433315, - -0.04256367, - -0.030856675, - -0.09314087, - -0.043470908, - 0.012043298, - -9.8401986e-05, - 0.040246293, - -0.04912119, - 0.014575804, - 0.017479645, - -0.00515073, - -0.033331197, - 0.0075505474, - 0.07488009, - 0.06460031, - 0.044803377, - -0.028485151 + -0.03428553, + 0.09004888, + -0.11458894, + 0.0021527493, + 0.05904814, + -0.027502729, + -0.020575836, + 0.03378457, + -0.038715836, + 0.026000869, + 0.11478867, + 0.027114356, + -0.015911782, + -0.021798763, + -0.04674167, + -0.046405133, + -0.074190386, + -0.05286571, + -0.028126542, + 0.06323515, + 0.02913013, + 0.047108278, + -0.052707225, + -0.0053599635, + 0.03071732, + 0.017738523, + -0.016880909, + 0.08683748, + -0.01419749, + -0.083865836, + 0.020033062, + 0.071156204, + 0.083663985, + 0.030905709, + 0.11826464, + 0.02876898, + 0.06954055, + -0.017332977, + -0.005812741, + 0.0058015552, + 0.001208471, + -0.06535491, + 0.037350487, + 0.018552719, + -0.0034832722, + -0.001124515, + -0.029755933, + -0.021265727, + 0.0058143395, + -0.035625655, + -0.03724204, + 0.012374368, + -0.066953905, + -0.023154013, + 0.056864116, + 0.0014606857, + 0.014412622, + -0.017193878, + 0.009222129, + 0.060872346, + 0.024618814, + 0.03699705, + -0.050617803, + 0.051762927, + 0.10159892, + 0.008498099, + -0.04801456, + -0.012997251, + 0.031116826, + -0.1659354, + -0.14099391, + 0.009771681, + -0.025979118, + 0.052322462, + -0.007871116, + 0.007861781, + -0.08469375, + -0.04453351, + 0.054181393, + -0.07046408, + -0.057691414, + -0.10079021, + 0.02186296, + 0.022151157, + 0.0071818396, + 0.130646, + 0.08021881, + -0.0044269706, + -0.018767677, + 0.0076321233, + -0.031633127, + 0.031931527, + -0.022182738, + 0.030723765, + -0.023784049, + 0.069556564, + 0.016621906, + 0.009541423, + 0.027459256, + 0.102094576, + 0.021432728, + -0.021382928, + 0.015117344, + 0.039430253, + -0.09436079, + -0.11549412, + 0.094706915, + -0.011174707, + 0.07267626, + -0.03601918, + -0.011763209, + -0.066555545, + -0.034689933, + -0.10300218, + 0.030211166, + -0.06319931, + -0.09080848, + 0.041160528, + -0.03372365, + 0.04571954, + 0.07133777, + -0.03177294, + -0.059663862, + -0.017204959, + -0.032270484, + -0.05857379, + 0.067352176, + -5.0251458e-33, + -0.005811169, + -0.07199202, + -0.009300383, + 0.096577324, + 0.03708445, + -0.034742005, + -0.0047524897, + 0.016684553, + -0.098613314, + 0.005455344, + -0.014082916, + -0.08406552, + 0.0027243053, + 0.044460878, + -0.012708549, + 0.03457976, + -0.0005862883, + 0.063180104, + -0.026798664, + -0.013535706, + 0.024189947, + 0.01542626, + -0.041350108, + -0.055188444, + -0.06456418, + 0.031478163, + -0.007293317, + -0.03944318, + 0.05984358, + 0.02667509, + 0.013961637, + -0.038835857, + -0.0485192, + 0.017592456, + 0.02095435, + 0.035228003, + 0.011563164, + -0.008445899, + -0.044658076, + 0.014642002, + 5.8537742e-05, + -0.046962045, + 0.027041595, + 0.0066561843, + 0.06440716, + -0.04475169, + -0.026170205, + -0.016300367, + 0.0551575, + 0.014121041, + -0.008471725, + 0.04206057, + 0.050532088, + 0.021643365, + 0.011242044, + 0.048596855, + 0.017674237, + -0.0049935156, + 0.0019010587, + 0.06328416, + 0.03586134, + 0.035088714, + -0.06643235, + 0.008815076, + -0.027297651, + -0.059867114, + -0.027219879, + -0.08726865, + 0.11245166, + 0.05882553, + -0.041703966, + -0.06924601, + 0.064341605, + 0.015860816, + -0.027766522, + -0.037580114, + -0.011743611, + 0.06949358, + -0.07105207, + -0.039093148, + -0.043085232, + -0.11208843, + -0.030707585, + -0.06380492, + -0.03527061, + -0.06121885, + -0.015268978, + -0.100922786, + 0.04748757, + -0.083198026, + -0.0029790367, + 0.013129155, + -0.056719888, + -0.057915524, + 0.06138452, + 2.76823e-33, + 0.0036890432, + 0.06695775, + -0.055907723, + 0.025152251, + 0.014722569, + 0.033783082, + 0.09345767, + -0.010525945, + -0.04667415, + 0.14253052, + -0.015412643, + 0.006669673, + 0.07681041, + -0.04577685, + 0.079887144, + 0.0036023448, + 0.023597728, + -0.06528208, + -0.042549107, + -0.025877435, + -0.07481574, + 0.10019824, + 0.12577929, + 0.064089745, + -0.016686304, + 0.01409427, + 0.025257608, + 0.0017210066, + -0.013362902, + 0.011713427, + 0.037738074, + 0.04061518, + -0.12053303, + 0.024357164, + 0.03439261, + -0.10164916, + 0.11861079, + -0.035714135, + -0.012694357, + 0.022589708, + 0.039240547, + 0.106231034, + 0.010664901, + 0.07653826, + 0.020890983, + 0.06468378, + 0.08584671, + -0.03213069, + 0.0435966, + 0.011061552, + 0.023196135, + -0.067093305, + 0.055348866, + -0.008123861, + -0.026925996, + 0.07702015, + -0.01161366, + 0.045000453, + -0.02460899, + 0.020922417, + -0.0016905216, + 0.02905479, + -0.038986016, + -0.013623761, + -0.019841073, + -0.057056155, + -0.014542025, + 0.010135319, + 0.01689078, + 0.011984185, + 0.01991723, + 0.019205214, + -0.06552643, + -0.050277457, + 0.050829098, + -0.07556213, + -0.018830225, + -0.012219267, + 0.0019397368, + -0.0035257766, + 0.07000847, + -0.029260997, + -0.008443407, + 0.04745947, + -0.0004566185, + -0.014023967, + -0.0035412489, + 0.084373, + -0.0015863521, + 0.0016559219, + -0.02315912, + 0.059896436, + 0.019620532, + 0.054353774, + 0.012328795, + -1.5288656e-08, + -0.038075536, + -0.08422955, + -0.013584843, + -0.03280181, + -0.020946743, + 0.089246586, + 0.0054381313, + -0.070446074, + -0.039640933, + -0.018214736, + 0.057154264, + -0.02636421, + -0.09882496, + 0.01748733, + -0.019522436, + -0.062379386, + -0.019562414, + -0.011194671, + -0.03005611, + 0.010603683, + -0.0055661174, + 0.053237215, + 0.044146214, + 0.02581067, + 0.0058922465, + 0.059643324, + 0.06885044, + 0.08893949, + -0.062240638, + -0.038882267, + 0.028826952, + 0.08772289, + 0.017748002, + -0.05002541, + -0.0009826778, + 0.1297349, + 0.08316373, + 0.08159867, + 0.01174721, + 0.0068597244, + -0.072790615, + -0.0019851946, + -0.018349772, + 0.008917563, + -0.038223803, + -0.09057707, + -0.064334795, + -0.042570896, + -0.030840263, + -0.09316567, + -0.043464772, + 0.01205224, + -8.986558e-05, + 0.0402598, + -0.04913751, + 0.014560711, + 0.017480103, + -0.0051642335, + -0.033332866, + 0.007570478, + 0.07488999, + 0.06458834, + 0.0448589, + -0.02847636 ] ] } diff --git a/tests/integration/recordings/responses/61be36ad8ccd.json b/tests/integration/recordings/responses/61be36ad8ccd.json index 56719cabc..fa23b4e78 100644 --- a/tests/integration/recordings/responses/61be36ad8ccd.json +++ b/tests/integration/recordings/responses/61be36ad8ccd.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 44225125, - "load_duration": 36913875, + "total_duration": 18669659, + "load_duration": 7831248, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.021802, - 0.088129535, - -0.10867403, - 0.0027561262, - 0.04917365, - -0.030165128, - -0.0155558735, - 0.027549915, - -0.025064131, - 0.016137881, - 0.124836035, - 0.0027821937, - -0.033310093, - -0.0071708336, - -0.07004796, - -0.027996853, - -0.09748515, - -0.091607764, - 0.013367206, - 0.08752305, - 0.013990884, - 0.03663788, - -0.036330026, - -0.019752761, - 0.04456914, - -0.009629443, - -0.01832647, - 0.048832405, - -0.015315298, - -0.07147843, - 0.04094573, - 0.082709365, - 0.063961774, - 0.01448001, - 0.13194442, - 0.0303949, - 0.101027474, - -0.030359762, - -0.047630757, - 0.044637363, - 0.027034018, - -0.029368822, - 0.038537122, - 0.0053882804, - 0.01478374, - 0.025617138, - 0.0041860593, - 0.0034900715, - 0.029765956, - -0.036669906, - -0.04589116, - 0.031120853, - -0.07786974, - -0.019517597, - 0.053876307, - -0.0152282175, - -0.0016955235, - 0.016938528, - 0.019939963, - 0.07106882, - 0.009938938, - 0.03114348, - -0.010335175, - 0.029952966, - 0.115054145, - 0.025746102, - -0.052842245, - -0.042447682, - 0.0053093657, - -0.09987591, - -0.12741813, - -0.012022532, - -0.013787561, - 0.05265948, - -0.01723935, - 0.009638554, - -0.0775266, - 0.0014047497, - 0.06974368, - -0.08465856, - -0.061480872, - -0.14244927, - 0.0096944375, - -0.008611519, - -0.0318523, - 0.12823504, - 0.053257603, - 0.021978743, - 0.0026468195, - 0.015444479, - -0.042528655, - 0.031551417, - -0.06209267, - 0.044017885, - -0.0060390937, - 0.06959196, - 0.0050514904, - 0.059341036, - 0.00658094, - 0.08397857, - -0.0067914296, - -0.041901726, - 0.027081704, - 0.106456675, - -0.039408114, - -0.053899165, - 0.09689717, - -0.0084604705, - 0.03398384, - -0.033843804, - 0.002225838, - -0.08180734, - -0.008216738, - -0.11271415, - 0.0058824755, - -0.095151186, - -0.07958445, - 0.052868627, - -0.08120183, - 0.034291897, - 0.07903789, - -0.02675632, - -0.04391073, - 0.0067707864, - -0.05438546, - -0.021719433, - 0.080597855, - -3.9388086e-33, - -0.0072714644, - -0.079664536, - 0.024838887, - 0.115598045, - 0.03591746, - -0.07254434, - 0.012642099, - 0.050809097, - -0.100082524, - 0.019521356, - 0.0035883472, - -0.07001022, - 0.007977421, - 0.029305879, - -0.017785804, - 0.02702277, - 0.016827941, - 0.035956737, - -0.0209356, - -0.032321777, - 0.056705642, - -0.009747762, - -0.059722506, - -0.053817417, - -0.055837773, - 0.06526892, - -0.024752634, - -0.07778206, - 0.038636208, - 0.008998632, - 0.009699391, - -0.02798574, - -0.024878206, - -0.0017547129, - 0.025541965, - 0.034623418, - -8.975541e-06, - 0.0034556785, - -0.04525613, - 0.03461154, - -0.025307115, - -0.02981576, - -0.019071916, - -0.023184983, - 0.049324982, - -0.061433185, - 0.00038017757, - 0.0028894164, - 0.027610173, - 0.0069347974, - -0.020659719, - 0.060771395, - 0.015200205, - 0.038918514, - -0.025353896, - -0.0017897633, - -0.019378036, - -0.0056970986, - -0.017806012, - 0.038060427, - 0.0320353, - 0.03998783, - -0.09612384, - 0.0006942505, - -0.018478483, - -0.06866618, - -0.0077035497, - -0.083554305, - 0.10223985, - 0.05141575, - -0.033018276, - -0.05033401, - 0.043923385, - 0.017748218, - -0.006601344, - -0.018691983, - 0.012763011, - 0.016694913, - -0.095070764, - -0.023533016, - 0.006879241, - -0.07225332, - -0.0029991802, - -0.06930797, - -0.027289826, - -0.0672911, - -0.006683099, - -0.06801406, - 0.04452207, - -0.09788058, - 0.050909285, - 0.010051549, - -0.04617998, - -0.067622505, - 0.04447288, - 2.5643933e-33, - 0.014783131, - 0.071710624, - -0.05237768, - 0.011041238, - -0.013921518, - 0.07072471, - 0.091977395, - -0.01916791, - -0.015780058, - 0.14812021, - 0.031904023, - 0.022344623, - 0.07071857, - -0.037060503, - 0.08806883, - -0.018145561, - -0.013254877, - -0.041782882, - -0.052317847, - -0.00279131, - -0.024807084, - 0.13974102, - 0.074973755, - 0.056424167, - -0.029412953, - 0.017093861, - 0.03373144, - 0.06874087, - 0.020454561, - -0.018965451, - 0.081238694, - 0.06527906, - -0.09342225, - 0.0037720343, - 0.06347132, - -0.08775714, - 0.09286548, - -0.024266576, - 0.029101077, - 0.0034162905, - 0.05528427, - 0.102037616, - -0.023588225, - 0.065829135, - 0.01520327, - 0.034344077, - 0.10559419, - 0.011605323, - 0.0409873, - -0.056635953, - 0.037730522, - -0.04976337, - 0.047961522, - 0.0042118295, - -0.014172872, - 0.07564937, - -0.009671058, - 0.05520304, - -0.031121492, - 0.019924358, - -0.024975697, - 0.031822197, - -0.019536836, - -0.009870229, - -0.020225972, - -0.03319855, - -0.026266782, - 0.038882248, - 0.012940086, - -0.041266225, - 0.012833021, - 0.028703777, - -0.054075323, - -0.07628176, - 0.021953572, - -0.023357453, - -0.026714878, - -0.029401133, - 0.005280363, - 0.012325193, - 0.05232579, - 0.0054451786, - -0.0063759633, - 0.04604998, - 0.042399842, - -0.018433316, - 0.01260558, - 0.09300185, - -0.005949781, - -0.015193224, - -0.011673769, - 0.048114438, - 0.02588804, - 0.050943956, - 0.005536351, - -1.5059804e-08, - -0.03100338, - -0.07003323, - -0.032613333, - -0.008732137, - -0.0045523546, - 0.0759239, - -0.032725554, - -0.08790561, - -0.032228027, - -0.02459868, - 0.051224917, - -0.034561895, - -0.08266327, - 0.013319846, - -0.020541467, - -0.056271035, - -0.009450659, - -0.015903467, - -0.036625408, - 0.010096497, - -0.03440534, - 0.0315293, - -0.00013937108, - 0.010463861, - 0.017065981, - 0.015492903, - 0.074808784, - 0.07079003, - -0.050000764, - -0.047017526, - 0.01375958, - 0.060757488, - -0.009361379, - -0.01570009, - -0.01836736, - 0.12301148, - 0.1185397, - 0.12366319, - 0.022782512, - -0.020027133, - -0.07401259, - -0.0047104736, - -0.024872223, - 0.006070436, - -0.06660639, - -0.08130306, - -0.0873992, - -0.0634906, - -0.039198957, - -0.11274462, - -0.030654918, - 0.026607778, - -0.063220546, - 0.042023618, - -0.039010853, - -0.009214424, - 0.005044682, - 0.0015641748, - -0.058640927, - 0.043107104, - 0.06682025, - 0.062172387, - 0.021147223, - -0.041068073 + -0.021797279, + 0.08814402, + -0.10868957, + 0.0027341088, + 0.049185295, + -0.030170735, + -0.015565467, + 0.027587239, + -0.025064457, + 0.016123094, + 0.12483694, + 0.002735925, + -0.033303194, + -0.0071613337, + -0.07005802, + -0.028024055, + -0.09749922, + -0.09159195, + 0.013367305, + 0.0874955, + 0.014002874, + 0.036639757, + -0.03636182, + -0.019740878, + 0.04459328, + -0.009643348, + -0.018319484, + 0.048830714, + -0.0152804, + -0.07148693, + 0.040963966, + 0.08269608, + 0.06397198, + 0.0145023735, + 0.13194914, + 0.030426234, + 0.10101107, + -0.030376758, + -0.047626566, + 0.04463136, + 0.027045978, + -0.029361075, + 0.038553316, + 0.005380632, + 0.014782317, + 0.025612796, + 0.0041573737, + 0.0035170745, + 0.029783405, + -0.03664018, + -0.0459057, + 0.031118676, + -0.077901915, + -0.01951666, + 0.05389714, + -0.015227032, + -0.0016507138, + 0.016938176, + 0.019922407, + 0.07105241, + 0.009955439, + 0.031143824, + -0.010342315, + 0.0299448, + 0.115018405, + 0.025722643, + -0.052856576, + -0.042419422, + 0.0053135715, + -0.099866174, + -0.12745431, + -0.012013655, + -0.013812364, + 0.052661266, + -0.017216302, + 0.009661314, + -0.07750365, + 0.001425789, + 0.06971633, + -0.08466273, + -0.061505307, + -0.1424137, + 0.009696796, + -0.008596895, + -0.031801328, + 0.12823558, + 0.053274382, + 0.02196283, + 0.0026299024, + 0.015462265, + -0.042509567, + 0.031536907, + -0.062131215, + 0.04401508, + -0.0060322434, + 0.06963364, + 0.005069902, + 0.059349127, + 0.0066066287, + 0.083945125, + -0.0067983367, + -0.04187391, + 0.027067436, + 0.10645863, + -0.039466046, + -0.053930666, + 0.09689939, + -0.008489689, + 0.033982914, + -0.033854645, + 0.0022207978, + -0.08181357, + -0.008203118, + -0.112689435, + 0.005881858, + -0.09516723, + -0.07958026, + 0.05286301, + -0.08119332, + 0.034290165, + 0.07901507, + -0.026746603, + -0.043884493, + 0.0067500784, + -0.054359503, + -0.021698626, + 0.08062436, + -3.9372978e-33, + -0.0072650607, + -0.07970752, + 0.024809107, + 0.1155797, + 0.035922393, + -0.072518565, + 0.012635176, + 0.050813816, + -0.10010529, + 0.019547075, + 0.0035949259, + -0.07004452, + 0.007995194, + 0.029300675, + -0.017782843, + 0.026989916, + 0.016807383, + 0.035927042, + -0.020967118, + -0.032325625, + 0.05671912, + -0.009719085, + -0.05972821, + -0.053807173, + -0.055842206, + 0.065258704, + -0.024726693, + -0.077762, + 0.03861746, + 0.008987917, + 0.009739114, + -0.028010633, + -0.02491916, + -0.0017105616, + 0.025539458, + 0.0346136, + 3.9485058e-05, + 0.0034435065, + -0.045235515, + 0.034653082, + -0.025328144, + -0.029821398, + -0.019025166, + -0.02314655, + 0.049356878, + -0.061453078, + 0.00034613282, + 0.0028801307, + 0.027612487, + 0.006939868, + -0.020667072, + 0.06074888, + 0.01522031, + 0.038911674, + -0.025372753, + -0.0018010045, + -0.019389275, + -0.0056944923, + -0.017822273, + 0.038047276, + 0.03205162, + 0.04001528, + -0.0961084, + 0.0007117376, + -0.018443316, + -0.06868148, + -0.0076998174, + -0.08358278, + 0.10225404, + 0.051446233, + -0.03301962, + -0.05037479, + 0.043945532, + 0.017751444, + -0.0066287024, + -0.01868368, + 0.012750775, + 0.016747138, + -0.09506785, + -0.023539655, + 0.0068607777, + -0.07226867, + -0.0030067777, + -0.069316946, + -0.027342388, + -0.067299545, + -0.0067162975, + -0.06797568, + 0.04455736, + -0.097934015, + 0.050929137, + 0.010035259, + -0.046227023, + -0.06760485, + 0.04445212, + 2.562595e-33, + 0.014783255, + 0.07173777, + -0.052347645, + 0.011015672, + -0.013930196, + 0.07069973, + 0.09197335, + -0.019221101, + -0.015802069, + 0.14809151, + 0.031869162, + 0.022357255, + 0.070741944, + -0.037042238, + 0.08803802, + -0.018144036, + -0.013264365, + -0.04176153, + -0.052341193, + -0.0027917302, + -0.024827031, + 0.13969763, + 0.07499699, + 0.056436434, + -0.029428342, + 0.017082963, + 0.033736177, + 0.06876884, + 0.020432826, + -0.018958652, + 0.08124247, + 0.06528793, + -0.0933768, + 0.0037903648, + 0.06345718, + -0.08775565, + 0.092871055, + -0.024276976, + 0.029103147, + 0.003399683, + 0.05533184, + 0.10196994, + -0.023569867, + 0.06581559, + 0.015236517, + 0.034391418, + 0.10560325, + 0.011587524, + 0.040974785, + -0.05662303, + 0.037732083, + -0.049770575, + 0.04793812, + 0.004231376, + -0.01415405, + 0.075640246, + -0.009698359, + 0.05522304, + -0.03112681, + 0.019937888, + -0.024967762, + 0.0318396, + -0.019503184, + -0.009845991, + -0.020246677, + -0.03324142, + -0.026290817, + 0.038862564, + 0.012934493, + -0.04129811, + 0.012831314, + 0.028768215, + -0.05400383, + -0.07626407, + 0.021966536, + -0.023368899, + -0.026754307, + -0.029407034, + 0.0053001987, + 0.012337391, + 0.05231288, + 0.005433406, + -0.0063848183, + 0.04605393, + 0.042325705, + -0.01845249, + 0.0126290405, + 0.093028955, + -0.0059780106, + -0.0152219515, + -0.011663129, + 0.048099615, + 0.025889266, + 0.05090448, + 0.005562377, + -1.5056981e-08, + -0.03096952, + -0.07003743, + -0.032617524, + -0.008757707, + -0.004564154, + 0.07594425, + -0.032733086, + -0.08789985, + -0.032205302, + -0.02457474, + 0.0512304, + -0.034549378, + -0.08262979, + 0.013313169, + -0.020548707, + -0.056250956, + -0.009471762, + -0.015904719, + -0.036591273, + 0.010126428, + -0.034383, + 0.031482615, + -0.0001312433, + 0.010469896, + 0.017070647, + 0.015479776, + 0.07480599, + 0.07080731, + -0.050010458, + -0.047061216, + 0.0137453, + 0.060734108, + -0.009365188, + -0.015720002, + -0.018347824, + 0.12303049, + 0.118518114, + 0.12366621, + 0.02281813, + -0.019984957, + -0.07401524, + -0.0047247335, + -0.024880406, + 0.006057382, + -0.066578485, + -0.08131662, + -0.087398425, + -0.06347802, + -0.039209016, + -0.1127259, + -0.030658804, + 0.026613072, + -0.06321768, + 0.042032808, + -0.03901875, + -0.009210964, + 0.00502309, + 0.0015242217, + -0.058664218, + 0.04312288, + 0.066781215, + 0.062229507, + 0.021180226, + -0.04108164 ] ] } diff --git a/tests/integration/recordings/responses/6906a6e71988.json b/tests/integration/recordings/responses/6906a6e71988.json index c488687de..44ae5c0eb 100644 --- a/tests/integration/recordings/responses/6906a6e71988.json +++ b/tests/integration/recordings/responses/6906a6e71988.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:23.01809Z", + "created_at": "2025-07-31T17:45:33.679033503Z", "done": true, "done_reason": "stop", - "total_duration": 448882000, - "load_duration": 83673125, + "total_duration": 4349645498, + "load_duration": 42994476, "prompt_eval_count": 317, - "prompt_eval_duration": 351902458, + "prompt_eval_duration": 4262865760, "eval_count": 2, - "eval_duration": 12722875, + "eval_duration": 43296572, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/6cc063bbd7d3.json b/tests/integration/recordings/responses/6cc063bbd7d3.json index 6eadff3aa..161ca1071 100644 --- a/tests/integration/recordings/responses/6cc063bbd7d3.json +++ b/tests/integration/recordings/responses/6cc063bbd7d3.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.509395Z", + "created_at": "2025-07-31T17:56:51.813281519Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.561227Z", + "created_at": "2025-07-31T17:56:51.989711694Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.604344Z", + "created_at": "2025-07-31T17:56:52.166114062Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.647038Z", + "created_at": "2025-07-31T17:56:52.343105056Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.688732Z", + "created_at": "2025-07-31T17:56:52.520952283Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.730495Z", + "created_at": "2025-07-31T17:56:52.698509394Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.772148Z", + "created_at": "2025-07-31T17:56:52.874946532Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.813191Z", + "created_at": "2025-07-31T17:56:53.056294341Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.85447Z", + "created_at": "2025-07-31T17:56:53.236117913Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.896136Z", + "created_at": "2025-07-31T17:56:53.414303003Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,7 +201,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.937588Z", + "created_at": "2025-07-31T17:56:53.59508044Z", "done": false, "done_reason": null, "total_duration": null, @@ -219,7 +219,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.978357Z", + "created_at": "2025-07-31T17:56:53.77302497Z", "done": false, "done_reason": null, "total_duration": null, @@ -237,7 +237,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.019403Z", + "created_at": "2025-07-31T17:56:53.952164995Z", "done": false, "done_reason": null, "total_duration": null, @@ -255,7 +255,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.06055Z", + "created_at": "2025-07-31T17:56:54.138962187Z", "done": false, "done_reason": null, "total_duration": null, @@ -273,7 +273,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.101456Z", + "created_at": "2025-07-31T17:56:54.316918504Z", "done": false, "done_reason": null, "total_duration": null, @@ -291,7 +291,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.142967Z", + "created_at": "2025-07-31T17:56:54.494254106Z", "done": false, "done_reason": null, "total_duration": null, @@ -309,7 +309,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.184487Z", + "created_at": "2025-07-31T17:56:54.675122971Z", "done": false, "done_reason": null, "total_duration": null, @@ -327,7 +327,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.226323Z", + "created_at": "2025-07-31T17:56:54.849660284Z", "done": false, "done_reason": null, "total_duration": null, @@ -345,7 +345,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.269043Z", + "created_at": "2025-07-31T17:56:55.027127762Z", "done": false, "done_reason": null, "total_duration": null, @@ -363,15 +363,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:59.311737Z", + "created_at": "2025-07-31T17:56:55.201988285Z", "done": true, "done_reason": "stop", - "total_duration": 1014917792, - "load_duration": 140789542, + "total_duration": 4724422329, + "load_duration": 43306862, "prompt_eval_count": 26, - "prompt_eval_duration": 70044833, + "prompt_eval_duration": 1291013267, "eval_count": 20, - "eval_duration": 803278042, + "eval_duration": 3389578393, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/6d35c91287e2.json b/tests/integration/recordings/responses/6d35c91287e2.json index 0cefca691..dbcc6c470 100644 --- a/tests/integration/recordings/responses/6d35c91287e2.json +++ b/tests/integration/recordings/responses/6d35c91287e2.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.498603Z", + "created_at": "2025-07-31T17:46:20.369406477Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.549619Z", + "created_at": "2025-07-31T17:46:20.5518691Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.598236Z", + "created_at": "2025-07-31T17:46:20.736317928Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.645776Z", + "created_at": "2025-07-31T17:46:20.930122964Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.695135Z", + "created_at": "2025-07-31T17:46:21.125702287Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.743802Z", + "created_at": "2025-07-31T17:46:21.317606091Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.793065Z", + "created_at": "2025-07-31T17:46:21.507713677Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.840802Z", + "created_at": "2025-07-31T17:46:21.695066353Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.887121Z", + "created_at": "2025-07-31T17:46:21.891869708Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.932548Z", + "created_at": "2025-07-31T17:46:22.085099601Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:26.977911Z", + "created_at": "2025-07-31T17:46:22.269187077Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:27.02202Z", + "created_at": "2025-07-31T17:46:22.453591089Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,15 +238,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:27.067458Z", + "created_at": "2025-07-31T17:46:22.638173365Z", "done": true, "done_reason": "stop", - "total_duration": 1040164917, - "load_duration": 169886125, + "total_duration": 4335218946, + "load_duration": 40997571, "prompt_eval_count": 417, - "prompt_eval_duration": 299516583, + "prompt_eval_duration": 2023436216, "eval_count": 13, - "eval_duration": 570085417, + "eval_duration": 2270231002, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/6fbea1abca7c.json b/tests/integration/recordings/responses/6fbea1abca7c.json index 801bed18f..62f182d3c 100644 --- a/tests/integration/recordings/responses/6fbea1abca7c.json +++ b/tests/integration/recordings/responses/6fbea1abca7c.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.219117Z", + "created_at": "2025-07-31T17:46:08.878347326Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.261232Z", + "created_at": "2025-07-31T17:46:09.062828335Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.30374Z", + "created_at": "2025-07-31T17:46:09.243638066Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.346151Z", + "created_at": "2025-07-31T17:46:09.423285311Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.388371Z", + "created_at": "2025-07-31T17:46:09.606591486Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.430754Z", + "created_at": "2025-07-31T17:46:09.78961117Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.474105Z", + "created_at": "2025-07-31T17:46:09.970261883Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.521006Z", + "created_at": "2025-07-31T17:46:10.154172091Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.565821Z", + "created_at": "2025-07-31T17:46:10.340795908Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.612477Z", + "created_at": "2025-07-31T17:46:10.526545522Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.660102Z", + "created_at": "2025-07-31T17:46:10.714952054Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.708431Z", + "created_at": "2025-07-31T17:46:10.910927483Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.757736Z", + "created_at": "2025-07-31T17:46:11.095386909Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.801918Z", + "created_at": "2025-07-31T17:46:11.281660364Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.845865Z", + "created_at": "2025-07-31T17:46:11.464474672Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.889535Z", + "created_at": "2025-07-31T17:46:11.650350705Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.938274Z", + "created_at": "2025-07-31T17:46:11.834464866Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:24.986637Z", + "created_at": "2025-07-31T17:46:12.019307704Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,15 +346,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:25.035053Z", + "created_at": "2025-07-31T17:46:12.209293507Z", "done": true, "done_reason": "stop", - "total_duration": 1672758667, - "load_duration": 82323250, + "total_duration": 35638042382, + "load_duration": 48273168, "prompt_eval_count": 386, - "prompt_eval_duration": 773194583, + "prompt_eval_duration": 32256935234, "eval_count": 19, - "eval_duration": 816461625, + "eval_duration": 3332216342, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/70adef2c30c4.json b/tests/integration/recordings/responses/70adef2c30c4.json index 2609d2c87..fe1697d6f 100644 --- a/tests/integration/recordings/responses/70adef2c30c4.json +++ b/tests/integration/recordings/responses/70adef2c30c4.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:58.183439Z", + "created_at": "2025-07-31T17:56:50.453323367Z", "done": true, "done_reason": "stop", - "total_duration": 3440514791, - "load_duration": 61560708, + "total_duration": 14856284585, + "load_duration": 44791770, "prompt_eval_count": 30, - "prompt_eval_duration": 92499375, + "prompt_eval_duration": 2452957312, "eval_count": 70, - "eval_duration": 3284810375, + "eval_duration": 12358070057, "response": "The answer is Saturn! Saturn's ring system is one of the most iconic and well-known in our solar system. The rings are made up of ice particles, rock debris, and dust that orbit around the planet due to its gravitational pull.\n\nWould you like to know more about Saturn's rings or is there something else I can help you with?", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/7354ec181984.json b/tests/integration/recordings/responses/7354ec181984.json index b76c85d93..0404c6a6a 100644 --- a/tests/integration/recordings/responses/7354ec181984.json +++ b/tests/integration/recordings/responses/7354ec181984.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:26.022261Z", + "created_at": "2025-07-31T17:51:16.201313167Z", "done": true, "done_reason": "stop", - "total_duration": 6323034167, - "load_duration": 107456542, + "total_duration": 27475921912, + "load_duration": 40564716, "prompt_eval_count": 25, - "prompt_eval_duration": 69246125, + "prompt_eval_duration": 964907432, "eval_count": 150, - "eval_duration": 6145775458, + "eval_duration": 26469935419, "response": "The smallest country in the world is the Vatican City, which has a total area of approximately 0.44 km\u00b2 (0.17 sq mi). It is an independent city-state located within Rome, Italy, and is home to the Pope and the central government of the Catholic Church.\n\nTo put that into perspective, the Vatican City is smaller than a golf course! Despite its tiny size, it has its own government, currency, postal system, and even its own police force. It's also home to numerous iconic landmarks like St. Peter's Basilica and the Sistine Chapel.\n\nInterestingly, the Vatican City is not only the smallest country in the world but also the most densely populated, with a population of just over 800 people!", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/75d0dd9d0fa3.json b/tests/integration/recordings/responses/75d0dd9d0fa3.json index 003218266..65225a11e 100644 --- a/tests/integration/recordings/responses/75d0dd9d0fa3.json +++ b/tests/integration/recordings/responses/75d0dd9d0fa3.json @@ -45,15 +45,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:28.736819Z", + "created_at": "2025-07-31T17:53:29.277838934Z", "done": true, "done_reason": "stop", - "total_duration": 1520367458, - "load_duration": 59997042, + "total_duration": 15295815581, + "load_duration": 42102876, "prompt_eval_count": 119, - "prompt_eval_duration": 198841625, + "prompt_eval_duration": 9859380831, "eval_count": 29, - "eval_duration": 1259800500, + "eval_duration": 5393566178, "response": "{ \"name\": \"Michael Jordan\", \"year_born\": \"1963\", \"year_retired\": \"2003\"}\n ", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/76b89a84cd6f.json b/tests/integration/recordings/responses/76b89a84cd6f.json index 3f5d84ec0..09cd3ae8d 100644 --- a/tests/integration/recordings/responses/76b89a84cd6f.json +++ b/tests/integration/recordings/responses/76b89a84cd6f.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 54951792, - "load_duration": 30643292, + "total_duration": 15551190, + "load_duration": 6064047, "prompt_eval_count": 5, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.04308226, - 0.008707138, - 0.06876158, - 0.018115537, - 0.04603657, - 0.0026118131, - -0.0032358477, - -0.041284926, - -0.09074888, - -0.033087812, - -0.026611822, - 0.0077352105, - 0.020191023, - -0.03254043, - -0.035847843, - 0.031108031, - -0.039247137, - -0.011286401, - -0.109710276, - -0.12942196, - 0.018077252, - 0.011446383, - -0.07231236, - -0.013655743, - 0.035438832, - 0.024783252, - 0.03387316, - 0.0726014, - -0.012643238, - -0.058606703, - 0.057943814, - -0.08163548, - 0.064962864, - 0.0013675748, - -0.06751009, - 0.03504323, - -0.044962864, - -0.004789603, - 0.039971247, - -0.010461211, - 0.019703588, - -0.09856083, - -0.01284534, - 0.018876119, - 0.09569305, - 0.11571406, - -0.040684983, - -0.026837468, - -0.046950106, - 0.022655226, - -0.0884734, - -0.023497678, - -0.022986038, - -0.031128721, - -0.052087843, - 0.04241795, - 0.011578454, - 0.06702011, - 0.027121129, - -0.0021518404, - 0.04675332, - -0.082024105, - -0.038331598, - 0.05215799, - 0.097757615, - -0.0006708623, - -0.051935766, - 0.09100271, - -0.016111707, - -0.06877312, - 0.00767068, - 0.076737314, - -0.0017499238, - 0.014369293, - 0.038031887, - -0.0044654603, - 0.011287075, - 0.0006178959, - 0.08834809, - -0.05933476, - -0.042706404, - -0.048178285, - -0.053068914, - 0.033110976, - 0.008051986, - -0.042581946, - -0.038104057, - -0.007202849, - 0.010891519, - -0.05466173, - 0.03903238, - -0.06774145, - -0.02356764, - -0.03883483, - 0.03464186, - 0.015297014, - 0.0073803077, - -0.12351391, - 0.036168184, - 0.13193323, - -0.06441449, - 0.033508655, - -0.01435515, - 0.0014314495, - 0.031048443, - -0.03981852, - 0.0236718, - -0.0028333638, - 0.096959464, - -0.13331193, - -0.054209094, - 0.019610135, - 0.06984815, - -0.05347757, - 0.0018131314, - 0.02127606, - 0.01981612, - 0.036502477, - 0.008825069, - 0.018954003, - -0.07161326, - -0.018733062, - 0.031044634, - 0.09102944, - 0.016508427, - -0.08625295, - -0.08300717, - -1.4044197e-34, - -0.072007515, - -0.045496386, - -0.027986562, - 0.05823018, - -0.010462877, - -0.06121516, - 0.026053715, - -0.06574638, - 0.029178392, - 0.012307141, - -0.06338016, - 0.040593755, - 0.03648161, - 0.01977942, - 0.08755496, - 0.028216325, - 0.044194777, - 0.076237544, - 0.02949726, - -0.0022650051, - 0.04304541, - 0.025918182, - 1.2261046e-05, - -0.038463842, - -0.0161955, - 0.03338553, - 0.02112944, - -0.023382189, - 0.009846733, - 0.033575017, - 0.030112585, - 0.060389582, - -0.06522927, - -0.016030189, - 0.019156763, - -0.002600835, - -0.04663393, - 0.02794595, - 0.021004112, - 0.0074595963, - -0.048745092, - -0.0070450655, - 0.019834043, - 0.016411202, - -0.06381404, - 0.031237993, - 0.091976196, - -0.0313931, - 0.022238847, - -0.015018542, - 0.0025784613, - -0.031382624, - -0.0152902305, - -0.025491757, - 0.08233924, - 0.14333151, - -0.0255008, - -0.005104579, - -0.02309693, - -0.03117742, - 0.06995927, - 0.030787794, - 0.04810884, - 0.037135385, - 0.0068392092, - 0.06759879, - 0.049763102, - 0.008472162, - 0.07170584, - 0.0076969583, - -0.005139827, - -0.0031728086, - 0.024646448, - -0.06879641, - 0.05249289, - -0.009404918, - 0.10184627, - -0.013639711, - -0.022681188, - 0.021382388, - -0.09593746, - 0.024071718, - -0.072101034, - -0.04462981, - 0.033456877, - -0.03942254, - 0.020099705, - -0.07495305, - -0.008311987, - 0.013811793, - -0.09847922, - 0.0336409, - 0.08235891, - -0.0034134828, - -0.05005179, - -2.0283256e-33, - -0.13664234, - 0.06463093, - 0.05221015, - 0.10102781, - 0.016344123, - -0.01269384, - -0.09024102, - -0.023596523, - 0.0057664234, - 0.10294541, - -0.025930807, - -0.040247634, - 0.034446176, - 0.019228913, - -0.056902077, - 0.019905953, - 0.018969242, - -0.039362065, - 0.011287794, - 0.056024995, - -0.016000811, - 0.058928564, - -0.038211577, - -0.030445429, - -0.02130076, - 0.031401403, - -0.021228284, - -0.01400283, - -0.051042903, - 0.048970606, - 0.018451849, - -0.015488385, - -0.05033241, - 0.053844187, - -0.050984643, - 0.016940817, - -0.032773405, - -0.02502497, - 0.000826887, - 0.10213942, - 0.04724571, - 0.010156266, - -0.11653258, - 0.012165439, - -0.029735534, - -0.09959623, - -0.052066926, - 0.06851813, - 0.054645896, - -0.066007115, - 0.025503889, - 0.013539478, - 0.008429433, - -0.10756056, - -0.08184448, - 0.07179834, - 0.007978949, - -0.013011469, - 0.020322459, - 0.07827889, - -0.07320297, - -0.1153648, - 0.04087073, - 0.04355079, - -0.0012279376, - 0.045840748, - -0.004366462, - 0.074786335, - -0.017625354, - -0.046014115, - 0.022716347, - 0.057738, - -0.015408269, - 0.007771719, - -0.04381374, - -0.05289107, - -0.08783473, - 0.016243288, - -0.018398289, - -0.05679973, - 0.036058675, - -0.040418148, - 0.039242174, - 0.083593465, - -0.019223504, - 0.05582025, - 0.04756948, - -0.07378718, - 0.03371102, - -0.08680738, - -0.010659349, - 0.0524085, - 0.009771544, - 0.023841262, - -0.086208895, - -1.7164519e-08, - 0.021028979, - -0.051292755, - 0.11877283, - -0.04687027, - 0.06566496, - 0.058750976, - -0.050496, - 0.055720143, - -0.040577173, - 0.055665523, - 0.025019526, - -0.001681203, - -0.031047702, - 0.022228474, - 0.028109053, - 0.03163934, - -0.025502652, - 0.020898303, - -0.023064507, - 0.013436037, - 0.07504084, - 0.022279648, - 0.028908938, - -0.014271217, - 0.025474275, - -0.051414162, - -0.014502164, - 0.014646399, - -0.028023712, - 0.08406334, - -0.07755092, - 0.038713943, - -0.0043370826, - 0.025676368, - 0.12571524, - 0.06996381, - 0.0059321956, - -0.10410214, - -0.041439336, - 0.016119901, - -0.040744506, - 0.017772397, - -0.09114363, - -0.026066387, - 0.055598073, - 0.016705057, - 0.016444646, - -0.11935461, - 0.02789905, - 0.0151745565, - 0.042357437, - 0.06817164, - 0.05782822, - 0.063278705, - 0.06748475, - 0.059781626, - 0.06468886, - -0.06749451, - -0.035589237, - 0.0640055, - 0.008595763, - 0.003157698, - 0.009343837, - -0.08392565 + -0.043116726, + 0.008762204, + 0.06876551, + 0.018154446, + 0.046014935, + 0.0026075789, + -0.0032023117, + -0.041317876, + -0.09075395, + -0.033095382, + -0.026616864, + 0.0077496697, + 0.020235015, + -0.03252394, + -0.035852764, + 0.031051183, + -0.039273262, + -0.011288686, + -0.1097229, + -0.1294173, + 0.018025845, + 0.011457748, + -0.072298184, + -0.013640621, + 0.035424188, + 0.024748072, + 0.03384521, + 0.07259934, + -0.012619609, + -0.058619946, + 0.05801557, + -0.08161403, + 0.064947575, + 0.001362615, + -0.06751795, + 0.035076257, + -0.044989314, + -0.0047651185, + 0.039968234, + -0.01040782, + 0.019694408, + -0.09855809, + -0.012823173, + 0.01890794, + 0.09571055, + 0.115746535, + -0.04069045, + -0.02680746, + -0.04694828, + 0.022680543, + -0.08849123, + -0.023503046, + -0.022990132, + -0.031088889, + -0.052039262, + 0.042437814, + 0.011638174, + 0.0670399, + 0.027137857, + -0.0021473, + 0.046701267, + -0.08202047, + -0.03838818, + 0.052178502, + 0.09778113, + -0.0006593349, + -0.051926874, + 0.09100724, + -0.016121143, + -0.06875496, + 0.007669393, + 0.0767284, + -0.0017276715, + 0.01433612, + 0.038044788, + -0.004464167, + 0.011304684, + 0.0006356988, + 0.08827611, + -0.0593169, + -0.04269011, + -0.04821714, + -0.053063165, + 0.033073585, + 0.008053761, + -0.04257827, + -0.03806659, + -0.007204833, + 0.010878339, + -0.054683488, + 0.03902928, + -0.06775304, + -0.023547798, + -0.038845137, + 0.034658056, + 0.015339489, + 0.0073782727, + -0.123479486, + 0.03620066, + 0.13195266, + -0.06441574, + 0.03353106, + -0.014377511, + 0.0014469373, + 0.031026753, + -0.039825514, + 0.023753827, + -0.0028518923, + 0.097023025, + -0.13331918, + -0.05414865, + 0.019609982, + 0.06981739, + -0.05350471, + 0.0017914361, + 0.02123565, + 0.019793358, + 0.03652012, + 0.008757212, + 0.018963538, + -0.07163105, + -0.018745081, + 0.03101379, + 0.09108634, + 0.016522765, + -0.08622687, + -0.083000556, + -1.4120944e-34, + -0.07204795, + -0.0454863, + -0.027961366, + 0.05822211, + -0.010502984, + -0.061186932, + 0.026028452, + -0.06577434, + 0.029192366, + 0.0122880135, + -0.063328385, + 0.040566593, + 0.036464784, + 0.019783257, + 0.087558, + 0.028167294, + 0.044170283, + 0.076228, + 0.029474925, + -0.0022399179, + 0.043097593, + 0.025920164, + 2.2395505e-05, + -0.03851217, + -0.016177999, + 0.03336546, + 0.021168817, + -0.0234282, + 0.009798394, + 0.03357645, + 0.03013725, + 0.060346767, + -0.065224335, + -0.015980415, + 0.019157061, + -0.0025903578, + -0.0466378, + 0.027946537, + 0.02100925, + 0.0074677323, + -0.048780043, + -0.007029379, + 0.019858357, + 0.016415412, + -0.0638162, + 0.031251393, + 0.09193477, + -0.031419653, + 0.022252394, + -0.01500479, + 0.0025760988, + -0.031387188, + -0.015299431, + -0.02546342, + 0.08234505, + 0.14331058, + -0.025547996, + -0.005159215, + -0.023134073, + -0.031177375, + 0.06997682, + 0.030782456, + 0.048128117, + 0.037137397, + 0.0068431245, + 0.06755107, + 0.049724884, + 0.008470394, + 0.071763836, + 0.0076690027, + -0.005134569, + -0.0031549458, + 0.024666298, + -0.068752035, + 0.05245442, + -0.00935608, + 0.10184588, + -0.013630788, + -0.022619508, + 0.021436714, + -0.095911086, + 0.024078555, + -0.07211819, + -0.04461444, + 0.03340193, + -0.039429355, + 0.020107377, + -0.074943066, + -0.008328609, + 0.013822816, + -0.098455004, + 0.03365863, + 0.0823121, + -0.003396206, + -0.0499897, + -2.0274605e-33, + -0.13668069, + 0.06462584, + 0.052196007, + 0.100969166, + 0.016337585, + -0.012695544, + -0.09028159, + -0.02356374, + 0.005722851, + 0.10296686, + -0.025897449, + -0.04024869, + 0.034410108, + 0.019260515, + -0.056976542, + 0.019898219, + 0.018957885, + -0.03935449, + 0.011276682, + 0.056057617, + -0.015980229, + 0.058943093, + -0.038266417, + -0.03044946, + -0.021351444, + 0.031391833, + -0.021235785, + -0.013972733, + -0.051025163, + 0.048980854, + 0.018483348, + -0.015526178, + -0.050335266, + 0.053864546, + -0.051029664, + 0.016963435, + -0.032785386, + -0.024993876, + 0.000785349, + 0.10212783, + 0.047282238, + 0.010184973, + -0.11654175, + 0.0121573685, + -0.029727131, + -0.09965936, + -0.052029923, + 0.06852823, + 0.054656398, + -0.065998994, + 0.025493179, + 0.013522059, + 0.008406762, + -0.10758715, + -0.08182221, + 0.07181613, + 0.008019134, + -0.013016491, + 0.020291433, + 0.07824832, + -0.07321792, + -0.11531359, + 0.040891647, + 0.04357381, + -0.0012217899, + 0.045807034, + -0.0044387314, + 0.07476453, + -0.017609056, + -0.045983914, + 0.022713736, + 0.05770199, + -0.01545519, + 0.007768293, + -0.043805454, + -0.052848887, + -0.08787122, + 0.016240425, + -0.018441278, + -0.05685473, + 0.036065023, + -0.04042333, + 0.039200664, + 0.08363829, + -0.019230386, + 0.055873565, + 0.047601603, + -0.073787, + 0.03373948, + -0.08676912, + -0.010690602, + 0.052389733, + 0.009766554, + 0.023823792, + -0.08622312, + -1.7164984e-08, + 0.02098433, + -0.051278073, + 0.118761696, + -0.046744175, + 0.06562692, + 0.05873887, + -0.05050669, + 0.05573273, + -0.040608365, + 0.05565212, + 0.025007756, + -0.0016481675, + -0.030990602, + 0.022212071, + 0.028143445, + 0.031635366, + -0.025462067, + 0.020907363, + -0.0230822, + 0.013461223, + 0.075033404, + 0.02227541, + 0.028905692, + -0.014230868, + 0.025434028, + -0.051393002, + -0.014499751, + 0.0146705145, + -0.027979253, + 0.084023595, + -0.07754991, + 0.038684603, + -0.004379834, + 0.025665374, + 0.12571882, + 0.06997437, + 0.0059620147, + -0.10417642, + -0.041501433, + 0.016078206, + -0.040800873, + 0.017794596, + -0.091134265, + -0.026061513, + 0.055591512, + 0.016653784, + 0.016463995, + -0.119408, + 0.027920188, + 0.015162422, + 0.04233569, + 0.06814759, + 0.057812255, + 0.06331279, + 0.06752363, + 0.05975722, + 0.06466393, + -0.06744258, + -0.035575487, + 0.0640129, + 0.00857161, + 0.0031763979, + 0.009383616, + -0.08389455 ] ] } diff --git a/tests/integration/recordings/responses/7b4815aba6c5.json b/tests/integration/recordings/responses/7b4815aba6c5.json index 1d3b3a05a..f8f55d673 100644 --- a/tests/integration/recordings/responses/7b4815aba6c5.json +++ b/tests/integration/recordings/responses/7b4815aba6c5.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.15813Z", + "created_at": "2025-07-31T17:44:12.691972914Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.200509Z", + "created_at": "2025-07-31T17:44:12.873445514Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.242546Z", + "created_at": "2025-07-31T17:44:13.058391013Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.28427Z", + "created_at": "2025-07-31T17:44:13.241305582Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.326031Z", + "created_at": "2025-07-31T17:44:13.426830341Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.367844Z", + "created_at": "2025-07-31T17:44:13.613434143Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.409826Z", + "created_at": "2025-07-31T17:44:13.798507122Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.45127Z", + "created_at": "2025-07-31T17:44:13.985391256Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.493038Z", + "created_at": "2025-07-31T17:44:14.168605602Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.534837Z", + "created_at": "2025-07-31T17:44:14.348657505Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.576616Z", + "created_at": "2025-07-31T17:44:14.536885315Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.618384Z", + "created_at": "2025-07-31T17:44:14.721523959Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.660068Z", + "created_at": "2025-07-31T17:44:14.905534546Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.702659Z", + "created_at": "2025-07-31T17:44:15.086739434Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.745358Z", + "created_at": "2025-07-31T17:44:15.267128106Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.787884Z", + "created_at": "2025-07-31T17:44:15.453189872Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.830511Z", + "created_at": "2025-07-31T17:44:15.637366291Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.872944Z", + "created_at": "2025-07-31T17:44:15.824993856Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,15 +346,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:08.915128Z", + "created_at": "2025-07-31T17:44:16.009001804Z", "done": true, "done_reason": "stop", - "total_duration": 1584166500, - "load_duration": 126445334, + "total_duration": 33974512186, + "load_duration": 44406815, "prompt_eval_count": 368, - "prompt_eval_duration": 699644917, + "prompt_eval_duration": 30611002922, "eval_count": 19, - "eval_duration": 757663250, + "eval_duration": 3318539690, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/7e6806cba34a.json b/tests/integration/recordings/responses/7e6806cba34a.json index 23eae3ac9..7b1d5261e 100644 --- a/tests/integration/recordings/responses/7e6806cba34a.json +++ b/tests/integration/recordings/responses/7e6806cba34a.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.546801Z", + "created_at": "2025-07-31T17:59:14.382398152Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.588481Z", + "created_at": "2025-07-31T17:59:14.561084788Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.630066Z", + "created_at": "2025-07-31T17:59:14.743154167Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.671027Z", + "created_at": "2025-07-31T17:59:14.920818124Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.712294Z", + "created_at": "2025-07-31T17:59:15.099067906Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.753866Z", + "created_at": "2025-07-31T17:59:15.274401879Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.795863Z", + "created_at": "2025-07-31T17:59:15.449669669Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.83722Z", + "created_at": "2025-07-31T17:59:15.626501213Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.878343Z", + "created_at": "2025-07-31T17:59:15.802614623Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.919504Z", + "created_at": "2025-07-31T17:59:15.978698104Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,7 +201,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:42.960515Z", + "created_at": "2025-07-31T17:59:16.160654179Z", "done": false, "done_reason": null, "total_duration": null, @@ -219,7 +219,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.001631Z", + "created_at": "2025-07-31T17:59:16.338412914Z", "done": false, "done_reason": null, "total_duration": null, @@ -237,15 +237,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:43.043353Z", + "created_at": "2025-07-31T17:59:16.521646436Z", "done": true, "done_reason": "stop", - "total_duration": 834188250, - "load_duration": 75182417, + "total_duration": 4555044563, + "load_duration": 43101307, "prompt_eval_count": 29, - "prompt_eval_duration": 261107458, + "prompt_eval_duration": 2371036213, "eval_count": 13, - "eval_duration": 497358667, + "eval_duration": 2140342701, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/80e4404d8987.json b/tests/integration/recordings/responses/80e4404d8987.json index d28c14b31..5f2d0fc5f 100644 --- a/tests/integration/recordings/responses/80e4404d8987.json +++ b/tests/integration/recordings/responses/80e4404d8987.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.720409Z", + "created_at": "2025-07-31T17:43:33.008989003Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.76184Z", + "created_at": "2025-07-31T17:43:33.18634026Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.803087Z", + "created_at": "2025-07-31T17:43:33.363360212Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.844065Z", + "created_at": "2025-07-31T17:43:33.540067474Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.885088Z", + "created_at": "2025-07-31T17:43:33.717434082Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.926149Z", + "created_at": "2025-07-31T17:43:33.901524865Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:05.967374Z", + "created_at": "2025-07-31T17:43:34.086510094Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:06.008267Z", + "created_at": "2025-07-31T17:43:34.268793455Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:06.049424Z", + "created_at": "2025-07-31T17:43:34.447206727Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,15 +184,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:06.090397Z", + "created_at": "2025-07-31T17:43:34.629253797Z", "done": true, "done_reason": "stop", - "total_duration": 1022446958, - "load_duration": 63964125, + "total_duration": 14403810481, + "load_duration": 10124930315, "prompt_eval_count": 31, - "prompt_eval_duration": 586685542, + "prompt_eval_duration": 2656282776, "eval_count": 10, - "eval_duration": 371051666, + "eval_duration": 1621858907, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/80f09f27dd61.json b/tests/integration/recordings/responses/80f09f27dd61.json index 1294ab795..349b3671a 100644 --- a/tests/integration/recordings/responses/80f09f27dd61.json +++ b/tests/integration/recordings/responses/80f09f27dd61.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -20,14 +20,14 @@ "body": { "__type__": "openai.types.chat.chat_completion.ChatCompletion", "__data__": { - "id": "chatcmpl-33", + "id": "chatcmpl-544", "choices": [ { "finish_reason": "stop", "index": 0, "logprobs": null, "message": { - "content": "Hello! Welcome. How can I assist you today?", + "content": "Hello! It's nice to meet you. Is there something I can help you with, or would you like to chat?", "refusal": null, "role": "assistant", "annotations": null, @@ -37,15 +37,15 @@ } } ], - "created": 1753814886, + "created": 1753984701, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion", "service_tier": null, "system_fingerprint": "fp_ollama", "usage": { - "completion_tokens": 12, + "completion_tokens": 26, "prompt_tokens": 29, - "total_tokens": 41, + "total_tokens": 55, "completion_tokens_details": null, "prompt_tokens_details": null } diff --git a/tests/integration/recordings/responses/836f51dfb3c5.json b/tests/integration/recordings/responses/836f51dfb3c5.json index 727a8b873..cd75fa2f5 100644 --- a/tests/integration/recordings/responses/836f51dfb3c5.json +++ b/tests/integration/recordings/responses/836f51dfb3c5.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:11.406221Z", + "created_at": "2025-07-31T17:44:32.660113677Z", "done": true, "done_reason": "stop", - "total_duration": 235117291, - "load_duration": 88000791, + "total_duration": 2803904109, + "load_duration": 43469066, "prompt_eval_count": 214, - "prompt_eval_duration": 132949959, + "prompt_eval_duration": 2716196330, "eval_count": 2, - "eval_duration": 13414416, + "eval_duration": 43729281, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/83c2ffb72daa.json b/tests/integration/recordings/responses/83c2ffb72daa.json index 0d5b5698d..86e1d2762 100644 --- a/tests/integration/recordings/responses/83c2ffb72daa.json +++ b/tests/integration/recordings/responses/83c2ffb72daa.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 52467750, - "load_duration": 31533500, + "total_duration": 18512787, + "load_duration": 11031698, "prompt_eval_count": 2, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - 0.06829306, - 0.06173801, - -0.0064223157, - 0.08267553, - -0.078277536, - 0.026545998, - 0.13129343, - 0.041391026, - -0.019504873, - -0.02713137, - 0.08875854, - -0.10276947, - 0.05070561, - -0.071384996, - -0.00928895, - -0.039247785, - 0.028884366, - -0.010484679, - -0.024695162, - -0.035464898, - -0.040930223, - -0.009903115, - -0.026185343, - 0.057967443, - -0.0006098045, - 0.0076593193, - 0.013928812, - -0.0016587632, - 0.044655178, - -0.05899092, - -0.03795896, - 0.037799176, - -0.0332701, - 0.071682036, - 0.097220846, - -0.08261943, - 0.02762241, - -0.014190529, - 0.018169386, - -0.0027171622, - -0.024265053, - -0.11493207, - 0.08515992, - -0.01675261, - -0.0063101193, - 0.06525532, - -0.05800194, - 0.09667521, - -0.014198328, - -0.0068260604, - -0.09889978, - -0.01510962, - -0.07833434, - -0.03558934, - -0.008278174, - -0.013655411, - -0.07625151, - -0.030405695, - -0.013589355, - 0.05011788, - -0.010591766, - -0.038398705, - 0.067407176, - 0.035656955, - 0.010748781, - -0.0782303, - -0.0068980707, - -0.03009224, - 0.055957098, - -0.07684975, - -0.009063114, - -0.0028242331, - -0.02941445, - 0.06881706, - 0.013745152, - 0.030784354, - -0.036471423, - -0.071473554, - 0.054742932, - -0.028959777, - -0.0646612, - -0.059742935, - -0.067661926, - 0.02277713, - 0.07953034, - 0.05176706, - 0.14789894, - -0.0024908802, - -0.055424616, - -0.027760211, - 0.019384153, - 0.06692775, - -0.07952429, - 0.019047037, - -0.0009761573, - 0.013479472, - 0.03820792, - -0.040212464, - 0.06499357, - 0.13929029, - 0.05928682, - 0.018087227, - -0.049103815, - -0.05746931, - -0.17034934, - 0.009854012, - 0.04478709, - -0.08707101, - 0.046889856, - -0.020303955, - -0.062274978, - 0.03028755, - 0.049917854, - -0.030625027, - -0.0071967863, - -0.060630836, - -0.0057445974, - 0.02869731, - -0.055902474, - -0.006085085, - 0.075516894, - 0.07304867, - -0.03200334, - -0.02799431, - -0.0013179934, - 0.023734178, - 0.08233767, - -2.0787383e-33, - 0.014712576, - -0.08495617, - 0.059368838, - -0.0078545045, - -0.015981605, - 0.025985476, - 0.03761475, - 0.12561654, - -0.040023252, - 0.024720326, - 0.014450719, - -0.06304022, - 0.034111224, - -0.0076677934, - 0.008186544, - 0.104618765, - 0.01885282, - -0.021535598, - -0.043817643, - 0.056795686, - 0.0162111, - -0.073493764, - 0.02015092, - 0.05246774, - 0.015011722, - -0.065883316, - -0.032571133, - 0.025002327, - 0.018430093, - -0.00030110884, - -0.06266603, - -0.0061966996, - -0.16044672, - 0.028114, - 0.032982383, - 0.03726186, - 0.05405662, - -0.007922701, - -0.008597104, - 0.054075304, - -0.046998195, - -0.03870265, - 0.08493373, - -0.005938321, - 0.021924786, - -0.052063633, - -0.0474363, - -0.054906394, - 0.03400279, - -0.028335832, - -0.03204598, - -0.0013805361, - -0.04042138, - -0.017744347, - 0.05225112, - 0.0038320313, - 0.008692027, - 0.032701842, - 0.010805374, - 0.111949906, - -0.019722536, - -0.04577441, - -0.0020288338, - 0.020897591, - -0.0061685205, - -0.0017238781, - -0.0068083988, - -0.08133369, - 0.091827765, - 0.048646387, - 0.07771223, - -0.05870432, - 0.0063732844, - 0.003602972, - -0.071249805, - 0.022061156, - 0.019477166, - 0.101326875, - 0.006618201, - -0.044631816, - 0.061397545, - -0.091977604, - -0.013284187, - 0.014608401, - -0.017614143, - 0.0073858355, - 0.0062043285, - -0.04802106, - 0.013127447, - -0.07759211, - 0.01413356, - 0.035386372, - -0.026163345, - 0.002707529, - 0.086350374, - 9.1322365e-34, - -0.022040654, - 0.050855946, - -0.027267559, - 0.028623927, - 0.013727834, - -0.07108624, - 0.090404175, - -0.090647236, - -0.06563531, - 0.066881575, - 0.067018434, - -0.050155967, - 0.01906401, - -0.041479547, - 0.012601864, - 0.06909683, - 0.028203063, - -0.07096439, - -0.061153483, - 0.031663455, - -0.09626923, - 0.13134155, - -0.003593555, - -0.027185703, - -0.062974066, - -0.0009243527, - -0.0086801, - -0.03132579, - -0.01858645, - 0.011512133, - 0.07186438, - -0.071975954, - -0.0058840294, - 0.0935521, - 0.046686247, - -0.0319705, - 0.06956754, - -0.04588064, - 0.010095534, - 0.06409261, - 0.072478145, - 0.047231663, - 0.048781574, - 0.06763336, - 0.00544567, - 0.035764705, - 0.018254025, - -0.038195167, - 0.05008257, - 0.041405946, - -0.025459182, - 0.021584406, - 0.014274052, - -0.0071268557, - -0.014267975, - -0.010105019, - -0.09164536, - 0.009354, - 0.0043337494, - -0.009582353, - -0.029860858, - 0.1747107, - -0.004588478, - 0.05782761, - -0.044819914, - -0.05143084, - -0.045887187, - 0.0074449596, - 0.0054387185, - 0.03959965, - -0.056232695, - -0.0022210428, - 0.047835756, - -0.039582185, - 0.027316226, - 0.03971807, - -0.079697974, - 0.035112984, - 0.029242193, - 0.010144024, - -0.039045013, - -0.027879896, - -0.04085825, - 0.04611513, - -0.06931006, - 0.06197763, - 0.03922113, - 0.025860274, - 0.0064425287, - 0.053613506, - 0.069628745, - -0.007990118, - -0.038263954, - -0.10954398, - 0.018542193, - -1.3334614e-08, - -0.025668537, - -0.074732535, - -0.019855397, - 0.03849193, - 0.027314082, - -0.0108753685, - -0.03520762, - 0.03607515, - -0.063237555, - 0.011492363, - 0.03342595, - -0.012063489, - 0.003983985, - 0.016522186, - -0.008002231, - -0.041689247, - -0.07092196, - 0.008746665, - 0.0044521443, - -0.03877822, - -0.05125361, - 0.017749831, - -0.018253427, - 0.04394152, - -0.042883433, - 0.08245374, - 0.015452847, - 0.022076963, - 0.04442366, - 0.022832835, - 0.0829697, - -0.012612379, - 0.013092737, - -0.06689179, - 0.047846217, - -0.0450767, - 0.006519167, - 0.0055980817, - -0.019575235, - -0.017305179, - -0.038374994, - -0.0004379076, - -0.008650625, - -0.026787048, - -0.065987535, - -0.14336497, - 0.0415435, - -0.04859031, - 0.012749022, - -0.08499327, - -0.01095022, - -0.038154606, - 0.03009022, - -0.038868725, - -0.036706466, - 0.046492293, - 0.036234695, - 0.052362718, - -0.09623827, - -0.041491255, - 0.050219566, - -2.0856969e-05, - 0.0019338039, - 0.019553944 + 0.06828853, + 0.06176443, + -0.0064164964, + 0.08271423, + -0.07821361, + 0.026539708, + 0.13127756, + 0.041380048, + -0.019513987, + -0.027114647, + 0.088755146, + -0.10272858, + 0.05070812, + -0.07138208, + -0.009237211, + -0.03925189, + 0.028860727, + -0.010505408, + -0.024712363, + -0.035447232, + -0.040921025, + -0.009929597, + -0.026157938, + 0.057920404, + -0.000610026, + 0.0076649976, + 0.013957126, + -0.0016385933, + 0.044628702, + -0.058993056, + -0.037964217, + 0.037799507, + -0.03327965, + 0.07171921, + 0.09722134, + -0.08262229, + 0.02760268, + -0.014165805, + 0.01818236, + -0.0026569532, + -0.024224523, + -0.11491416, + 0.0851865, + -0.016761899, + -0.006304915, + 0.065241225, + -0.058053154, + 0.0966831, + -0.014210973, + -0.006840772, + -0.09888544, + -0.0151145635, + -0.07833236, + -0.03558869, + -0.008314475, + -0.013676866, + -0.0761953, + -0.030428959, + -0.013569219, + 0.05013825, + -0.0106057385, + -0.038397208, + 0.06740368, + 0.035669673, + 0.0107887, + -0.0782119, + -0.0068797, + -0.03011727, + 0.05595264, + -0.0768514, + -0.009096549, + -0.0027871567, + -0.029414741, + 0.06881623, + 0.013754744, + 0.030826487, + -0.036464185, + -0.07146736, + 0.05470558, + -0.02897486, + -0.06466289, + -0.05974617, + -0.06765741, + 0.02273816, + 0.07949564, + 0.05178505, + 0.14790222, + -0.0025215826, + -0.05544018, + -0.027736945, + 0.019415101, + 0.06687858, + -0.07953003, + 0.019072771, + -0.00095781777, + 0.01352604, + 0.038187943, + -0.040192492, + 0.064972185, + 0.13924964, + 0.059315108, + 0.018057114, + -0.049068965, + -0.05745058, + -0.17035483, + 0.009827581, + 0.044772815, + -0.087117024, + 0.046921514, + -0.020351866, + -0.06224432, + 0.030254036, + 0.04995525, + -0.030603461, + -0.007250532, + -0.060660776, + -0.0057841516, + 0.028707923, + -0.05595239, + -0.0061211013, + 0.0755027, + 0.073071584, + -0.031982183, + -0.028019294, + -0.0012888111, + 0.023758534, + 0.08233939, + -2.0778123e-33, + 0.014715223, + -0.08498444, + 0.05937457, + -0.007845196, + -0.01597936, + 0.025986308, + 0.037590146, + 0.12566452, + -0.039989363, + 0.024738252, + 0.014438816, + -0.06302637, + 0.0340797, + -0.00766197, + 0.008190336, + 0.10460811, + 0.018833369, + -0.021560388, + -0.04381885, + 0.056816146, + 0.016241364, + -0.07346648, + 0.020132408, + 0.052436877, + 0.014993327, + -0.06591125, + -0.032540314, + 0.024993971, + 0.018408103, + -0.00031402768, + -0.06270013, + -0.0062176185, + -0.1604857, + 0.028161483, + 0.032995526, + 0.037255995, + 0.054070372, + -0.007917325, + -0.008593869, + 0.054075062, + -0.047028862, + -0.038699273, + 0.084944956, + -0.005901343, + 0.021917563, + -0.052044198, + -0.047473878, + -0.05488473, + 0.0340269, + -0.028332518, + -0.03206542, + -0.0013664847, + -0.040407866, + -0.017748341, + 0.052257605, + 0.003831341, + 0.008690302, + 0.0326634, + 0.010794847, + 0.11198066, + -0.0197124, + -0.045761418, + -0.0020336432, + 0.020911695, + -0.0061559565, + -0.0017037267, + -0.006830028, + -0.08132594, + 0.09181586, + 0.048670176, + 0.07772897, + -0.058737844, + 0.006335169, + 0.0036056917, + -0.07125556, + 0.022052463, + 0.019465465, + 0.10136954, + 0.0066308854, + -0.04465323, + 0.061420094, + -0.09194896, + -0.01327561, + 0.014581149, + -0.017611397, + 0.007349155, + 0.00623538, + -0.048055336, + 0.013159856, + -0.077642046, + 0.014176555, + 0.035402596, + -0.026225176, + 0.0027164302, + 0.08640326, + 9.131446e-34, + -0.022036942, + 0.050841562, + -0.027241774, + 0.028640023, + 0.0137315225, + -0.07108254, + 0.090386234, + -0.09063743, + -0.06564483, + 0.06686015, + 0.06697802, + -0.050123725, + 0.019029083, + -0.04146069, + 0.012556864, + 0.06908357, + 0.028190888, + -0.07095583, + -0.061176844, + 0.031665504, + -0.09627965, + 0.13132389, + -0.0035763069, + -0.027172986, + -0.0630171, + -0.00091286737, + -0.008730202, + -0.031350046, + -0.018555095, + 0.011510101, + 0.071851164, + -0.07196774, + -0.0059023383, + 0.0935336, + 0.046666197, + -0.031923246, + 0.06959583, + -0.045899943, + 0.010128291, + 0.06411297, + 0.07247715, + 0.04719049, + 0.048774146, + 0.06759043, + 0.005473298, + 0.03578301, + 0.018263457, + -0.038228214, + 0.050064407, + 0.04143578, + -0.025405152, + 0.021583116, + 0.014285433, + -0.0071130656, + -0.014282598, + -0.010110906, + -0.09167918, + 0.009367985, + 0.0043357764, + -0.009596076, + -0.029858816, + 0.17470205, + -0.0045499653, + 0.057866864, + -0.0448407, + -0.05143361, + -0.04589322, + 0.00746687, + 0.0054419166, + 0.039566983, + -0.05621581, + -0.0022154362, + 0.047840577, + -0.039587613, + 0.027299056, + 0.039714534, + -0.079680495, + 0.03507584, + 0.029217845, + 0.010124337, + -0.03903408, + -0.027895331, + -0.040882032, + 0.046143655, + -0.06931917, + 0.061984897, + 0.039232828, + 0.025852159, + 0.0064182505, + 0.053619858, + 0.06963339, + -0.007952825, + -0.038268514, + -0.109513305, + 0.01855092, + -1.3333404e-08, + -0.025683647, + -0.0746983, + -0.019864988, + 0.038491003, + 0.027304374, + -0.010895902, + -0.035199717, + 0.036022183, + -0.0632235, + 0.01150509, + 0.03342266, + -0.012038254, + 0.0039724396, + 0.016541637, + -0.007991224, + -0.041665304, + -0.07092065, + 0.008739235, + 0.0044709085, + -0.0388358, + -0.051271453, + 0.017737802, + -0.018253798, + 0.04398574, + -0.042873986, + 0.08246222, + 0.0154647175, + 0.022104649, + 0.044403654, + 0.022859568, + 0.08297405, + -0.012621729, + 0.013065749, + -0.066898905, + 0.04787577, + -0.045054216, + 0.006517585, + 0.005591374, + -0.019547338, + -0.017292783, + -0.03841781, + -0.00047104285, + -0.008667157, + -0.026772907, + -0.06594723, + -0.14334334, + 0.041484844, + -0.048574504, + 0.012758157, + -0.08496632, + -0.0109471325, + -0.038168024, + 0.030072162, + -0.038850017, + -0.036700998, + 0.04647737, + 0.036238283, + 0.052374676, + -0.096244395, + -0.041510135, + 0.050267737, + -1.2173862e-06, + 0.0019191783, + 0.019568525 ] ] } diff --git a/tests/integration/recordings/responses/840fbb380b73.json b/tests/integration/recordings/responses/840fbb380b73.json index 003de6b81..5b8c5ef9d 100644 --- a/tests/integration/recordings/responses/840fbb380b73.json +++ b/tests/integration/recordings/responses/840fbb380b73.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:06.949885Z", + "created_at": "2025-07-31T17:43:40.884537604Z", "done": true, "done_reason": "stop", - "total_duration": 268717875, - "load_duration": 91961875, + "total_duration": 3179669897, + "load_duration": 48206132, "prompt_eval_count": 233, - "prompt_eval_duration": 130104500, + "prompt_eval_duration": 2966799546, "eval_count": 5, - "eval_duration": 45715417, + "eval_duration": 164144409, "response": "unsafe\nS1", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/84cab42e1f5c.json b/tests/integration/recordings/responses/84cab42e1f5c.json index 04c3f9663..69555f93e 100644 --- a/tests/integration/recordings/responses/84cab42e1f5c.json +++ b/tests/integration/recordings/responses/84cab42e1f5c.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -17,7 +17,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -26,7 +26,7 @@ "text": "Blue" } ], - "created": 1753814831, + "created": 1753984323, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -36,7 +36,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -45,7 +45,7 @@ "text": ".\n\n" } ], - "created": 1753814831, + "created": 1753984323, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -55,16 +55,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": "My" + "text": "This" } ], - "created": 1753814831, + "created": 1753984323, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -74,7 +74,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -83,7 +83,7 @@ "text": " response" } ], - "created": 1753814831, + "created": 1753984323, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -93,16 +93,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " is" + "text": " completes" } ], - "created": 1753814831, + "created": 1753984323, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -112,16 +112,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " based" + "text": " the" } ], - "created": 1753814831, + "created": 1753984324, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -131,45 +131,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " on" - } - ], - "created": 1753814831, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " a" - } - ], - "created": 1753814831, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -178,7 +140,7 @@ "text": " traditional" } ], - "created": 1753814831, + "created": 1753984324, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -188,16 +150,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " poem" + "text": " English" } ], - "created": 1753814831, + "created": 1753984324, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -207,7 +169,83 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " rhyme" + } + ], + "created": 1753984324, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " that" + } + ], + "created": 1753984324, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " often" + } + ], + "created": 1753984324, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " begins" + } + ], + "created": 1753984325, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -216,7 +254,7 @@ "text": " with" } ], - "created": 1753814831, + "created": 1753984325, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -226,83 +264,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " the" - } - ], - "created": 1753814831, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " first" - } - ], - "created": 1753814831, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " line" - } - ], - "created": 1753814831, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " being" - } - ], - "created": 1753814831, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -311,7 +273,7 @@ "text": " \"" } ], - "created": 1753814831, + "created": 1753984325, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -321,7 +283,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -330,7 +292,7 @@ "text": "R" } ], - "created": 1753814831, + "created": 1753984325, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -340,7 +302,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -349,7 +311,7 @@ "text": "oses" } ], - "created": 1753814831, + "created": 1753984325, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -359,7 +321,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -368,7 +330,7 @@ "text": " are" } ], - "created": 1753814832, + "created": 1753984326, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -378,7 +340,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -387,7 +349,7 @@ "text": " red" } ], - "created": 1753814832, + "created": 1753984326, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -397,83 +359,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": "\"," - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " but" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " in" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " reality" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -482,7 +368,7 @@ "text": "," } ], - "created": 1753814832, + "created": 1753984326, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -492,16 +378,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " roses" + "text": " v" } ], - "created": 1753814832, + "created": 1753984326, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -511,349 +397,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " come" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " in" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " various" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " colors" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " such" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " as" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " red" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": "," - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " pink" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": "," - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " yellow" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": "," - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " white" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": "," - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " and" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " purple" - } - ], - "created": 1753814832, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": "." - } - ], - "created": 1753814833, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", - "choices": [ - { - "finish_reason": null, - "index": 0, - "logprobs": null, - "text": " V" - } - ], - "created": 1753814833, - "model": "llama3.2:3b-instruct-fp16", - "object": "text_completion", - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.completion.Completion", - "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -862,7 +406,7 @@ "text": "io" } ], - "created": 1753814833, + "created": 1753984326, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -872,7 +416,7 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, @@ -881,7 +425,7 @@ "text": "lets" } ], - "created": 1753814833, + "created": 1753984326, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -891,16 +435,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": "," + "text": " are" } ], - "created": 1753814833, + "created": 1753984327, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -910,16 +454,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " on" + "text": "...\"" } ], - "created": 1753814833, + "created": 1753984327, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -929,16 +473,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " the" + "text": " The" } ], - "created": 1753814833, + "created": 1753984327, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -948,16 +492,16 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", "choices": [ { "finish_reason": null, "index": 0, "logprobs": null, - "text": " other" + "text": " full" } ], - "created": 1753814833, + "created": 1753984327, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", @@ -967,7 +511,463 @@ { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-313", + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " rhyme" + } + ], + "created": 1753984327, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " is" + } + ], + "created": 1753984328, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " typically" + } + ], + "created": 1753984328, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": ":\n\n" + } + ], + "created": 1753984328, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "R" + } + ], + "created": 1753984328, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "oses" + } + ], + "created": 1753984328, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " are" + } + ], + "created": 1753984328, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " red" + } + ], + "created": 1753984329, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": ",\n" + } + ], + "created": 1753984329, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "V" + } + ], + "created": 1753984329, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "io" + } + ], + "created": 1753984329, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "lets" + } + ], + "created": 1753984329, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " are" + } + ], + "created": 1753984330, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " blue" + } + ], + "created": 1753984330, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": ",\n" + } + ], + "created": 1753984330, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "Sugar" + } + ], + "created": 1753984330, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " is" + } + ], + "created": 1753984330, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " sweet" + } + ], + "created": 1753984330, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": ",\n" + } + ], + "created": 1753984331, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": "And" + } + ], + "created": 1753984331, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " so" + } + ], + "created": 1753984331, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " are" + } + ], + "created": 1753984331, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": " you" + } + ], + "created": 1753984331, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", + "choices": [ + { + "finish_reason": null, + "index": 0, + "logprobs": null, + "text": ".\"\n\n" + } + ], + "created": 1753984332, + "model": "llama3.2:3b-instruct-fp16", + "object": "text_completion", + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.completion.Completion", + "__data__": { + "id": "cmpl-447", "choices": [ { "finish_reason": "length", @@ -976,7 +976,7 @@ "text": "" } ], - "created": 1753814833, + "created": 1753984332, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", diff --git a/tests/integration/recordings/responses/85594a69d74a.json b/tests/integration/recordings/responses/85594a69d74a.json index ccfa6136d..252da82e6 100644 --- a/tests/integration/recordings/responses/85594a69d74a.json +++ b/tests/integration/recordings/responses/85594a69d74a.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:56.756402Z", + "created_at": "2025-07-31T17:59:06.788329966Z", "done": true, "done_reason": "stop", - "total_duration": 144652458, - "load_duration": 69191125, + "total_duration": 1049049073, + "load_duration": 42143902, "prompt_eval_count": 213, - "prompt_eval_duration": 63417917, + "prompt_eval_duration": 960926699, "eval_count": 2, - "eval_duration": 11229208, + "eval_duration": 45508201, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/8bba71367e87.json b/tests/integration/recordings/responses/8bba71367e87.json index e381541bb..0cf7b90d5 100644 --- a/tests/integration/recordings/responses/8bba71367e87.json +++ b/tests/integration/recordings/responses/8bba71367e87.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -24,7 +24,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -39,7 +39,7 @@ "logprobs": null } ], - "created": 1753819487, + "created": 1753984093, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -50,423 +50,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "'d", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " be", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " happy", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " help", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " you", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " with", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " current", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " weather", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " in", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Tokyo", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819487, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "!", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " However", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " I", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -481,7 +65,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984093, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -492,11 +76,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " a", + "content": " not", "function_call": null, "refusal": null, "role": "assistant", @@ -507,7 +91,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984093, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -518,11 +102,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " large", + "content": " able", "function_call": null, "refusal": null, "role": "assistant", @@ -533,7 +117,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984093, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -544,11 +128,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " language", + "content": " to", "function_call": null, "refusal": null, "role": "assistant", @@ -559,7 +143,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984094, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -570,11 +154,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " model", + "content": " provide", "function_call": null, "refusal": null, "role": "assistant", @@ -585,7 +169,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984094, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -596,137 +180,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " I", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " don", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "'t", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " have", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -741,7 +195,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984094, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -752,7 +206,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -767,7 +221,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984094, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -778,85 +232,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " access", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " current", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819488, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -871,7 +247,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984094, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -882,11 +258,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " conditions", + "content": " information", "function_call": null, "refusal": null, "role": "assistant", @@ -897,7 +273,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984095, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -908,7 +284,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -923,7 +299,7 @@ "logprobs": null } ], - "created": 1753819488, + "created": 1753984095, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -934,11 +310,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " But", + "content": " However", "function_call": null, "refusal": null, "role": "assistant", @@ -949,7 +325,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984095, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -960,7 +336,33 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ",", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984095, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -975,7 +377,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984095, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -986,7 +388,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1001,7 +403,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984095, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1012,7 +414,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1027,7 +429,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984096, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1038,7 +440,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1053,7 +455,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984096, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1064,7 +466,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1079,7 +481,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984096, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1090,7 +492,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1105,7 +507,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984096, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1116,7 +518,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1131,7 +533,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984096, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1142,7 +544,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1157,7 +559,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984097, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1168,7 +570,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1183,7 +585,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984097, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1194,7 +596,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1209,7 +611,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984097, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1220,7 +622,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1235,7 +637,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984097, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1246,7 +648,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1261,7 +663,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984097, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1272,7 +674,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1287,7 +689,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984097, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1298,7 +700,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1313,7 +715,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984098, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1324,7 +726,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1339,7 +741,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984098, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1350,7 +752,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1365,7 +767,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984098, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1376,7 +778,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1391,7 +793,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984098, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1402,7 +804,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1417,7 +819,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984098, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1428,7 +830,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1443,7 +845,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984099, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1454,7 +856,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1469,7 +871,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984099, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1480,7 +882,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1495,7 +897,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984099, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1506,7 +908,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1521,7 +923,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984099, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1532,7 +934,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1547,7 +949,7 @@ "logprobs": null } ], - "created": 1753819489, + "created": 1753984099, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1558,7 +960,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1573,7 +975,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984100, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1584,7 +986,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1599,7 +1001,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984100, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1610,7 +1012,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1625,7 +1027,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984100, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1636,33 +1038,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " weather", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1677,7 +1053,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984100, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1688,7 +1064,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1703,7 +1079,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984100, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1714,7 +1090,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1729,7 +1105,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984100, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1740,7 +1116,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1755,7 +1131,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984101, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1766,7 +1142,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1781,7 +1157,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984101, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1792,7 +1168,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1807,7 +1183,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984101, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1818,111 +1194,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Dark", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Sky", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " or", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1937,7 +1209,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984101, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1948,7 +1220,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -1963,7 +1235,7 @@ "logprobs": null } ], - "created": 1753819490, + "created": 1753984101, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -1974,813 +1246,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " for", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " current", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " weather", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " conditions", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " and", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " forecast", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " in", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Tokyo", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819490, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ".\n", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "2", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ".", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Use", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " a", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " search", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " engine", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ":", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " You", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " can", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " type", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " \"", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "Tok", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "yo", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " weather", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\"", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " in", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " a", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " search", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " engine", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " like", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Google", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819491, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -2795,7 +1261,7 @@ "logprobs": null } ], - "created": 1753819491, + "created": 1753984102, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2806,11 +1272,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " and", + "content": " or", "function_call": null, "refusal": null, "role": "assistant", @@ -2821,7 +1287,7 @@ "logprobs": null } ], - "created": 1753819491, + "created": 1753984102, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2832,11 +1298,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " it", + "content": " BBC", "function_call": null, "refusal": null, "role": "assistant", @@ -2847,7 +1313,7 @@ "logprobs": null } ], - "created": 1753819492, + "created": 1753984102, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2858,11 +1324,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " will", + "content": " Weather", "function_call": null, "refusal": null, "role": "assistant", @@ -2873,7 +1339,7 @@ "logprobs": null } ], - "created": 1753819492, + "created": 1753984102, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2884,11 +1350,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " show", + "content": " to", "function_call": null, "refusal": null, "role": "assistant", @@ -2899,7 +1365,7 @@ "logprobs": null } ], - "created": 1753819492, + "created": 1753984102, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2910,11 +1376,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " you", + "content": " get", "function_call": null, "refusal": null, "role": "assistant", @@ -2925,7 +1391,7 @@ "logprobs": null } ], - "created": 1753819492, + "created": 1753984102, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2936,7 +1402,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -2951,7 +1417,7 @@ "logprobs": null } ], - "created": 1753819492, + "created": 1753984103, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -2962,865 +1428,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " current", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " weather", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " conditions", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " and", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " forecast", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " for", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Tokyo", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ".\n", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "3", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ".", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Check", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " social", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " media", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ":", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " You", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " can", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " also", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " check", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819492, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " official", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Twitter", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " account", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " of", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Japan", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Meteor", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "ological", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Agency", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "J", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "MA", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ")", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " for", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -3835,7 +1443,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984103, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -3846,7 +1454,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -3861,7 +1469,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984103, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -3872,11 +1480,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " updates", + "content": " forecast", "function_call": null, "refusal": null, "role": "assistant", @@ -3887,7 +1495,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984103, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -3898,11 +1506,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " and", + "content": " for", "function_call": null, "refusal": null, "role": "assistant", @@ -3913,7 +1521,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984103, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -3924,59 +1532,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " forecasts", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " in", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -3991,7 +1547,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984104, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4002,11 +1558,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": ".\n\n", + "content": ".\n", "function_call": null, "refusal": null, "role": "assistant", @@ -4017,7 +1573,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984104, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4028,11 +1584,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": "Please", + "content": "2", "function_call": null, "refusal": null, "role": "assistant", @@ -4043,7 +1599,7 @@ "logprobs": null } ], - "created": 1753819493, + "created": 1753984104, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4054,371 +1610,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " note", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819493, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " that", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " weather", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " in", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Tokyo", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " can", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " change", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " frequently", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " especially", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " during", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " seasons", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -4433,7 +1625,7 @@ "logprobs": null } ], - "created": 1753819494, + "created": 1753984104, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4444,11 +1636,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " Here", + "content": " Use", "function_call": null, "refusal": null, "role": "assistant", @@ -4459,7 +1651,7 @@ "logprobs": null } ], - "created": 1753819494, + "created": 1753984104, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4470,33 +1662,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "'s", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -4511,7 +1677,7 @@ "logprobs": null } ], - "created": 1753819494, + "created": 1753984104, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4522,111 +1688,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " general", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " idea", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " of", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " the", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819494, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -4641,7 +1703,7 @@ "logprobs": null } ], - "created": 1753819494, + "created": 1753984105, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4652,11 +1714,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " patterns", + "content": " app", "function_call": null, "refusal": null, "role": "assistant", @@ -4667,7 +1729,7 @@ "logprobs": null } ], - "created": 1753819494, + "created": 1753984105, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4678,11 +1740,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " in", + "content": ":", "function_call": null, "refusal": null, "role": "assistant", @@ -4693,7 +1755,7 @@ "logprobs": null } ], - "created": 1753819494, + "created": 1753984105, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4704,11 +1766,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " Tokyo", + "content": " Many", "function_call": null, "refusal": null, "role": "assistant", @@ -4719,7 +1781,7 @@ "logprobs": null } ], - "created": 1753819495, + "created": 1753984105, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4730,11 +1792,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": ":\n\n", + "content": " smartphones", "function_call": null, "refusal": null, "role": "assistant", @@ -4745,7 +1807,7 @@ "logprobs": null } ], - "created": 1753819495, + "created": 1753984105, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -4756,2893 +1818,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "*", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Spring", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "March", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " May", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "):", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Mild", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " temperatures", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " with", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " averages", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ranging", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " from", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "12", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0C", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "20", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0C", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819495, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "54", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0F", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "68", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0F", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ").\n", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "*", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Summer", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "June", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " August", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "):", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Hot", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " and", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " humid", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " with", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " temperatures", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " often", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " reaching", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " over", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819496, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "30", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0C", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "86", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0F", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ").\n", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "*", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Autumn", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "September", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " November", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "):", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Cool", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " temperatures", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " with", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " averages", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ranging", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " from", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "10", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0C", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819497, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "20", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0C", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "50", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0F", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "68", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0F", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ").\n", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "*", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Winter", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "December", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " to", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " February", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "):", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Cold", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ",", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " with", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " temperatures", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " often", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " below", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819498, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " ", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "0", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0C", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " (", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "32", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "\u00b0F", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": ").\n\n", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "I", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " hope", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " this", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " helps", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": "!", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " Let", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " me", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " know", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " if", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", - "choices": [ - { - "delta": { - "content": " you", - "function_call": null, - "refusal": null, - "role": "assistant", - "tool_calls": null - }, - "finish_reason": null, - "index": 0, - "logprobs": null - } - ], - "created": 1753819499, - "model": "llama3.2:3b-instruct-fp16", - "object": "chat.completion.chunk", - "service_tier": null, - "system_fingerprint": "fp_ollama", - "usage": null - } - }, - { - "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", - "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -7657,7 +1833,7 @@ "logprobs": null } ], - "created": 1753819499, + "created": 1753984106, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -7668,11 +1844,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " any", + "content": " built", "function_call": null, "refusal": null, "role": "assistant", @@ -7683,7 +1859,7 @@ "logprobs": null } ], - "created": 1753819499, + "created": 1753984106, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -7694,11 +1870,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " other", + "content": "-in", "function_call": null, "refusal": null, "role": "assistant", @@ -7709,7 +1885,7 @@ "logprobs": null } ], - "created": 1753819499, + "created": 1753984106, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -7720,11 +1896,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", "choices": [ { "delta": { - "content": " questions", + "content": " weather", "function_call": null, "refusal": null, "role": "assistant", @@ -7735,7 +1911,7 @@ "logprobs": null } ], - "created": 1753819499, + "created": 1753984106, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -7746,7 +1922,423 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " apps", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984106, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " that", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984107, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " can", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984107, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " provide", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984107, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " real", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984107, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": "-time", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984107, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " weather", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984107, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " information", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984108, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " for", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984108, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " various", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984108, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " locations", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984108, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ",", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984108, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " including", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984109, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " Tokyo", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984109, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ".\n", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984109, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": "3", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984109, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -7761,7 +2353,7 @@ "logprobs": null } ], - "created": 1753819499, + "created": 1753984109, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -7772,7 +2364,1229 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-97", + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " Tune", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984110, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " into", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984110, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " local", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984110, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " news", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984110, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ":", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984110, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " You", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984110, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " can", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984111, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " watch", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984111, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " local", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984111, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " news", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984111, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " programs", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984111, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " or", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984112, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " listen", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984112, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " to", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984112, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " radio", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984112, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " stations", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984112, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " in", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984112, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " Tokyo", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984113, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " to", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984113, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " get", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984113, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " updates", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984113, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " on", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984113, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " the", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984114, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " current", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984114, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " weather", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984114, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " conditions", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984114, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ".\n\n", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984114, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": "If", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984114, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " you", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984115, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": "'d", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984115, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " like", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984115, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ",", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984115, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " I", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984115, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " can", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984116, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " tell", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984116, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " you", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984116, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " about", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984116, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " the", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984116, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " typical", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984117, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " weather", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984117, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " patterns", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984117, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " in", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984117, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " Tokyo", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984117, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " throughout", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984117, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " the", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984118, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": " year", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984118, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", + "choices": [ + { + "delta": { + "content": ".", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984118, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-117", "choices": [ { "delta": { @@ -7787,7 +3601,7 @@ "logprobs": null } ], - "created": 1753819499, + "created": 1753984118, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/90fec951fdb9.json b/tests/integration/recordings/responses/90fec951fdb9.json index 23afe7927..6794a1f7f 100644 --- a/tests/integration/recordings/responses/90fec951fdb9.json +++ b/tests/integration/recordings/responses/90fec951fdb9.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 32707708, - "load_duration": 23390000, + "total_duration": 16580079, + "load_duration": 6456308, "prompt_eval_count": 8, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.054516047, - -0.016456056, - -0.010628294, - 0.022998175, - 0.011771307, - -0.11192805, - -0.009638266, - 0.019111464, - 0.048958372, - -0.040184658, - -0.022362057, - 0.016236247, - 0.009179422, - 0.054799747, - 0.049246185, - -0.095869735, - -0.031108288, - -0.010185289, - -0.02914681, - -0.08954776, - -0.0006788293, - 0.03496997, - 0.016079746, - 0.003440155, - 0.039660316, - -0.016080642, - -0.028411511, - 0.021429215, - 0.046082154, - -0.062199906, - -0.023051145, - 0.10141082, - 0.025186997, - -0.03625052, - -0.032918967, - 0.034433577, - -0.016646268, - -0.066217534, - -0.06070787, - 0.0006243064, - -0.06383077, - 0.0077886702, - -0.005127284, - -0.036702275, - -0.023532037, - 0.074247204, - -0.017199293, - 0.064781435, - -0.00963324, - -0.0011216484, - -0.094671436, - 0.029772488, - -0.0828219, - -0.053136364, - -0.014507852, - -0.015170829, - 0.03712605, - 0.071739994, - -0.018907284, - -0.11193762, - -0.11859575, - 0.029719124, - 0.030655412, - 0.10308374, - -0.027978238, - -0.045611758, - 0.0013704232, - 0.004602404, - 0.032320693, - -0.027153788, - -0.06603313, - -0.015827695, - 0.01920783, - 0.06879109, - 0.047088612, - -0.1058506, - 0.046279814, - -0.030967912, - -0.06984916, - -0.014879451, - -0.0014568317, - 0.026731879, - -0.04702097, - 0.076069675, - 0.05755153, - -0.020301627, - 0.038702164, - 0.06855233, - -0.06817319, - -0.017392006, - 0.057020444, - -0.0795406, - -0.014256318, - 0.0036161602, - -0.05289696, - 0.049625576, - 0.021482797, - 0.034989595, - 0.025457244, - -0.004806878, - 0.051217325, - -0.085426696, - 0.07142323, - 0.04465428, - 0.039311107, - -0.013488202, - 0.07088864, - -0.06598805, - 0.05922822, - -0.023026757, - -0.027465338, - -0.046879534, - -0.03751372, - -0.0085191075, - 0.05315477, - 0.0037932945, - -0.020239882, - 0.043557003, - -0.03434906, - 0.04282584, - -0.007332412, - -0.0016165953, - 0.041878954, - -0.025151564, - -0.0301328, - 0.05601688, - -0.03388191, - -4.802144e-33, - 0.008930927, - -0.10549414, - -0.022485359, - -0.00461374, - 0.10122854, - -0.024063904, - 0.072040126, - 0.00826307, - -0.017573163, - -0.012551788, - 0.011197847, - 0.09432378, - 0.025232295, - 0.061275084, - 0.028605146, - 0.070148624, - -0.028050693, - 0.042055413, - 0.012653081, - 0.051212482, - 0.06987365, - 0.113007665, - 0.063927636, - 0.04614841, - 0.00071471, - -0.04746817, - -0.007670411, - -0.016275087, - -0.039374933, - -0.0060473024, - -0.057836913, - -0.032802302, - 0.030103875, - 0.049495216, - 0.006514002, - -0.015127479, - 0.027406687, - -0.13926439, - 0.04688173, - -0.00014261098, - 0.023295157, - 0.014260961, - 0.00048042598, - -0.019151432, - -0.02166308, - 0.012344319, - -0.03541818, - -0.014996304, - -0.12476534, - 0.017857043, - -0.015367026, - -0.030933712, - 0.0775453, - 0.067932405, - -0.002991927, - 0.034482367, - 0.07207725, - -0.008732087, - -0.0038812195, - -0.048092995, - 0.021236168, - 0.06584243, - 0.07847724, - 0.014562048, - 0.066736475, - 0.07221872, - 0.03357779, - 0.084165, - 0.01657892, - 0.04212138, - -0.059364557, - 0.020403123, - -0.065706775, - 0.045810685, - 0.0029439582, - 0.0034878643, - -0.008467763, - -0.14005418, - 0.056226924, - 0.05473064, - -0.060421, - -0.035074305, - -0.05707729, - -0.0104098, - -0.089569785, - -0.023614792, - 0.0344653, - 0.033663824, - 0.06720568, - -0.0725603, - -0.04185905, - -0.08224899, - 0.010631505, - -0.042881776, - -0.0014539668, - 8.40692e-34, - -0.07032476, - 0.0070766173, - -0.03506184, - 0.021500606, - -0.11258514, - -0.045659322, - 0.08482931, - 0.050339974, - 0.0533988, - 0.01208183, - -0.0019384808, - -0.0860773, - 0.09599927, - 0.0037235345, - 0.060938608, - 0.015288853, - -0.040593054, - 0.10491757, - 0.07109598, - -0.0050172145, - -0.049021836, - 0.091859885, - -0.09862007, - -0.012040684, - -0.016914355, - -0.028067894, - -0.12471722, - -0.078632146, - -0.018693453, - 0.021743925, - 0.0057838396, - 0.051090635, - -0.08270728, - 0.07299018, - 0.014088154, - 0.0010067249, - -0.03681869, - 0.005664378, - 0.017898101, - 0.01379136, - 0.049959406, - 0.021462437, - 0.11088524, - 0.061694097, - 0.018546695, - 0.036211833, - -0.06682083, - 0.036322806, - -0.021121122, - -0.079697676, - 0.065231666, - 0.002995329, - 0.0188468, - -0.008694769, - -0.058170997, - -0.040058907, - 0.051831294, - 0.016280394, - -0.08779952, - -0.022270929, - -0.013231236, - -0.03801554, - 0.0254927, - 0.030549657, - -0.054053955, - 0.040396415, - -0.116118245, - -0.026093038, - -0.004378966, - -0.15024145, - 0.08058958, - -0.05766716, - 0.02520104, - -0.0038984206, - -0.06448939, - 0.020477816, - -0.034754846, - -0.029315596, - -0.052802563, - 0.050487537, - -0.03663958, - -0.009309272, - -0.031305738, - -0.0010610216, - -0.089741714, - 0.0445201, - -0.058746234, - 0.028397618, - 0.057035178, - -0.021242462, - 0.024774676, - 0.023253858, - -0.025503494, - 0.066465355, - 0.011176001, - -1.5780694e-08, - -0.043592602, - 0.050871234, - 0.009062051, - 0.03658537, - 0.002769079, - 0.038917493, - -0.013205564, - 0.006855097, - -0.006784634, - 0.020516934, - -0.029890155, - -0.005596517, - -0.06777992, - -0.05436433, - 0.02436097, - 0.13761573, - -0.07139558, - 0.007746665, - 0.051632155, - 0.059728563, - 0.0424793, - -0.035606194, - -0.05791164, - 0.044417217, - -0.105627485, - 0.009701339, - -0.016052725, - 0.03566595, - 0.023313522, - -0.079250954, - 0.0054293363, - -0.060480006, - -0.044735, - 0.013152052, - -0.015912784, - -0.012098195, - 0.0058634495, - -0.070984975, - 0.017616477, - 0.03611389, - 0.023517592, - -0.007936504, - -0.03601146, - 0.0059993765, - 0.059939068, - 0.0058700717, - -0.05880679, - -0.04119574, - -0.038231015, - -0.030013425, - 0.01916342, - -0.020920184, - -0.008940394, - -0.025874808, - 0.08722286, - 0.042265054, - -0.09463029, - -0.034977533, - 0.05149754, - 0.042541843, - -0.01818799, - 0.06035198, - 0.1938343, - 0.01467125 + -0.054535713, + -0.01647897, + -0.010625582, + 0.023008224, + 0.011759344, + -0.111917414, + -0.009629101, + 0.019088669, + 0.048962217, + -0.040159587, + -0.022340449, + 0.016200319, + 0.009202871, + 0.054802068, + 0.049234793, + -0.09586973, + -0.03113037, + -0.010191666, + -0.029160725, + -0.08953834, + -0.0006810638, + 0.034986537, + 0.016088286, + 0.003424259, + 0.039658964, + -0.016078588, + -0.028393669, + 0.021416757, + 0.046096187, + -0.06216622, + -0.023037015, + 0.1013916, + 0.025191637, + -0.036257233, + -0.03293213, + 0.034415286, + -0.016662724, + -0.06623385, + -0.060700428, + 0.0006065483, + -0.06384223, + 0.0078050094, + -0.0051191156, + -0.03669438, + -0.023559183, + 0.07428184, + -0.01721832, + 0.06481015, + -0.009654798, + -0.0010900846, + -0.09464753, + 0.029771274, + -0.08282523, + -0.05314006, + -0.01452814, + -0.015171761, + 0.037140947, + 0.07173332, + -0.018937344, + -0.11191488, + -0.11860731, + 0.029728845, + 0.030656187, + 0.103080384, + -0.027999118, + -0.04562904, + 0.0013753629, + 0.0045906254, + 0.032285035, + -0.027159423, + -0.066034995, + -0.015820919, + 0.019237159, + 0.06880823, + 0.04709089, + -0.10583619, + 0.04628416, + -0.030952096, + -0.06985726, + -0.01489977, + -0.0014784886, + 0.026742237, + -0.047020886, + 0.07605717, + 0.057562124, + -0.020296026, + 0.038699575, + 0.068533406, + -0.06815694, + -0.017370922, + 0.057017475, + -0.079542115, + -0.014235115, + 0.003623275, + -0.052890334, + 0.04962317, + 0.021482611, + 0.03502143, + 0.025463292, + -0.0048156767, + 0.051208895, + -0.08542178, + 0.07141962, + 0.044664312, + 0.03932115, + -0.013527642, + 0.07087505, + -0.06600328, + 0.05924568, + -0.023052538, + -0.027463019, + -0.046869196, + -0.03748807, + -0.008505666, + 0.053133655, + 0.0037986652, + -0.020212771, + 0.043556064, + -0.034358878, + 0.042805903, + -0.0073572295, + -0.0015867023, + 0.04185437, + -0.02514704, + -0.030102273, + 0.05600693, + -0.033871204, + -4.8017078e-33, + 0.008951275, + -0.10545955, + -0.022469193, + -0.0046515064, + 0.101191446, + -0.024060266, + 0.0720429, + 0.008255562, + -0.017577993, + -0.012554701, + 0.011185812, + 0.094320215, + 0.02520196, + 0.06127928, + 0.028601166, + 0.0701666, + -0.028017957, + 0.042041674, + 0.012637978, + 0.051187877, + 0.06988971, + 0.11301735, + 0.06395204, + 0.04612332, + 0.0007098928, + -0.047456842, + -0.007656803, + -0.016268259, + -0.039380968, + -0.0060521755, + -0.057859622, + -0.032785412, + 0.030082524, + 0.049507707, + 0.0064935936, + -0.015132811, + 0.027426424, + -0.1392769, + 0.046868317, + -0.00012954268, + 0.023331605, + 0.014258741, + 0.00046936277, + -0.019173825, + -0.021654885, + 0.012334302, + -0.035438117, + -0.015027805, + -0.12477716, + 0.017864512, + -0.0153814005, + -0.030930284, + 0.07757873, + 0.06792478, + -0.0030147473, + 0.034457013, + 0.072105244, + -0.008725219, + -0.0039085858, + -0.04809878, + 0.021228116, + 0.06583196, + 0.078512274, + 0.014584778, + 0.06673733, + 0.07221583, + 0.033577427, + 0.084169276, + 0.016580611, + 0.042122263, + -0.05934277, + 0.020412177, + -0.06569441, + 0.045803223, + 0.0029223328, + 0.0034806612, + -0.008458956, + -0.14005856, + 0.056258015, + 0.0547175, + -0.060439304, + -0.03508705, + -0.057089172, + -0.010407182, + -0.0895699, + -0.023597935, + 0.03446976, + 0.033670787, + 0.06720999, + -0.0725774, + -0.041841872, + -0.08223708, + 0.010613598, + -0.042911, + -0.0014353823, + 8.40787e-34, + -0.07033168, + 0.0070736655, + -0.035051774, + 0.0215116, + -0.11255857, + -0.045679986, + 0.08481424, + 0.050349806, + 0.053389013, + 0.012061718, + -0.0019460444, + -0.08606971, + 0.09598181, + 0.0037098164, + 0.060968198, + 0.015265811, + -0.040588457, + 0.1049424, + 0.07111727, + -0.0050206287, + -0.048998516, + 0.091846675, + -0.09864027, + -0.0120658465, + -0.016912753, + -0.028071038, + -0.12469634, + -0.07860433, + -0.018691944, + 0.021778468, + 0.0057638944, + 0.051103488, + -0.08267645, + 0.07295661, + 0.014084549, + 0.0009908162, + -0.0368374, + 0.0056820577, + 0.017890759, + 0.013773187, + 0.04993143, + 0.021441808, + 0.11090027, + 0.0616793, + 0.01855691, + 0.03620899, + -0.066839695, + 0.03632399, + -0.021137934, + -0.07972023, + 0.06525973, + 0.0030076413, + 0.018850671, + -0.0086968895, + -0.05816279, + -0.040063396, + 0.0518331, + 0.01626531, + -0.087793276, + -0.022269959, + -0.013210075, + -0.038003217, + 0.025518952, + 0.030541278, + -0.05405989, + 0.040391784, + -0.116105065, + -0.026072625, + -0.004357362, + -0.15028392, + 0.080575064, + -0.05765118, + 0.025196219, + -0.003874716, + -0.064485975, + 0.020491246, + -0.03476756, + -0.029334074, + -0.05278849, + 0.05050434, + -0.03664279, + -0.009323289, + -0.031295024, + -0.0010326867, + -0.08972744, + 0.044515517, + -0.058742493, + 0.028400177, + 0.05706043, + -0.021254258, + 0.024761314, + 0.023270002, + -0.025499929, + 0.06648831, + 0.011162858, + -1.5781294e-08, + -0.043614857, + 0.050878897, + 0.009046982, + 0.036620628, + 0.0027418374, + 0.038922437, + -0.013206618, + 0.006840255, + -0.006766401, + 0.020513676, + -0.0298916, + -0.0055974717, + -0.067749254, + -0.054363597, + 0.024360858, + 0.13762148, + -0.07138463, + 0.007756891, + 0.05164215, + 0.05973908, + 0.04245239, + -0.0355896, + -0.05790089, + 0.04440935, + -0.10564156, + 0.009710743, + -0.016090317, + 0.03566852, + 0.023329897, + -0.07925744, + 0.0054290486, + -0.060488198, + -0.044711344, + 0.013145073, + -0.015926179, + -0.012088508, + 0.0058591575, + -0.0709944, + 0.017609226, + 0.03612532, + 0.023530722, + -0.007937178, + -0.036036156, + 0.0060040886, + 0.059937168, + 0.0058717513, + -0.05882341, + -0.041207276, + -0.03821708, + -0.030004062, + 0.019161234, + -0.02088437, + -0.008930527, + -0.025869865, + 0.08722182, + 0.042273775, + -0.09460558, + -0.034966514, + 0.051498678, + 0.042548534, + -0.01819233, + 0.06034196, + 0.19383828, + 0.014679242 ] ] } diff --git a/tests/integration/recordings/responses/97d3812bfccb.json b/tests/integration/recordings/responses/97d3812bfccb.json index 450c3e1d9..2562216d4 100644 --- a/tests/integration/recordings/responses/97d3812bfccb.json +++ b/tests/integration/recordings/responses/97d3812bfccb.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:13.069497Z", + "created_at": "2025-07-31T17:44:42.081968601Z", "done": true, "done_reason": "stop", - "total_duration": 296277708, - "load_duration": 68882333, + "total_duration": 2973322204, + "load_duration": 42260755, "prompt_eval_count": 217, - "prompt_eval_duration": 185153000, + "prompt_eval_duration": 2760398671, "eval_count": 5, - "eval_duration": 41690709, + "eval_duration": 170139491, "response": "unsafe\nS1", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/97e259c0d3e5.json b/tests/integration/recordings/responses/97e259c0d3e5.json index aac1f6b3c..da4646193 100644 --- a/tests/integration/recordings/responses/97e259c0d3e5.json +++ b/tests/integration/recordings/responses/97e259c0d3e5.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.586434Z", + "created_at": "2025-07-31T17:44:48.550106343Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.628283Z", + "created_at": "2025-07-31T17:44:48.733721083Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.670142Z", + "created_at": "2025-07-31T17:44:48.914478642Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.712246Z", + "created_at": "2025-07-31T17:44:49.096327727Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.754828Z", + "created_at": "2025-07-31T17:44:49.276270052Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.79698Z", + "created_at": "2025-07-31T17:44:49.455584591Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.839206Z", + "created_at": "2025-07-31T17:44:49.635566453Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.88144Z", + "created_at": "2025-07-31T17:44:49.814260821Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.92423Z", + "created_at": "2025-07-31T17:44:49.994127858Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:13.966797Z", + "created_at": "2025-07-31T17:44:50.176174312Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.009087Z", + "created_at": "2025-07-31T17:44:50.363377723Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.050988Z", + "created_at": "2025-07-31T17:44:50.543615455Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.093655Z", + "created_at": "2025-07-31T17:44:50.722758469Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.136425Z", + "created_at": "2025-07-31T17:44:50.904084798Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.179625Z", + "created_at": "2025-07-31T17:44:51.085643411Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.22262Z", + "created_at": "2025-07-31T17:44:51.27077086Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.268355Z", + "created_at": "2025-07-31T17:44:51.4518987Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.31532Z", + "created_at": "2025-07-31T17:44:51.633427977Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,15 +346,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:14.358392Z", + "created_at": "2025-07-31T17:44:51.814721336Z", "done": true, "done_reason": "stop", - "total_duration": 948553666, - "load_duration": 64505458, + "total_duration": 6854120324, + "load_duration": 41359627, "prompt_eval_count": 384, - "prompt_eval_duration": 110383875, + "prompt_eval_duration": 3546083190, "eval_count": 19, - "eval_duration": 772755125, + "eval_duration": 3266092572, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/9b812cbcb88d.json b/tests/integration/recordings/responses/9b812cbcb88d.json index 75cd80a6d..b9a8ae5b2 100644 --- a/tests/integration/recordings/responses/9b812cbcb88d.json +++ b/tests/integration/recordings/responses/9b812cbcb88d.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:30.907069Z", + "created_at": "2025-07-31T17:54:04.489140565Z", "done": true, "done_reason": "stop", - "total_duration": 978723208, - "load_duration": 82950875, + "total_duration": 29123702298, + "load_duration": 40831143, "prompt_eval_count": 324, - "prompt_eval_duration": 453827625, + "prompt_eval_duration": 27253456381, "eval_count": 11, - "eval_duration": 439485709, + "eval_duration": 1828867055, "response": "[get_weather(location=\"San Francisco, CA\")]", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/9c140a29ae09.json b/tests/integration/recordings/responses/9c140a29ae09.json index d817f1e9c..887a1185c 100644 --- a/tests/integration/recordings/responses/9c140a29ae09.json +++ b/tests/integration/recordings/responses/9c140a29ae09.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.506573Z", + "created_at": "2025-07-31T17:44:59.929905409Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.555673Z", + "created_at": "2025-07-31T17:45:00.110734553Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.60425Z", + "created_at": "2025-07-31T17:45:00.290949416Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.650587Z", + "created_at": "2025-07-31T17:45:00.475960339Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.698731Z", + "created_at": "2025-07-31T17:45:00.659481498Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.750779Z", + "created_at": "2025-07-31T17:45:00.843632147Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.800299Z", + "created_at": "2025-07-31T17:45:01.033425275Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.849125Z", + "created_at": "2025-07-31T17:45:01.224890954Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.896216Z", + "created_at": "2025-07-31T17:45:01.412700377Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.942094Z", + "created_at": "2025-07-31T17:45:01.602294122Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:15.985438Z", + "created_at": "2025-07-31T17:45:01.79213394Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:16.033126Z", + "created_at": "2025-07-31T17:45:01.982631333Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,15 +238,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:16.082319Z", + "created_at": "2025-07-31T17:45:02.172457365Z", "done": true, "done_reason": "stop", - "total_duration": 755349958, - "load_duration": 97536083, + "total_duration": 4304742260, + "load_duration": 41766451, "prompt_eval_count": 415, - "prompt_eval_duration": 78861250, + "prompt_eval_duration": 2018894885, "eval_count": 13, - "eval_duration": 578291875, + "eval_duration": 2243484258, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/9c4bc9c3e7ac.json b/tests/integration/recordings/responses/9c4bc9c3e7ac.json index 0072c87c2..e28052a74 100644 --- a/tests/integration/recordings/responses/9c4bc9c3e7ac.json +++ b/tests/integration/recordings/responses/9c4bc9c3e7ac.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 51147375, - "load_duration": 33379959, + "total_duration": 16658862, + "load_duration": 6238467, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.055990793, - 0.076004684, - -0.09247725, - 0.014340361, - 0.058780864, - -0.032434482, - 0.020954052, - 0.028818125, - -0.06591213, - 0.013541593, - 0.12999941, - 0.004603084, - -0.0069239275, - -0.055457443, - -0.047553156, - -0.029139794, - -0.12236376, - -0.05360872, - -0.014706594, - 0.05984688, - 0.034442738, - 0.02076038, - -0.048697792, - 0.0135388365, - 0.058592733, - -0.003076384, - -0.031565297, - 0.082541116, - -0.031259205, - -0.12057633, - 0.038319625, - 0.06574785, - 0.06415721, - 0.038382582, - 0.12570712, - 0.03108174, - 0.10821103, - -0.0019794356, - -0.024704305, - 0.028765837, - 0.01268161, - -0.039844505, - 0.043253522, - -0.015898596, - -0.0135526005, - -0.0050831717, - -0.007911988, - 0.039783813, - 0.0036548872, - -0.033632487, - -0.058547974, - 0.0048877494, - -0.089586094, - -0.010457663, - 0.059202507, - -0.020414542, - 0.014278556, - 0.013986488, - -0.0046022516, - 0.0383391, - 0.0048145773, - 0.029772853, - -0.020863408, - 0.018640704, - 0.12422993, - -0.023236223, - -0.040323637, - -0.023598222, - -0.007448043, - -0.09083128, - -0.16859712, - 0.01012451, - -0.035808884, - 0.010595173, - -0.02050494, - 0.0020821376, - -0.10925222, - 0.00793264, - 0.048889533, - -0.11391199, - -0.06072707, - -0.13435508, - 0.0063265716, - -0.008838073, - -0.03153269, - 0.099169336, - 0.055310693, - 0.0068571265, - -0.023463152, - -0.0031599961, - 0.036782328, - 0.014336826, - 0.022220163, - 0.047114056, - 0.007079763, - 0.06806425, - 0.01851431, - 0.040882625, - 0.055058856, - 0.09488346, - -0.015833577, - -7.924328e-05, - 0.010821554, - 0.09177704, - -0.07464829, - -0.06471165, - 0.07013805, - -0.04499751, - 0.057702336, - -0.0260911, - 0.006323043, - -0.09500501, - -0.010549514, - -0.07887475, - 0.039744847, - -0.04154404, - -0.055268157, - 0.07540271, - -0.04667509, - 0.036143072, - 0.080297194, - -0.036381353, - -0.03477274, - 0.01701203, - -0.047007203, - -0.06519774, - 0.062141683, - -4.222482e-33, - -0.0017580023, - -0.09383388, - -0.02982657, - 0.1257841, - 0.03802007, - -0.03654342, - 0.0060920226, - 0.05906885, - -0.11074452, - 0.005664566, - -0.0259852, - -0.074819505, - 0.008342821, - 0.027451068, - -0.05248069, - 0.02401768, - -0.004380289, - 0.039321493, - -0.04213744, - -0.027290314, - 0.054677974, - 0.02707243, - -0.03329442, - -0.060589895, - -0.050737355, - 0.017969057, - -0.0035060972, - -0.04666249, - 0.073946096, - 0.01333894, - -0.0033873583, - -0.046544433, - -0.060105033, - 0.03406923, - 0.001542676, - 0.039177947, - 0.03989323, - -0.012346489, - -0.030511485, - -0.0019157606, - -0.014608986, - -0.012997742, - 0.019522104, - -0.022349002, - 0.074362256, - -0.053366993, - -0.023993475, - 0.029225096, - 0.027534606, - 0.015111057, - -0.020442221, - 0.043327376, - 0.019660354, - 0.017330697, - -0.0035011724, - 0.019482937, - -0.0003428041, - 0.0004143988, - -0.005117252, - 0.06624799, - 0.027922852, - 0.041020587, - -0.067166425, - 0.028737254, - -0.03478325, - -0.055551115, - -0.032713737, - -0.08099247, - 0.09216284, - 0.06395264, - -0.049168136, - -0.039908994, - 0.036915958, - -0.001602359, - 0.00033041168, - -0.026015632, - -0.005999889, - 0.05474541, - -0.09568287, - -0.05186289, - -0.048838183, - -0.08639551, - -0.034023147, - -0.033257127, - -0.05651867, - -0.051131375, - 0.00809173, - -0.08581851, - 0.06507323, - -0.085427366, - 0.027997404, - 0.029847065, - -0.031673994, - -0.08560956, - 0.1017672, - 2.1855676e-33, - 0.01160785, - 0.077607885, - -0.017380483, - 0.005239329, - 0.0009684126, - 0.06543702, - 0.07256893, - -0.044318836, - -0.04749324, - 0.14031002, - -0.025741624, - 0.0057860985, - 0.040946104, - -0.054880083, - 0.074413285, - -0.023610368, - 0.018364722, - -0.060585637, - -0.044149306, - 0.0027854694, - -0.04580664, - 0.1172219, - 0.10268574, - 0.07907412, - -0.0466143, - 0.018618405, - 0.029834948, - 0.037265483, - 0.02273822, - -0.0026589038, - 0.041726097, - 0.06439532, - -0.089163445, - 0.018188318, - 0.024064727, - -0.096389584, - 0.08642254, - -0.05389359, - 0.01923105, - 0.045092683, - 0.045125954, - 0.09655961, - 0.014908797, - 0.059611585, - 0.03066662, - 0.05882299, - 0.111484826, - 0.016632542, - 0.011590394, - -0.023702666, - -0.008617484, - -0.055030316, - 0.047606383, - -0.014632687, - -0.014156344, - 0.069926, - 0.032047603, - 0.042642817, - -0.053942375, - 0.031047028, - 0.009216673, - 0.033024028, - -0.019033706, - 0.005568194, - -0.014985451, - -0.09193244, - -0.03210824, - 0.015367608, - 0.029150328, - 0.01250386, - -0.004827391, - 0.023345906, - -0.028271332, - -0.08454125, - 0.051068563, - -0.0133641455, - -0.029022738, - -0.02258452, - 0.010884119, - -0.009810021, - 0.049751773, - -0.0032637494, - -0.038813565, - 0.027924104, - 0.017925078, - 0.005337612, - 0.058691237, - 0.09577674, - -0.014308608, - 0.006972794, - -0.02733344, - 0.06912433, - 0.05727631, - 0.03206042, - 0.0042422824, - -1.6766318e-08, - -0.036354303, - -0.09146416, - -0.026319364, - -0.007941995, - -0.024127059, - 0.09896698, - -0.04723083, - -0.03767135, - -0.029419973, - -0.022513283, - 0.04125822, - -0.0011487947, - -0.05570366, - 0.020679709, - -0.038118906, - -0.0524994, - -0.02624128, - -0.05336954, - -0.040593866, - -0.0073642326, - -0.0014442836, - 0.02714257, - 0.027141048, - 0.00932513, - -0.00026505854, - 0.038233075, - 0.037096914, - 0.08405413, - -0.06340637, - -0.014856458, - 0.05038612, - 0.06703033, - 0.027668556, - -0.04360097, - -0.012041474, - 0.08500689, - 0.111594744, - 0.1046117, - 0.019726463, - -0.0003025109, - -0.04110389, - 0.009575226, - -0.05285304, - -0.0026365265, - -0.031144748, - -0.08860188, - -0.06762232, - -0.07451522, - -0.053012833, - -0.09560941, - -0.05273455, - 0.013032144, - 0.0029190276, - 0.041905046, - -0.04522114, - 0.016730292, - 0.017214278, - 0.021578068, - -0.03718778, - 0.02353425, - 0.052041385, - 0.06444499, - 0.02387539, - -0.025236009 + -0.055982195, + 0.076081045, + -0.092433155, + 0.014348906, + 0.058773536, + -0.032442085, + 0.020963855, + 0.028817905, + -0.06589273, + 0.013538555, + 0.12999825, + 0.0045920787, + -0.006909012, + -0.055475082, + -0.047584433, + -0.029108394, + -0.12240743, + -0.053608917, + -0.01469641, + 0.059820697, + 0.034437098, + 0.020720147, + -0.04865705, + 0.013548686, + 0.0586202, + -0.0030596491, + -0.03157386, + 0.08258042, + -0.03125673, + -0.12059294, + 0.038315985, + 0.06574817, + 0.06415242, + 0.038379226, + 0.12571476, + 0.031110935, + 0.1082007, + -0.001945952, + -0.024701951, + 0.028813468, + 0.0127212815, + -0.039838783, + 0.043286823, + -0.01592365, + -0.013537497, + -0.005082726, + -0.007908334, + 0.03983111, + 0.003692527, + -0.03362251, + -0.058581255, + 0.0048970715, + -0.08960359, + -0.010458434, + 0.059172295, + -0.020381752, + 0.014312179, + 0.013974074, + -0.004614098, + 0.038257506, + 0.0048046876, + 0.029754879, + -0.020845208, + 0.018606082, + 0.12421981, + -0.0232267, + -0.040324386, + -0.023599995, + -0.0074541806, + -0.090772845, + -0.16855139, + 0.010115335, + -0.035782356, + 0.010589896, + -0.020527367, + 0.0020988148, + -0.10928735, + 0.007924992, + 0.04890187, + -0.113910094, + -0.060739312, + -0.13437672, + 0.00634339, + -0.008847756, + -0.031561375, + 0.09914905, + 0.055312507, + 0.0068518873, + -0.023520693, + -0.0031644772, + 0.03681182, + 0.01432514, + 0.022244632, + 0.047091875, + 0.007081216, + 0.06810163, + 0.01849201, + 0.040876437, + 0.0550351, + 0.09491265, + -0.015807496, + -7.8463614e-05, + 0.010782611, + 0.09179502, + -0.074678205, + -0.06468329, + 0.07012851, + -0.04499402, + 0.057698727, + -0.026082484, + 0.006340654, + -0.095005274, + -0.010555381, + -0.07885503, + 0.039741103, + -0.04152419, + -0.055226922, + 0.07541398, + -0.046700906, + 0.036149766, + 0.08029784, + -0.036375977, + -0.03479682, + 0.017001636, + -0.047019735, + -0.065193616, + 0.062124435, + -4.2218427e-33, + -0.0017721883, + -0.09384802, + -0.029816238, + 0.1257514, + 0.037986845, + -0.036565077, + 0.0060834913, + 0.059064236, + -0.11073993, + 0.0056816707, + -0.02600264, + -0.074865155, + 0.008368443, + 0.027423527, + -0.052494608, + 0.024021218, + -0.004405381, + 0.039293334, + -0.042134702, + -0.02727807, + 0.0547222, + 0.027060617, + -0.033269312, + -0.06058816, + -0.05074509, + 0.017971171, + -0.0035045573, + -0.04667931, + 0.073935926, + 0.013332051, + -0.0033654193, + -0.04656567, + -0.060090553, + 0.034060493, + 0.0015471254, + 0.039180268, + 0.03989967, + -0.012337066, + -0.030525556, + -0.0019079247, + -0.01460898, + -0.012969273, + 0.0195347, + -0.022350095, + 0.07438301, + -0.053325966, + -0.02395582, + 0.0292596, + 0.027539466, + 0.015120207, + -0.020437608, + 0.04332795, + 0.019630946, + 0.01733148, + -0.0034844875, + 0.019485317, + -0.0003494216, + 0.00042686076, + -0.005132303, + 0.066248745, + 0.027956294, + 0.04101923, + -0.06715309, + 0.028716685, + -0.034805898, + -0.0555862, + -0.032694634, + -0.081002966, + 0.092122965, + 0.06396523, + -0.049184497, + -0.03989705, + 0.0369496, + -0.0015811125, + 0.0003197812, + -0.026036698, + -0.0060006436, + 0.054738022, + -0.09564789, + -0.051893853, + -0.048824586, + -0.086377576, + -0.034007866, + -0.033272535, + -0.05652991, + -0.051139913, + 0.008112306, + -0.08580783, + 0.06505675, + -0.08542722, + 0.028040707, + 0.029830497, + -0.031668846, + -0.085617274, + 0.101759925, + 2.1862516e-33, + 0.0116090225, + 0.07760366, + -0.01735762, + 0.0052205133, + 0.000975667, + 0.065414004, + 0.07253235, + -0.04430889, + -0.047520436, + 0.14030467, + -0.025719156, + 0.005793378, + 0.04094832, + -0.054858785, + 0.07442516, + -0.023601599, + 0.018334351, + -0.060549837, + -0.044134542, + 0.0027906233, + -0.04578033, + 0.117253944, + 0.10268663, + 0.07909348, + -0.04660703, + 0.01862913, + 0.029805424, + 0.03725584, + 0.022773262, + -0.002695987, + 0.041698016, + 0.06442998, + -0.08914443, + 0.018201683, + 0.024052765, + -0.096402094, + 0.086406566, + -0.053927243, + 0.01922541, + 0.045094177, + 0.04512779, + 0.09656579, + 0.014912764, + 0.05959895, + 0.030681113, + 0.058840565, + 0.11149792, + 0.016654478, + 0.011594341, + -0.023696017, + -0.008650225, + -0.055055793, + 0.04757928, + -0.014643343, + -0.014162865, + 0.06991306, + 0.03207973, + 0.0426524, + -0.05397539, + 0.031076223, + 0.009192395, + 0.03302898, + -0.01901568, + 0.0055926023, + -0.014975381, + -0.09193982, + -0.032133788, + 0.015386497, + 0.029159075, + 0.0125140045, + -0.0048286216, + 0.023340749, + -0.028264781, + -0.08455668, + 0.051091656, + -0.013328911, + -0.029051399, + -0.022525461, + 0.0108897155, + -0.009800699, + 0.049736347, + -0.0032429877, + -0.03881739, + 0.027952043, + 0.017933605, + 0.0053533562, + 0.058684465, + 0.0957579, + -0.014328832, + 0.007005975, + -0.02730476, + 0.06912227, + 0.05728153, + 0.032041542, + 0.0042575295, + -1.6768182e-08, + -0.036365103, + -0.091483496, + -0.026359037, + -0.007929498, + -0.02413145, + 0.09893336, + -0.047247022, + -0.03765342, + -0.029432135, + -0.022519268, + 0.04128326, + -0.0011647358, + -0.055700593, + 0.02070786, + -0.0381106, + -0.052463897, + -0.02626158, + -0.053371817, + -0.040604327, + -0.007350147, + -0.0014813344, + 0.027132856, + 0.027119234, + 0.009304667, + -0.0002654552, + 0.03823097, + 0.03706377, + 0.08402369, + -0.06339772, + -0.014845903, + 0.05038436, + 0.06699758, + 0.027672214, + -0.04359535, + -0.012074228, + 0.08498164, + 0.111630745, + 0.10461064, + 0.01975323, + -0.0002619162, + -0.041080646, + 0.00960582, + -0.052899837, + -0.0026609222, + -0.031148938, + -0.08861712, + -0.06763032, + -0.074505664, + -0.05304462, + -0.09555963, + -0.052730557, + 0.013020395, + 0.0029235827, + 0.04188656, + -0.045235366, + 0.016710248, + 0.017224982, + 0.021616042, + -0.0371995, + 0.023536215, + 0.05206289, + 0.0644171, + 0.023864746, + -0.02523672 ] ] } diff --git a/tests/integration/recordings/responses/9e7a83d3d596.json b/tests/integration/recordings/responses/9e7a83d3d596.json index deb223dba..8b7b0dca5 100644 --- a/tests/integration/recordings/responses/9e7a83d3d596.json +++ b/tests/integration/recordings/responses/9e7a83d3d596.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -15,23 +15,23 @@ "body": { "__type__": "openai.types.completion.Completion", "__data__": { - "id": "cmpl-719", + "id": "cmpl-726", "choices": [ { "finish_reason": "stop", "index": 0, "logprobs": null, - "text": "Blue.\n\nExplanation: This is a classic example of an alliterative poem, often referred to as \"red roses.\" The original phrase, \"Roses are red,\" was actually coined by Ernest Thesiger in 1910 and was followed by the complementary phrase, making the complete sentence a poetic device called an \"alliterative couplet.\"" + "text": "Blue.\n\nExplanation: This is a play on words from the classic Valentine's Day poem \"Roses are red, violets are blue.\" The original poem typically ends with the line \"sent my love to you,\" but in this adaptation, I've altered it to rhyme by replacing \"blue\" with \"blue\" (a pun). However, it also makes a clever non-literary response." } ], - "created": 1753814830, + "created": 1753984322, "model": "llama3.2:3b-instruct-fp16", "object": "text_completion", "system_fingerprint": "fp_ollama", "usage": { - "completion_tokens": 71, + "completion_tokens": 81, "prompt_tokens": 50, - "total_tokens": 121, + "total_tokens": 131, "completion_tokens_details": null, "prompt_tokens_details": null } diff --git a/tests/integration/recordings/responses/9fadf5a3d68f.json b/tests/integration/recordings/responses/9fadf5a3d68f.json index 0ce5870f0..8d6e85d2e 100644 --- a/tests/integration/recordings/responses/9fadf5a3d68f.json +++ b/tests/integration/recordings/responses/9fadf5a3d68f.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:25.942494Z", + "created_at": "2025-07-31T17:46:18.2684163Z", "done": true, "done_reason": "stop", - "total_duration": 282345834, - "load_duration": 106002125, + "total_duration": 2939701988, + "load_duration": 43760413, "prompt_eval_count": 224, - "prompt_eval_duration": 148628959, + "prompt_eval_duration": 2854863241, "eval_count": 2, - "eval_duration": 26498000, + "eval_duration": 40546191, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/a410d4840402.json b/tests/integration/recordings/responses/a410d4840402.json index cfb29d78b..ab9f03746 100644 --- a/tests/integration/recordings/responses/a410d4840402.json +++ b/tests/integration/recordings/responses/a410d4840402.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 33714959, - "load_duration": 17011709, + "total_duration": 14442225, + "load_duration": 5908877, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.003961408, - 0.051414188, - -0.00058039324, - -0.03805786, - 0.00026862609, - -0.07164569, - -0.032947958, - 0.029143414, - 0.0895043, - 0.027018296, - 0.022992423, - 0.029479899, - 0.013462918, - 0.021877697, - 0.024697151, - 0.023186686, - -0.06790505, - 0.042193525, - -0.0668863, - -0.04484601, - -0.019504927, - -0.017638002, - -0.047011577, - 0.010105266, - -0.035193082, - 0.12793653, - -0.03992006, - -0.03702981, - 0.021819357, - -0.06665871, - 0.020533124, - 0.03142357, - 0.121719204, - 0.037876442, - -0.075640336, - 0.0359664, - 0.11100785, - -0.02567441, - -0.07788109, - 0.016981006, - -0.08081605, - 0.042523988, - 0.008232587, - 0.0731737, - 0.011123085, - 0.016207846, - 0.01944517, - -0.057269264, - -0.026940528, - 0.027561199, - -0.103662655, - 0.06181235, - -0.028062372, - 0.04553612, - 0.038513146, - 0.10225101, - 0.010200513, - 0.003872203, - -0.074381135, - -0.0097752875, - -0.014599097, - 0.0054576746, - -0.04897588, - 0.024681844, - 0.08043012, - -0.0014103616, - 0.0008604012, - 0.0016741438, - 0.016251745, - 0.00360708, - 0.058014695, - -0.010049014, - -0.0084027, - 0.06814959, - 0.033971835, - -0.011656133, - -0.04935883, - -0.03459291, - 0.022477727, - 0.01610207, - 0.025287844, - 0.03501659, - -0.018194117, - 0.06807382, - 0.059983365, - -0.025374522, - 0.04583719, - -0.04297365, - -0.104865946, - -0.028109012, - 0.079001896, - -0.017114554, - 0.012419278, - 0.04061318, - -0.020101532, - 0.026956845, - 0.041828763, - -0.044170532, - 0.08095696, - 0.021788325, - 0.081747636, - 0.033276387, - 0.021741632, - 0.092068955, - -0.05207143, - -0.13620017, - 0.013549487, - -0.019821124, - -0.036206715, - -0.050286006, - -0.032959178, - 0.04662646, - -0.062424622, - -0.056837536, - -0.027646665, - -0.15120761, - -0.093959294, - -0.010999317, - -0.02427833, - -0.046769585, - -0.002897303, - -0.06647176, - -0.025597623, - 0.018255977, - 0.0020313214, - -0.06226326, - -0.117481604, - -4.4295206e-33, - -0.009129055, - -0.037181977, - -0.02604801, - 0.052037112, - 0.00087297254, - 0.0065994835, - -0.0045263134, - -0.040167294, - 0.0041152886, - 0.042845216, - -0.049708433, - 0.045345027, - 0.04285296, - 0.044911012, - 0.11100636, - 0.021593297, - -0.03125754, - 0.072277226, - -0.01916381, - -0.03471753, - 0.06770263, - -0.016145714, - 0.05970865, - -0.02298266, - 0.028831182, - 0.015415605, - -0.00031274176, - -0.012733097, - -0.03328956, - -0.00013622487, - -0.024770694, - -0.042212497, - -0.0024302523, - 0.04124051, - 0.09191475, - 0.06856497, - -0.015284932, - -0.12650564, - 0.017038988, - -0.086213395, - 0.05503028, - 0.030287316, - 0.0043085497, - 0.03199775, - -0.032243066, - 0.004920853, - 0.009013211, - -0.023148343, - -0.04070659, - -0.091041416, - 0.036388315, - 0.024427423, - 0.013590955, - 0.032416057, - 0.040976506, - 0.037508775, - -0.041537814, - -0.0790035, - -0.05377612, - 0.06448428, - -0.080218546, - 0.021294411, - 0.062302276, - 0.045776673, - 0.032483075, - 0.08931608, - -0.04060625, - -0.031852096, - 0.09785858, - 0.01842136, - 0.005539284, - 0.033401128, - -0.069316946, - 0.0050071795, - -0.01113226, - 0.04040353, - -0.018702384, - -0.061634906, - -0.019955046, - 0.055725593, - -0.0339558, - -0.03284888, - 0.039789777, - 0.032518264, - -0.014831044, - -0.040828414, - 0.09042645, - -0.07117855, - -0.0452999, - 0.004429679, - -0.011286574, - 0.010456636, - -0.005107356, - -0.03228427, - -0.014561991, - 1.973978e-33, - -0.014741807, - -0.011373571, - -0.018968971, - -0.030024195, - -0.032379575, - 0.00021643718, - -0.012567692, - -0.121494584, - 0.0020773544, - 0.03192013, - -0.004760303, - 0.0094626825, - 0.070903994, - -0.10057645, - 0.025073227, - 0.0619163, - -0.0040503214, - -0.099229865, - -0.011797051, - -0.04770035, - -0.030485118, - 0.06268395, - -0.073855996, - -0.0061467164, - -0.01423362, - 0.0073681897, - -0.12381955, - -0.12358002, - 0.049814835, - 0.013639601, - -0.04231122, - -0.057728436, - 0.008867639, - -0.03936158, - -0.010378862, - 0.01995126, - 0.06864242, - -0.0034683226, - 0.034935873, - 0.01691657, - -0.041248, - 0.12756771, - -0.0109369, - -0.038407195, - 0.03351686, - 0.024284633, - -0.009186648, - 0.089450404, - -0.037300985, - -0.033677705, - 0.083595864, - 0.024388704, - 0.013052032, - -0.082466476, - 0.08174954, - 0.025851287, - -0.0407412, - 0.011634866, - 0.045149248, - 0.057999264, - -0.043137826, - -0.0218611, - 0.007614091, - 0.075013876, - -0.037117332, - -0.040271968, - -0.044543337, - -0.10995435, - -0.024011672, - -0.08962033, - 0.020206504, - 0.030622963, - -0.021175418, - 0.046819735, - -0.08388905, - -0.04419095, - -0.041822553, - 0.031128531, - 0.010744972, - 0.06392119, - -0.0031621107, - -0.012324199, - 0.039583333, - 0.03872388, - 0.04003792, - 0.012126796, - 0.060538515, - -0.046224117, - 0.009284271, - -0.051235553, - -0.049639463, - -0.015559349, - -0.08584357, - 0.07390804, - -0.029281551, - -1.4552155e-08, - -0.060234137, - -0.05653537, - -0.003924483, - -0.030553697, - 0.033688337, - -0.051516354, - 0.011325061, - 0.14125879, - 0.0239569, - 0.01933575, - 0.066012196, - 0.030753234, - -0.10696803, - 0.0034088665, - 0.073148385, - 0.02414587, - 0.080867074, - -0.07877004, - -0.032145467, - 0.07524812, - 0.0542984, - 0.009829384, - -0.1270656, - 0.06314169, - 0.09003407, - -0.0016169662, - 0.058391552, - 0.059590362, - -0.0047688517, - 0.022996303, - 0.035714924, - -0.034012605, - 0.07277301, - 0.0797266, - 0.0912049, - 0.022215161, - 0.045965668, - 0.04404474, - -0.083592154, - -0.10004596, - 0.020836696, - 0.023092525, - -0.047950342, - 0.08443384, - 0.0771323, - 0.009310225, - -0.080956854, - 0.09289323, - -0.020150434, - -0.00083508895, - -0.038630493, - 0.01606296, - 0.007031474, - -0.01770303, - -0.0022343053, - -0.021911092, - 0.03337036, - -0.032134622, - -0.012314019, - -0.0021285508, - 0.021125747, - 0.016543584, - 0.01756058, - -0.0771557 + -0.0039811856, + 0.051417787, + -0.0005715141, + -0.03803642, + 0.00027128452, + -0.07164157, + -0.03291995, + 0.029136118, + 0.08946447, + 0.027045254, + 0.02299213, + 0.029471328, + 0.0134518, + 0.02187473, + 0.02469151, + 0.02320869, + -0.06788812, + 0.042177398, + -0.06690857, + -0.044831563, + -0.019496728, + -0.017653666, + -0.04702048, + 0.010083961, + -0.03517946, + 0.12794615, + -0.039921533, + -0.037032884, + 0.021825306, + -0.06664962, + 0.020511147, + 0.03142631, + 0.12171163, + 0.03787702, + -0.07563969, + 0.035978485, + 0.11102433, + -0.025687309, + -0.077919014, + 0.01697996, + -0.08082786, + 0.04252522, + 0.008239131, + 0.07315216, + 0.01113036, + 0.016228424, + 0.019439362, + -0.05727214, + -0.02693478, + 0.027580295, + -0.103669524, + 0.06182539, + -0.028054044, + 0.0455263, + 0.038502, + 0.10223323, + 0.010206205, + 0.0038840948, + -0.074412525, + -0.009773545, + -0.014611934, + 0.0054425327, + -0.04896206, + 0.024686582, + 0.08042032, + -0.001383198, + 0.0008921756, + 0.0016657019, + 0.016278693, + 0.0036041273, + 0.058009814, + -0.01006157, + -0.008400023, + 0.06813469, + 0.033983845, + -0.011650561, + -0.04933034, + -0.034598112, + 0.02250701, + 0.016103996, + 0.025309183, + 0.03500135, + -0.018220501, + 0.06806079, + 0.0599987, + -0.02538144, + 0.04583358, + -0.042976376, + -0.10481837, + -0.02811274, + 0.0790001, + -0.01711748, + 0.0124125, + 0.04062369, + -0.020124251, + 0.026943244, + 0.041834604, + -0.04414842, + 0.08096912, + 0.02178193, + 0.081736214, + 0.03328645, + 0.021757673, + 0.09206164, + -0.05208259, + -0.1362023, + 0.013583276, + -0.019828305, + -0.036186047, + -0.05029691, + -0.0329588, + 0.04664088, + -0.062425017, + -0.0568364, + -0.027655112, + -0.1512015, + -0.09397598, + -0.011011233, + -0.02424874, + -0.046778478, + -0.0029182346, + -0.066472694, + -0.02560864, + 0.018279487, + 0.0020318548, + -0.062247586, + -0.1175178, + -4.4292554e-33, + -0.009136622, + -0.037174374, + -0.02604135, + 0.05202509, + 0.00085812004, + 0.006577624, + -0.00454986, + -0.040179957, + 0.0041096318, + 0.042830415, + -0.049696118, + 0.045353506, + 0.0428641, + 0.04491573, + 0.11103378, + 0.021595275, + -0.031244695, + 0.07227112, + -0.019163573, + -0.034743402, + 0.067720324, + -0.016135676, + 0.05970054, + -0.022954391, + 0.028844452, + 0.015425059, + -0.0003045761, + -0.012766137, + -0.033286437, + -0.00013060206, + -0.024786701, + -0.042230703, + -0.0024439848, + 0.041244082, + 0.09192393, + 0.06853773, + -0.015283744, + -0.12651399, + 0.017043643, + -0.086192474, + 0.05503255, + 0.030275606, + 0.0043217717, + 0.032015957, + -0.0322539, + 0.004916596, + 0.009010684, + -0.023150368, + -0.04071162, + -0.09106201, + 0.036371823, + 0.024427105, + 0.013570613, + 0.032411, + 0.04100558, + 0.03751363, + -0.041537277, + -0.078983925, + -0.053763084, + 0.064508595, + -0.080228396, + 0.02129479, + 0.0622995, + 0.045755383, + 0.03245333, + 0.089313425, + -0.040570408, + -0.03184067, + 0.097861506, + 0.018426524, + 0.005526579, + 0.03339057, + -0.069327235, + 0.0049843024, + -0.0111007085, + 0.04039599, + -0.018666456, + -0.061631028, + -0.01991712, + 0.055738986, + -0.03395009, + -0.032834787, + 0.03982438, + 0.032516815, + -0.014805643, + -0.040833063, + 0.09042099, + -0.0711825, + -0.045280255, + 0.0044581695, + -0.011292024, + 0.010472374, + -0.0051121535, + -0.032291505, + -0.0145741515, + 1.9738093e-33, + -0.014771771, + -0.011359673, + -0.018966977, + -0.03000902, + -0.032400236, + 0.00019931208, + -0.012589514, + -0.12148443, + 0.0020975752, + 0.03191172, + -0.004727992, + 0.009443682, + 0.07090955, + -0.100593194, + 0.025055854, + 0.06191594, + -0.0040894244, + -0.099229194, + -0.011800106, + -0.047672532, + -0.030504433, + 0.062661596, + -0.07383581, + -0.0061629685, + -0.014213279, + 0.0073416564, + -0.123823114, + -0.12355839, + 0.049825996, + 0.013651053, + -0.042307552, + -0.05771415, + 0.008855104, + -0.039362777, + -0.0103809675, + 0.01996238, + 0.06862416, + -0.0034638718, + 0.034936216, + 0.016902631, + -0.041256435, + 0.12757617, + -0.010952788, + -0.03840964, + 0.033540674, + 0.024278237, + -0.009185038, + 0.08946187, + -0.03729937, + -0.033711437, + 0.08362949, + 0.024390021, + 0.013046886, + -0.082465455, + 0.08171803, + 0.025880741, + -0.0407528, + 0.011633795, + 0.045149475, + 0.05799517, + -0.04314864, + -0.021866212, + 0.0076306188, + 0.07500617, + -0.03708559, + -0.040266518, + -0.044550084, + -0.10994247, + -0.0240191, + -0.089640714, + 0.020261075, + 0.030591883, + -0.02121829, + 0.046824012, + -0.08388275, + -0.04419642, + -0.041806895, + 0.03112276, + 0.010755989, + 0.06391166, + -0.0031454791, + -0.012299338, + 0.039625768, + 0.038725432, + 0.040054668, + 0.012143256, + 0.060516205, + -0.046224497, + 0.009262628, + -0.05123796, + -0.049621977, + -0.015547329, + -0.08584545, + 0.07391313, + -0.029286016, + -1.455054e-08, + -0.060226165, + -0.05653187, + -0.0039235186, + -0.030587185, + 0.03372064, + -0.05150516, + 0.011288491, + 0.14127062, + 0.023947941, + 0.01934539, + 0.06599671, + 0.030757299, + -0.10697992, + 0.0034128474, + 0.07314907, + 0.02416513, + 0.08087709, + -0.078763716, + -0.032150872, + 0.075260274, + 0.054309774, + 0.009813545, + -0.12706842, + 0.063152395, + 0.090030335, + -0.0016136293, + 0.05839544, + 0.05956212, + -0.0047300495, + 0.022987204, + 0.035680998, + -0.034018096, + 0.07276527, + 0.079732984, + 0.09122537, + 0.022239434, + 0.045964334, + 0.04402577, + -0.083599865, + -0.10004525, + 0.02080335, + 0.023068275, + -0.047920816, + 0.08444264, + 0.077142455, + 0.009301439, + -0.080961, + 0.09288208, + -0.020169992, + -0.0008043465, + -0.038628474, + 0.016085649, + 0.007008231, + -0.017678784, + -0.0022326205, + -0.021909356, + 0.033364836, + -0.03214099, + -0.012320089, + -0.0021274248, + 0.021111455, + 0.016534865, + 0.017552063, + -0.077128485 ] ] } diff --git a/tests/integration/recordings/responses/a59d0d7c1485.json b/tests/integration/recordings/responses/a59d0d7c1485.json index 2e896d01c..b4b76eee6 100644 --- a/tests/integration/recordings/responses/a59d0d7c1485.json +++ b/tests/integration/recordings/responses/a59d0d7c1485.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:27.609269Z", + "created_at": "2025-07-31T17:46:25.912484084Z", "done": true, "done_reason": "stop", - "total_duration": 326238958, - "load_duration": 79782250, + "total_duration": 3119768806, + "load_duration": 41781837, "prompt_eval_count": 238, - "prompt_eval_duration": 233571958, + "prompt_eval_duration": 3036042450, "eval_count": 2, - "eval_duration": 12258959, + "eval_duration": 41395778, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/a6810c23eda8.json b/tests/integration/recordings/responses/a6810c23eda8.json index 4d3b935da..a2c6fe9f4 100644 --- a/tests/integration/recordings/responses/a6810c23eda8.json +++ b/tests/integration/recordings/responses/a6810c23eda8.json @@ -23,7 +23,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.599113Z", + "created_at": "2025-07-31T17:53:05.617714834Z", "done": false, "done_reason": null, "total_duration": null, @@ -41,7 +41,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.643599Z", + "created_at": "2025-07-31T17:53:05.798542309Z", "done": false, "done_reason": null, "total_duration": null, @@ -59,7 +59,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.685747Z", + "created_at": "2025-07-31T17:53:05.980841562Z", "done": false, "done_reason": null, "total_duration": null, @@ -77,7 +77,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.727604Z", + "created_at": "2025-07-31T17:53:06.162303548Z", "done": false, "done_reason": null, "total_duration": null, @@ -95,7 +95,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.768014Z", + "created_at": "2025-07-31T17:53:06.340857636Z", "done": false, "done_reason": null, "total_duration": null, @@ -113,7 +113,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.809356Z", + "created_at": "2025-07-31T17:53:06.521595814Z", "done": false, "done_reason": null, "total_duration": null, @@ -131,7 +131,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.850402Z", + "created_at": "2025-07-31T17:53:06.701546566Z", "done": false, "done_reason": null, "total_duration": null, @@ -149,7 +149,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.891768Z", + "created_at": "2025-07-31T17:53:06.880153315Z", "done": false, "done_reason": null, "total_duration": null, @@ -167,7 +167,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.933421Z", + "created_at": "2025-07-31T17:53:07.060654891Z", "done": false, "done_reason": null, "total_duration": null, @@ -185,7 +185,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:24.976048Z", + "created_at": "2025-07-31T17:53:07.240380253Z", "done": false, "done_reason": null, "total_duration": null, @@ -203,7 +203,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.016922Z", + "created_at": "2025-07-31T17:53:07.419126616Z", "done": false, "done_reason": null, "total_duration": null, @@ -221,7 +221,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.058091Z", + "created_at": "2025-07-31T17:53:07.598790332Z", "done": false, "done_reason": null, "total_duration": null, @@ -239,7 +239,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.098992Z", + "created_at": "2025-07-31T17:53:07.777056009Z", "done": false, "done_reason": null, "total_duration": null, @@ -257,7 +257,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.140605Z", + "created_at": "2025-07-31T17:53:07.952924017Z", "done": false, "done_reason": null, "total_duration": null, @@ -275,7 +275,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.18202Z", + "created_at": "2025-07-31T17:53:08.131888967Z", "done": false, "done_reason": null, "total_duration": null, @@ -293,7 +293,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.223443Z", + "created_at": "2025-07-31T17:53:08.333057743Z", "done": false, "done_reason": null, "total_duration": null, @@ -311,7 +311,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.264829Z", + "created_at": "2025-07-31T17:53:08.5127717Z", "done": false, "done_reason": null, "total_duration": null, @@ -329,7 +329,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.306517Z", + "created_at": "2025-07-31T17:53:08.697392665Z", "done": false, "done_reason": null, "total_duration": null, @@ -347,7 +347,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.347967Z", + "created_at": "2025-07-31T17:53:08.880798014Z", "done": false, "done_reason": null, "total_duration": null, @@ -365,7 +365,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.389339Z", + "created_at": "2025-07-31T17:53:09.059630952Z", "done": false, "done_reason": null, "total_duration": null, @@ -383,7 +383,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.430357Z", + "created_at": "2025-07-31T17:53:09.239778855Z", "done": false, "done_reason": null, "total_duration": null, @@ -401,7 +401,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.471506Z", + "created_at": "2025-07-31T17:53:09.41796922Z", "done": false, "done_reason": null, "total_duration": null, @@ -419,7 +419,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.512744Z", + "created_at": "2025-07-31T17:53:09.594384168Z", "done": false, "done_reason": null, "total_duration": null, @@ -437,7 +437,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.55402Z", + "created_at": "2025-07-31T17:53:09.775287995Z", "done": false, "done_reason": null, "total_duration": null, @@ -455,7 +455,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.595747Z", + "created_at": "2025-07-31T17:53:09.956262386Z", "done": false, "done_reason": null, "total_duration": null, @@ -473,7 +473,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.637436Z", + "created_at": "2025-07-31T17:53:10.136372272Z", "done": false, "done_reason": null, "total_duration": null, @@ -491,7 +491,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.678551Z", + "created_at": "2025-07-31T17:53:10.31476923Z", "done": false, "done_reason": null, "total_duration": null, @@ -509,7 +509,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.719904Z", + "created_at": "2025-07-31T17:53:10.492983094Z", "done": false, "done_reason": null, "total_duration": null, @@ -527,7 +527,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.76118Z", + "created_at": "2025-07-31T17:53:10.673852463Z", "done": false, "done_reason": null, "total_duration": null, @@ -545,7 +545,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.802641Z", + "created_at": "2025-07-31T17:53:10.858527191Z", "done": false, "done_reason": null, "total_duration": null, @@ -563,7 +563,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.843247Z", + "created_at": "2025-07-31T17:53:11.038328378Z", "done": false, "done_reason": null, "total_duration": null, @@ -581,7 +581,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.88468Z", + "created_at": "2025-07-31T17:53:11.215788879Z", "done": false, "done_reason": null, "total_duration": null, @@ -599,7 +599,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.92653Z", + "created_at": "2025-07-31T17:53:11.399602557Z", "done": false, "done_reason": null, "total_duration": null, @@ -617,7 +617,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:25.968022Z", + "created_at": "2025-07-31T17:53:11.580123511Z", "done": false, "done_reason": null, "total_duration": null, @@ -635,7 +635,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.00935Z", + "created_at": "2025-07-31T17:53:11.757284813Z", "done": false, "done_reason": null, "total_duration": null, @@ -653,7 +653,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.050576Z", + "created_at": "2025-07-31T17:53:11.941379029Z", "done": false, "done_reason": null, "total_duration": null, @@ -671,7 +671,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.091784Z", + "created_at": "2025-07-31T17:53:12.127019036Z", "done": false, "done_reason": null, "total_duration": null, @@ -689,7 +689,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.133496Z", + "created_at": "2025-07-31T17:53:12.308766694Z", "done": false, "done_reason": null, "total_duration": null, @@ -707,7 +707,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.175442Z", + "created_at": "2025-07-31T17:53:12.497171833Z", "done": false, "done_reason": null, "total_duration": null, @@ -725,7 +725,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.217044Z", + "created_at": "2025-07-31T17:53:12.675256496Z", "done": false, "done_reason": null, "total_duration": null, @@ -743,7 +743,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.258582Z", + "created_at": "2025-07-31T17:53:12.86123582Z", "done": false, "done_reason": null, "total_duration": null, @@ -761,7 +761,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.300334Z", + "created_at": "2025-07-31T17:53:13.048911842Z", "done": false, "done_reason": null, "total_duration": null, @@ -779,15 +779,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:26.341814Z", + "created_at": "2025-07-31T17:53:13.236138052Z", "done": true, "done_reason": "stop", - "total_duration": 1862375416, - "load_duration": 73039291, + "total_duration": 7845771051, + "load_duration": 47957096, "prompt_eval_count": 18, - "prompt_eval_duration": 45477667, + "prompt_eval_duration": 178030745, "eval_count": 43, - "eval_duration": 1743432792, + "eval_duration": 7619284509, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/a97477559b10.json b/tests/integration/recordings/responses/a97477559b10.json index 60896815d..487f99d19 100644 --- a/tests/integration/recordings/responses/a97477559b10.json +++ b/tests/integration/recordings/responses/a97477559b10.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 23929167, - "load_duration": 17216625, + "total_duration": 14968443, + "load_duration": 6969545, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - 0.042460807, - -0.06189971, - -0.0784711, - 0.0064329687, - 0.03129365, - 0.00807445, - 0.05801836, - 0.025447326, - 0.016402787, - 0.045995634, - -0.028924342, - 0.04451832, - 0.05686613, - -0.015340794, - -0.07020505, - -0.057178136, - -0.07683263, - 0.006748679, - 0.0043323045, - -0.123651944, - 0.0031534543, - -0.03258051, - -0.02936216, - 0.024140852, - -0.028559243, - 0.10224467, - 0.0021632623, - -0.006975691, - 0.025292527, - -0.055500276, - 0.031231727, - -0.0070274337, - 0.08430815, - -0.028431177, - -0.083029, - 0.009555893, - -0.020029299, - -0.00243229, - -0.00768719, - -0.023077851, - -0.09293533, - -0.042625993, - -0.020000124, - 0.008240663, - 0.060970567, - 0.050315727, - -0.0510085, - -0.008543903, - -0.030227834, - -0.03582846, - -0.17836656, - -0.047279052, - 0.033892106, - 0.031623542, - -0.008832113, - 0.10480918, - 0.033559043, - 0.090348184, - -0.015757555, - -0.0125672715, - -0.084686965, - -0.114781834, - -0.13755985, - 0.021652374, - 0.047834594, - 0.043243896, - 0.008659893, - 0.038724966, - 0.046716973, - -0.077413626, - -0.04887495, - 0.031287406, - 0.022356613, - 0.00043283988, - 0.052321073, - -0.012254071, - -0.035172574, - -0.00825216, - -0.008866574, - -0.034267236, - -0.04576201, - 0.002467568, - -0.040877618, - 0.08047682, - 0.09472728, - 0.0413438, - 0.0057974122, - 0.044982508, - 0.025369909, - 0.006618073, - 0.010467276, - -0.07960384, - -0.03108485, - -0.03528749, - 0.01831391, - 0.053473305, - 0.06568304, - -0.07259002, - 0.02523736, - 0.10520362, - 0.035732146, - 0.028157586, - 0.011687256, - 0.044207197, - 0.012604437, - 0.0018819098, - 0.03926183, - 0.043135095, - 0.09784739, - -0.08801336, - -0.06060836, - 0.02681984, - 0.0041358666, - 0.033492945, - 0.011799116, - 0.009551661, - -0.0095491735, - -0.021212189, - -0.008917248, - 0.029352615, - -0.012693442, - -0.019269384, - 0.009901157, - -0.00812101, - 0.018603146, - -0.0007501193, - -0.056115113, - -3.8018077e-33, - 0.020848714, - 0.0047160466, - 0.019726405, - 0.06024251, - -0.0685974, - -0.07497267, - 0.007997452, - -0.047339544, - 0.057801835, - 0.049544968, - 0.01878086, - 0.03274472, - 0.017663997, - 0.07483022, - 0.02496901, - -0.011843339, - -0.11212756, - 0.0070379525, - 0.028099466, - -0.01746246, - 0.08173482, - -0.007920462, - 0.032095373, - -0.12300146, - 0.033773854, - 0.025873141, - -0.0045020077, - 0.079493225, - 0.0040725255, - 0.03305898, - 0.008061117, - 0.0134422695, - -0.03292251, - 0.031554114, - 0.04013794, - 0.0014983519, - 0.030762345, - 0.029481992, - 0.041350223, - -0.047438618, - 0.03944708, - -0.07526981, - 0.037927423, - -0.026016014, - 0.016933467, - 0.0136799775, - 0.0071263947, - -0.05386736, - -0.07443268, - -0.006070775, - 0.024427462, - -0.039844982, - -0.020661902, - -0.033354662, - 0.009005565, - 0.12111172, - -0.028260944, - -0.036192853, - -0.021332363, - 0.05333571, - 0.05161245, - -0.01204843, - 0.035563566, - 0.05408247, - 0.060722187, - 0.07159865, - 0.04299143, - 0.008544481, - 0.07421879, - 0.00841512, - -0.036342908, - -0.008549791, - -0.08816386, - -0.049075164, - 0.00029373015, - -0.05127952, - 0.03586739, - -0.030380003, - -0.012642127, - 0.018771531, - 0.01711824, - -0.06644723, - 0.023793438, - 0.0010271219, - -0.01939443, - -0.053452212, - -0.017060323, - -0.062207118, - -0.05962535, - -0.012172617, - -0.013190802, - -0.037036054, - 0.00082622556, - 0.098088354, - 0.024690514, - 2.1767905e-33, - -0.010088812, - -0.016811697, - -0.042140447, - 0.08837209, - -0.028899776, - -0.0048947735, - -0.082139015, - 0.029238816, - -0.043079354, - -0.014153092, - -0.028387645, - 0.025998218, - -0.017625, - 0.046511114, - -0.005768211, - 0.030010609, - 0.011375536, - 0.017426634, - 0.055062976, - 0.032230247, - -0.07995765, - 0.032486655, - -0.060016844, - -0.011561194, - 0.010211269, - 0.046528235, - 0.001191399, - 0.0786961, - -0.0446158, - 0.032789085, - 0.0023115936, - -0.03886269, - -0.017663589, - 0.07913024, - -0.004583343, - 0.043521065, - -0.031589273, - 0.008867868, - -0.05013296, - 0.068929516, - 0.043675046, - 0.019968731, - -0.08471742, - -0.046864275, - -0.0068198936, - -0.026138468, - -0.05107216, - 0.054374695, - 0.03069186, - -0.010925094, - 0.04721093, - -0.017387696, - -0.020754937, - -0.081763394, - -0.027709637, - 0.035980806, - 0.05396534, - 0.044874854, - 0.059699643, - 0.041227758, - -0.06664364, - -0.09201654, - 0.008915574, - 0.025849758, - -0.038651932, - -0.0044070315, - -0.052066546, - 0.027435115, - 0.012089562, - 0.048306923, - 0.059854515, - 0.097325735, - -0.053612895, - -0.07639326, - 0.015773866, - -0.0444848, - -0.13214406, - -0.0702488, - -0.10134438, - -0.11905995, - -0.027714504, - 0.006891868, - -0.0053650527, - 0.054135524, - -0.111159205, - 0.07835098, - 0.03506018, - 0.016036613, - 0.021490784, - -0.061526407, - 0.007425222, - 0.04833579, - -0.01361202, - 0.012450488, - -0.12729599, - -1.4009424e-08, - -0.040908325, - -0.01596458, - 0.060048707, - 0.03804525, - 0.0663794, - 0.04727275, - -0.016112225, - 0.09687414, - -0.04424251, - -0.028799534, - -0.01294642, - 0.013026413, - 0.022404836, - 0.04713173, - 0.06402557, - 0.12130648, - 0.06062839, - 0.10218965, - -0.0757528, - -0.023806982, - 0.12489501, - -0.045460615, - 0.09545599, - 0.021262301, - 0.03731495, - -0.075220875, - -0.0026194793, - 0.0472452, - 0.048499025, - 0.12358729, - 0.017998053, - 0.013811017, - -0.035893846, - -0.051789004, - 0.06182457, - 0.05160056, - 0.008895317, - -0.12500942, - 0.016453298, - -0.08590811, - -0.071096726, - 0.06987216, - -0.036072273, - -0.0053715096, - -0.048762616, - 0.00081640907, - -0.021502526, - -0.061078615, - 0.002485032, - -0.032720752, - 0.045743283, - 0.038934175, - -0.024666062, - 0.025897244, - 0.10301431, - -0.013001504, - 0.04783332, - -0.07114252, - 0.046031926, - 0.080549754, - -0.10302451, - 0.08449227, - 0.028010191, - -0.03697792 + 0.04246934, + -0.0618944, + -0.07845866, + 0.006401396, + 0.03128947, + 0.008068856, + 0.058044188, + 0.025462082, + 0.016408337, + 0.04599356, + -0.028957082, + 0.044486698, + 0.05686555, + -0.015339846, + -0.07019087, + -0.057180382, + -0.07685297, + 0.0067407293, + 0.0043030772, + -0.12366419, + 0.0031679699, + -0.03256299, + -0.029364487, + 0.024145795, + -0.028551238, + 0.10222551, + 0.0021702712, + -0.006973446, + 0.02528161, + -0.055480048, + 0.031220673, + -0.0070450194, + 0.08431401, + -0.028398452, + -0.08303615, + 0.009541477, + -0.020037198, + -0.0024062425, + -0.0076945405, + -0.02308465, + -0.092934616, + -0.04263402, + -0.020023424, + 0.008251729, + 0.060961593, + 0.050315246, + -0.051015448, + -0.008537156, + -0.030209582, + -0.035804313, + -0.17838922, + -0.047276836, + 0.033887845, + 0.03163823, + -0.0088280905, + 0.10479794, + 0.033603504, + 0.09033551, + -0.015729368, + -0.01254892, + -0.08468991, + -0.11477985, + -0.13756058, + 0.021674957, + 0.04781672, + 0.043237038, + 0.008655867, + 0.03871613, + 0.046686046, + -0.07744882, + -0.048885334, + 0.031265616, + 0.022349542, + 0.00042022156, + 0.05231528, + -0.012255999, + -0.035176765, + -0.008271301, + -0.008876192, + -0.034259606, + -0.04576233, + 0.0024646304, + -0.04088308, + 0.080499224, + 0.094756246, + 0.041366395, + 0.0058133435, + 0.04500413, + 0.025369063, + 0.006633623, + 0.010419986, + -0.07959981, + -0.031104978, + -0.03529105, + 0.018313587, + 0.053470284, + 0.06567532, + -0.072590366, + 0.025236402, + 0.10520805, + 0.03572949, + 0.02814476, + 0.011692066, + 0.04416959, + 0.012628916, + 0.0019061499, + 0.03925087, + 0.043107945, + 0.09784928, + -0.08801884, + -0.06063319, + 0.026810031, + 0.004124005, + 0.033488054, + 0.011797618, + 0.009536534, + -0.009537672, + -0.02117513, + -0.008926071, + 0.029353488, + -0.012708475, + -0.019247372, + 0.009906199, + -0.008120285, + 0.018623708, + -0.0007470326, + -0.056117814, + -3.8023727e-33, + 0.020859266, + 0.004704134, + 0.019713905, + 0.06026147, + -0.068616085, + -0.07496656, + 0.007965563, + -0.047342513, + 0.057801917, + 0.049556054, + 0.018799074, + 0.0327276, + 0.017686183, + 0.07481851, + 0.024969097, + -0.011860301, + -0.112135485, + 0.0070053833, + 0.028068872, + -0.017445812, + 0.08173, + -0.007909387, + 0.032085713, + -0.1230031, + 0.03378858, + 0.02584868, + -0.0044788863, + 0.07950368, + 0.0040581026, + 0.033058096, + 0.008016927, + 0.013449756, + -0.03286121, + 0.031516194, + 0.04012885, + 0.001483041, + 0.030742006, + 0.029510139, + 0.0413387, + -0.04742168, + 0.039420005, + -0.07528787, + 0.03795314, + -0.026037993, + 0.016913416, + 0.013665059, + 0.007120363, + -0.053870693, + -0.07446003, + -0.006065503, + 0.024451725, + -0.039831672, + -0.020631628, + -0.033301804, + 0.009001951, + 0.121108405, + -0.028260462, + -0.036188155, + -0.02130734, + 0.053347252, + 0.05162655, + -0.012022209, + 0.03556421, + 0.054093517, + 0.06071427, + 0.07156984, + 0.04303917, + 0.008578926, + 0.074212484, + 0.008422386, + -0.03633554, + -0.008555927, + -0.08816405, + -0.04909204, + 0.0002670324, + -0.05126362, + 0.035850402, + -0.030379485, + -0.012650656, + 0.018785939, + 0.017106218, + -0.06643792, + 0.023832165, + 0.0010230087, + -0.019367155, + -0.053461894, + -0.017073216, + -0.062169686, + -0.059630387, + -0.012175827, + -0.013204632, + -0.037058014, + 0.0008544243, + 0.098082356, + 0.024703579, + 2.1766385e-33, + -0.010099368, + -0.0168182, + -0.042137686, + 0.08836087, + -0.028925458, + -0.004915718, + -0.08213097, + 0.029251399, + -0.04311282, + -0.014153079, + -0.028403684, + 0.025994804, + -0.017619897, + 0.04652408, + -0.0057673934, + 0.030015819, + 0.011373319, + 0.017426124, + 0.055063453, + 0.032209422, + -0.079960845, + 0.03246185, + -0.06001134, + -0.011566254, + 0.01018926, + 0.046531614, + 0.0012229957, + 0.078700714, + -0.04463173, + 0.032767717, + 0.0022977523, + -0.03884795, + -0.017658522, + 0.07914583, + -0.0045653232, + 0.043508887, + -0.03161254, + 0.008865502, + -0.05014004, + 0.068952896, + 0.043693513, + 0.019953411, + -0.08469415, + -0.046883006, + -0.006852297, + -0.026142156, + -0.05107832, + 0.054367345, + 0.030704534, + -0.010923592, + 0.047204282, + -0.0173818, + -0.020731952, + -0.081755, + -0.027707007, + 0.035971012, + 0.053959154, + 0.044867415, + 0.059691377, + 0.041240178, + -0.06663067, + -0.09200673, + 0.00895469, + 0.02584983, + -0.038639534, + -0.0043883775, + -0.052067816, + 0.027416172, + 0.01205728, + 0.048341535, + 0.05988727, + 0.09733144, + -0.053652134, + -0.07638588, + 0.015775586, + -0.04448138, + -0.13214125, + -0.070245154, + -0.10134073, + -0.11906563, + -0.02768927, + 0.006883601, + -0.005372457, + 0.0541394, + -0.11116945, + 0.0783573, + 0.035053268, + 0.016029678, + 0.021500655, + -0.06151511, + 0.007436359, + 0.04835127, + -0.013626688, + 0.012427166, + -0.1272983, + -1.4008406e-08, + -0.040906746, + -0.015945766, + 0.06004726, + 0.03804035, + 0.06638126, + 0.047250524, + -0.016075904, + 0.09685192, + -0.04424075, + -0.028788049, + -0.012935697, + 0.012998786, + 0.022392461, + 0.047142737, + 0.0640327, + 0.12130623, + 0.060611896, + 0.10219882, + -0.075741336, + -0.023816003, + 0.12488407, + -0.04545764, + 0.095451295, + 0.021275673, + 0.037320722, + -0.0752353, + -0.002659371, + 0.047237474, + 0.048469283, + 0.123593695, + 0.018008605, + 0.013809098, + -0.035921507, + -0.05181155, + 0.06183518, + 0.051568378, + 0.008893763, + -0.1250224, + 0.01645567, + -0.085920095, + -0.07108048, + 0.0698563, + -0.036050897, + -0.005379485, + -0.04876276, + 0.00080607267, + -0.021491177, + -0.06106662, + 0.0024876762, + -0.03273843, + 0.04576169, + 0.03894835, + -0.024660163, + 0.025903016, + 0.10298729, + -0.013008437, + 0.04784895, + -0.071141616, + 0.046046503, + 0.08054465, + -0.10302495, + 0.08449142, + 0.02803045, + -0.036991965 ] ] } diff --git a/tests/integration/recordings/responses/ae1c22f18ecc.json b/tests/integration/recordings/responses/ae1c22f18ecc.json index b5b1604cc..595b6668d 100644 --- a/tests/integration/recordings/responses/ae1c22f18ecc.json +++ b/tests/integration/recordings/responses/ae1c22f18ecc.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:47.06444Z", + "created_at": "2025-07-31T17:59:32.661124541Z", "done": true, "done_reason": "stop", - "total_duration": 2620252041, - "load_duration": 196706333, + "total_duration": 11391290133, + "load_duration": 42154800, "prompt_eval_count": 20, - "prompt_eval_duration": 70100292, + "prompt_eval_duration": 1208581216, "eval_count": 58, - "eval_duration": 2352865167, + "eval_duration": 10140044676, "response": "I'm happy to help you with your test, but I don't see what kind of test we are testing. Could you please provide more context or clarify what kind of test you would like me to perform? Is it a programming test, a language proficiency test, or something else?", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/ae6835cfe70e.json b/tests/integration/recordings/responses/ae6835cfe70e.json index 82664b0b6..b6b80e2a9 100644 --- a/tests/integration/recordings/responses/ae6835cfe70e.json +++ b/tests/integration/recordings/responses/ae6835cfe70e.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:48:00.342705Z", + "created_at": "2025-07-31T17:57:28.280495733Z", "done": true, "done_reason": "stop", - "total_duration": 671224833, - "load_duration": 82344875, + "total_duration": 32855142429, + "load_duration": 40184388, "prompt_eval_count": 386, - "prompt_eval_duration": 545215084, + "prompt_eval_duration": 32634520132, "eval_count": 2, - "eval_duration": 43112416, + "eval_duration": 179879941, "response": "[]", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/afb33182f365.json b/tests/integration/recordings/responses/afb33182f365.json index 1c51c5a7f..0e1b0bfc0 100644 --- a/tests/integration/recordings/responses/afb33182f365.json +++ b/tests/integration/recordings/responses/afb33182f365.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -20,14 +20,14 @@ "body": { "__type__": "openai.types.chat.chat_completion.ChatCompletion", "__data__": { - "id": "chatcmpl-541", + "id": "chatcmpl-456", "choices": [ { "finish_reason": "stop", "index": 0, "logprobs": null, "message": { - "content": "Saturn is the planet that has rings around itself.", + "content": "Saturn is the planet known to have magnificent ring systems.", "refusal": null, "role": "assistant", "annotations": null, @@ -37,15 +37,15 @@ } } ], - "created": 1753814884, + "created": 1753984690, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion", "service_tier": null, "system_fingerprint": "fp_ollama", "usage": { - "completion_tokens": 12, + "completion_tokens": 13, "prompt_tokens": 39, - "total_tokens": 51, + "total_tokens": 52, "completion_tokens_details": null, "prompt_tokens_details": null } diff --git a/tests/integration/recordings/responses/b14ff438ca99.json b/tests/integration/recordings/responses/b14ff438ca99.json index d4d24a048..c445e7d42 100644 --- a/tests/integration/recordings/responses/b14ff438ca99.json +++ b/tests/integration/recordings/responses/b14ff438ca99.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:31.274826Z", + "created_at": "2025-07-31T17:51:39.104140157Z", "done": true, "done_reason": "stop", - "total_duration": 5223804916, - "load_duration": 61370666, + "total_duration": 22895811031, + "load_duration": 41692686, "prompt_eval_count": 23, - "prompt_eval_duration": 70195875, + "prompt_eval_duration": 793961939, "eval_count": 124, - "eval_duration": 5091701417, + "eval_duration": 22059637137, "response": "The official currency of Japan is the Japanese yen (\u00a5). It is abbreviated as \"JPY\" and its symbol is \u00a5. The yen is divided into 100 sen, although the sen has been officially discontinued since 1967.\n\nYou can exchange your money for yen at banks, currency exchange offices, or use ATMs to withdraw cash from an ATM. Credit cards are also widely accepted in Japan, especially among major retailers and restaurants.\n\nIt's worth noting that some businesses may not accept foreign currencies other than US dollars, so it's a good idea to have some local currency on hand when traveling to Japan.", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/b24590574a85.json b/tests/integration/recordings/responses/b24590574a85.json index 615b5618d..1f1e43679 100644 --- a/tests/integration/recordings/responses/b24590574a85.json +++ b/tests/integration/recordings/responses/b24590574a85.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -21,7 +21,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { @@ -36,7 +36,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984664, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -47,7 +47,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { @@ -62,7 +62,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984664, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -73,11 +73,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { - "content": " How", + "content": " It", "function_call": null, "refusal": null, "role": "assistant", @@ -88,7 +88,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984664, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -99,11 +99,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { - "content": " can", + "content": "'s", "function_call": null, "refusal": null, "role": "assistant", @@ -114,7 +114,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984665, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -125,11 +125,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { - "content": " I", + "content": " nice", "function_call": null, "refusal": null, "role": "assistant", @@ -140,7 +140,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984665, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -151,11 +151,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { - "content": " help", + "content": " to", "function_call": null, "refusal": null, "role": "assistant", @@ -166,7 +166,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984665, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -177,7 +177,33 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " meet", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984665, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", "choices": [ { "delta": { @@ -192,7 +218,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984665, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -203,11 +229,11 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { - "content": " today", + "content": ".", "function_call": null, "refusal": null, "role": "assistant", @@ -218,7 +244,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984666, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -229,7 +255,371 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " Is", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984666, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " there", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984666, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " something", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984666, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " I", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984666, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " can", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984666, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " help", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984667, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " you", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984667, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " with", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984667, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " or", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984667, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " would", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984667, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " you", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984668, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " like", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984668, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " to", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984668, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", + "choices": [ + { + "delta": { + "content": " chat", + "function_call": null, + "refusal": null, + "role": "assistant", + "tool_calls": null + }, + "finish_reason": null, + "index": 0, + "logprobs": null + } + ], + "created": 1753984668, + "model": "llama3.2:3b-instruct-fp16", + "object": "chat.completion.chunk", + "service_tier": null, + "system_fingerprint": "fp_ollama", + "usage": null + } + }, + { + "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", + "__data__": { + "id": "chatcmpl-968", "choices": [ { "delta": { @@ -244,7 +634,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984668, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -255,7 +645,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-528", + "id": "chatcmpl-968", "choices": [ { "delta": { @@ -270,7 +660,7 @@ "logprobs": null } ], - "created": 1753814882, + "created": 1753984669, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/b44cc7a7afc8.json b/tests/integration/recordings/responses/b44cc7a7afc8.json index 70b765861..dc770b693 100644 --- a/tests/integration/recordings/responses/b44cc7a7afc8.json +++ b/tests/integration/recordings/responses/b44cc7a7afc8.json @@ -23,1556 +23,1556 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 66326542, - "load_duration": 22228125, + "total_duration": 91561401, + "load_duration": 7565647, "prompt_eval_count": 162, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.07448108, - 0.027982691, - -0.025962545, - 0.028414156, - -0.04874927, - -0.124489374, - -0.03775365, - 0.041172747, - -0.048783444, - -0.027774421, - -0.09272271, - 0.051921174, - 0.08087506, - 0.023085767, - 0.103185095, - -0.06142812, - -0.046623003, - 0.031264473, - -0.009095788, - -0.110987656, - -0.020735977, - 0.036462996, - -0.013348663, - 0.007442654, - 0.019446686, - 0.0043880027, - -0.0123794135, - -0.04474342, - -0.00010696763, - 0.027796188, - -0.05249273, - 0.062042117, - 0.019623421, - 0.022298045, - -0.01876838, - 0.06636658, - -0.036940884, - -0.09439301, - -0.04989112, - -0.016055813, - -0.08934105, - 0.07278765, - -0.073312856, - -0.027571253, - -0.06639977, - 0.015506035, - -0.004176694, - -0.032542672, - -0.035769954, - -0.026245229, - -0.09129098, - 0.022831371, - -0.05601971, - -0.103505865, - -0.023430603, - -0.01617043, - 0.060298156, - -0.011999374, - -0.00982143, - -0.15203232, - -0.07311755, - 0.022391053, - 0.08800625, - 0.062195398, - -0.04764835, - -0.05545306, - -0.036078423, - 0.017782934, - 0.08492913, - -0.050706394, - -0.09958507, - -0.029495796, - -0.002121337, - 0.08148674, - 0.030521393, - -0.12159759, - 0.04639748, - 0.0054555144, - -0.0076237656, - 0.04930283, - 0.001018987, - 0.01823945, - -0.056388717, - 0.09080432, - 0.03544767, - -0.062846325, - 0.05177355, - 0.07175976, - -0.045391884, - 0.009686718, - 0.030302709, - -0.058896482, - 0.03719664, - 0.004174063, - -0.014313601, - 0.06214871, - 0.026443055, - -0.054081496, - -0.04056011, - 0.010876058, - -0.0033277434, - -0.07736001, - 0.055489365, - 0.011366925, - 0.049955327, - 0.011093621, - 0.044155005, - -0.08873286, - 0.04789806, - -0.029256178, - -0.021238709, - -0.059048988, - -0.006010105, - -0.036286995, - 0.045776833, - 0.07393597, - -0.0043319017, - 0.07591234, - -0.0006300352, - 0.0063326987, - 0.019833053, - -0.008920521, - -0.0074224886, - -0.014964156, - 0.012450781, - 0.003317517, - -0.009942644, - 1.525195e-33, - -0.030182399, - -0.056817565, - -0.009954876, - 0.02231213, - 0.057156544, - -0.018560076, - 0.07843683, - -0.003509288, - -0.031122614, - -0.0333474, - 0.019342642, - 0.03716782, - 0.030942772, - 0.13801146, - -0.0026788223, - 0.0060844175, - 0.024037478, - 0.028806396, - 0.0114514725, - 0.0028755309, - 0.009741409, - -0.010365574, - 0.025636459, - 0.04402703, - 0.00824972, - -0.023288164, - -0.025415357, - -0.02247272, - 0.016395057, - 0.0039686435, - -0.06683203, - -0.058984432, - -0.026139224, - 0.02571613, - -0.023981044, - -0.01542635, - -0.013025425, - -0.08132036, - 0.029904919, - -0.0048653325, - -0.02163821, - 0.025880665, - 0.004492511, - -0.013551861, - -0.014834658, - 0.046109095, - -0.00031146017, - 0.016851023, - -0.12182429, - 0.021024965, - -0.009434213, - -0.03510208, - 0.080137864, - 0.08463277, - 0.0019426581, - 0.051176246, - 0.05314091, - 0.032667853, - -0.041880205, - -0.05545038, - 0.014655727, - 0.034564327, - 0.09517278, - 0.0048721586, - 0.038064517, - 0.064016655, - 0.036886543, - 0.11732628, - 0.04750395, - 0.062849574, - -0.043793496, - 0.039535545, - -0.0414883, - 0.045276705, - -0.005626682, - 0.028326502, - 0.03510831, - -0.11158364, - 0.067508236, - 0.025473768, - -0.016454473, - -0.023138152, - 0.02560681, - -0.03489655, - -0.0143142305, - -0.043763783, - -0.006103266, - 0.044694975, - -0.007177529, - -0.038755096, - -0.06350946, - -0.05295245, - 0.044151388, - 0.024555689, - -0.01345332, - -5.1627547e-33, - -0.011461753, - -0.003969141, - -0.04658726, - 0.0008026091, - -0.090269305, - -0.0629358, - 0.009687034, - 0.00015354449, - 0.043152034, - 0.022057066, - -0.049155302, - -0.08511033, - 0.110782035, - 0.017681966, - 0.056186423, - 0.03724774, - -0.114085265, - 0.011197734, - 0.010572792, - 0.03503156, - -0.07397689, - 0.0156148635, - -0.032688703, - -0.06490581, - -0.010675779, - -0.041401856, - -0.097037986, - -0.07025277, - 0.021750104, - 0.05030694, - -0.017832309, - 0.032031614, - -0.03788665, - 0.03141082, - 0.07613352, - -0.0007763451, - 0.034961626, - -0.06256205, - -0.006801991, - -0.026741587, - 0.11656076, - 0.05023973, - 0.06515106, - 0.06511257, - 0.025219081, - 0.03180813, - -0.05966658, - 0.08190675, - -0.028054262, - -0.048548922, - -0.03486897, - 0.03020514, - 0.035033725, - -0.018610824, - -0.038684692, - -0.048875436, - 0.021133669, - 0.08319505, - -0.06746284, - -0.053462982, - -0.08098418, - -0.06340421, - 0.011191566, - 0.020785637, - -0.06575731, - 0.02211741, - -0.10775702, - -0.011597437, - -0.051947355, - -0.1501959, - 0.11516611, - -0.030521782, - -0.018723903, - 0.052845538, - -0.06679985, - 0.040416736, - -0.028146135, - -0.01644884, - -0.025731068, - 0.06570538, - 0.0866128, - 0.010937938, - -0.03865133, - 0.027389226, - -0.06712724, - -0.015267271, - -0.05265448, - 0.020899015, - 0.031420153, - 0.002802588, - 0.010436373, - 0.048363067, - 0.021981295, - 0.01690293, - -0.022728851, - -4.0744272e-08, - -0.0065167644, - 0.0014059767, - 0.05391456, - 0.015178632, - 0.018086514, - 0.08112959, - 0.005525823, - -0.037069544, - -0.01871401, - 0.051793523, - -0.014797383, - -0.044994324, - -0.09279006, - -0.07259356, - -0.004214306, - 0.14136177, - -0.022566888, - -0.030480398, - 0.047431417, - 0.06623071, - 0.07947818, - -0.023033215, - -0.05389834, - 0.10418305, - -0.08498801, - -0.032223985, - 0.058419, - 0.0036608635, - -0.02912376, - -0.09348434, - -0.004131768, - -0.035598896, - 0.007222825, - 0.040373847, - 0.04553802, - 0.018402338, - 0.021517321, - -0.06000489, - -0.028075347, - 0.018188315, - -0.021463133, - -0.003939297, - 0.012185079, - -0.016664179, - 0.021595497, - 0.02443412, - -0.044382285, - -0.047587246, - -0.057701204, - -0.057771184, - -0.0060019926, - -0.0099875815, - -0.016420204, - -0.049889106, - 0.020464808, - 0.076619074, - -0.13720629, - 0.00883673, - -0.032044746, - 0.035911836, - -0.006365476, - 0.11197782, - 0.15684035, - -0.00079191517 + -0.074518315, + 0.027976887, + -0.02593837, + 0.028388586, + -0.04876724, + -0.124505915, + -0.03774251, + 0.041209254, + -0.048753787, + -0.027770191, + -0.092701025, + 0.051893745, + 0.080871776, + 0.023072483, + 0.103199854, + -0.061396316, + -0.046632618, + 0.031259038, + -0.009099142, + -0.110972114, + -0.020746639, + 0.036444016, + -0.013380681, + 0.007444351, + 0.01946104, + 0.0044101896, + -0.012363551, + -0.044742297, + -0.000109245026, + 0.027794957, + -0.05248249, + 0.062051035, + 0.019644303, + 0.02233988, + -0.018772174, + 0.06638057, + -0.03694357, + -0.09439974, + -0.0498964, + -0.016085815, + -0.08936965, + 0.07279298, + -0.073282845, + -0.027557392, + -0.0663739, + 0.015465914, + -0.004212829, + -0.03255469, + -0.035763785, + -0.026250161, + -0.09131691, + 0.022791812, + -0.0560323, + -0.103517555, + -0.02337786, + -0.016169889, + 0.06033412, + -0.012003445, + -0.009792255, + -0.1520526, + -0.07313599, + 0.022364737, + 0.08799007, + 0.062190924, + -0.0476343, + -0.055481553, + -0.036080837, + 0.01777673, + 0.084963925, + -0.050720915, + -0.09959311, + -0.029466882, + -0.0020879637, + 0.08149215, + 0.030555075, + -0.12159375, + 0.04638196, + 0.0054617906, + -0.007600725, + 0.04925988, + 0.0010117136, + 0.01823397, + -0.056374155, + 0.0908255, + 0.03542638, + -0.06282811, + 0.05174182, + 0.07176561, + -0.04539055, + 0.009702367, + 0.03031262, + -0.05891284, + 0.037203796, + 0.0041589597, + -0.014310235, + 0.062144686, + 0.026470749, + -0.054097973, + -0.040584724, + 0.010875258, + -0.003349861, + -0.077305436, + 0.055475105, + 0.011331311, + 0.049933966, + 0.011079793, + 0.04419192, + -0.088725075, + 0.04790246, + -0.029256914, + -0.021242525, + -0.059049267, + -0.0059888517, + -0.036285046, + 0.045748435, + 0.07392407, + -0.0042937263, + 0.07591468, + -0.00059363164, + 0.006329638, + 0.019841122, + -0.008888848, + -0.0074318657, + -0.014973693, + 0.012456981, + 0.0033115426, + -0.009930274, + 1.5269222e-33, + -0.03017857, + -0.056806926, + -0.009980652, + 0.022316255, + 0.057149988, + -0.01857269, + 0.0784146, + -0.003503646, + -0.031156048, + -0.033383664, + 0.01937351, + 0.037160154, + 0.030936565, + 0.1380185, + -0.002635351, + 0.0060887556, + 0.02401934, + 0.028815405, + 0.011467783, + 0.0028821875, + 0.009709699, + -0.010394833, + 0.025645163, + 0.044017233, + 0.008246027, + -0.023243824, + -0.025415665, + -0.022480464, + 0.016407011, + 0.0039797607, + -0.06682885, + -0.058947742, + -0.026093839, + 0.025729727, + -0.023972526, + -0.015408932, + -0.013048789, + -0.08130767, + 0.029917423, + -0.004828957, + -0.021634426, + 0.02587896, + 0.0044811512, + -0.013536556, + -0.014813144, + 0.046061084, + -0.00032990836, + 0.016869118, + -0.12181025, + 0.021046987, + -0.009420413, + -0.035062335, + 0.08011807, + 0.08462047, + 0.0019942592, + 0.05117461, + 0.05312125, + 0.0326721, + -0.04189356, + -0.055460256, + 0.01466244, + 0.03459353, + 0.095160365, + 0.0048437407, + 0.038064692, + 0.06401175, + 0.036864925, + 0.11731751, + 0.04750967, + 0.06286565, + -0.04375349, + 0.039547894, + -0.041467424, + 0.04528996, + -0.005650938, + 0.028340634, + 0.03510358, + -0.111572064, + 0.06749655, + 0.025440717, + -0.016470913, + -0.023179049, + 0.0256079, + -0.03489901, + -0.01430054, + -0.043748833, + -0.0060837497, + 0.044692438, + -0.0072164233, + -0.038763802, + -0.063516915, + -0.052999448, + 0.04417511, + 0.024537848, + -0.013432413, + -5.162713e-33, + -0.0114407325, + -0.003955193, + -0.04661703, + 0.0007875603, + -0.09029818, + -0.062948115, + 0.009710563, + 0.0001300855, + 0.04312154, + 0.022073459, + -0.04914153, + -0.08508943, + 0.11078909, + 0.017684652, + 0.056212854, + 0.03725169, + -0.114067726, + 0.011182615, + 0.0105617605, + 0.035042927, + -0.07395952, + 0.015640577, + -0.032705046, + -0.06488826, + -0.010690244, + -0.041422527, + -0.09704262, + -0.070222415, + 0.021709241, + 0.05029499, + -0.017807636, + 0.032029808, + -0.03788697, + 0.03136548, + 0.076138325, + -0.0007745447, + 0.034956176, + -0.06253955, + -0.006809682, + -0.026719663, + 0.11657212, + 0.050194807, + 0.06518189, + 0.06511228, + 0.025210718, + 0.03180309, + -0.059656575, + 0.08190252, + -0.028029623, + -0.04854541, + -0.034874525, + 0.030208217, + 0.035034187, + -0.018606044, + -0.038674414, + -0.048887372, + 0.021132758, + 0.08317319, + -0.0675027, + -0.05348525, + -0.080962874, + -0.06341586, + 0.011199907, + 0.0207855, + -0.06572508, + 0.022130286, + -0.10779961, + -0.011599286, + -0.051970255, + -0.15018743, + 0.11517539, + -0.03052435, + -0.0187536, + 0.052858524, + -0.06682251, + 0.04038606, + -0.028126227, + -0.016444748, + -0.02575468, + 0.06569969, + 0.08660793, + 0.010974402, + -0.0386448, + 0.027382996, + -0.06711591, + -0.0152584985, + -0.052659295, + 0.020921137, + 0.031379428, + 0.002811196, + 0.010419629, + 0.048391167, + 0.02201258, + 0.016886525, + -0.022732206, + -4.073636e-08, + -0.006513384, + 0.0014004525, + 0.053950045, + 0.015167113, + 0.018088367, + 0.08111558, + 0.0055300333, + -0.03707988, + -0.018737856, + 0.051793147, + -0.014813838, + -0.044947825, + -0.09278965, + -0.07260186, + -0.0041794567, + 0.14137252, + -0.022569552, + -0.030528586, + 0.047419064, + 0.066193394, + 0.07945365, + -0.023014557, + -0.053888306, + 0.104186185, + -0.08501846, + -0.03223301, + 0.05844058, + 0.0036631415, + -0.02911171, + -0.09349268, + -0.004123487, + -0.035597004, + 0.007244818, + 0.04035152, + 0.045556862, + 0.01838623, + 0.02155509, + -0.060004886, + -0.028096678, + 0.018191703, + -0.021443348, + -0.003914473, + 0.012175833, + -0.01663914, + 0.021617427, + 0.024437096, + -0.04434746, + -0.04760396, + -0.057683956, + -0.057734974, + -0.006014961, + -0.009957316, + -0.016418923, + -0.049850997, + 0.02046306, + 0.07664182, + -0.13724001, + 0.008844773, + -0.032054316, + 0.035961926, + -0.0063517457, + 0.11198241, + 0.1568678, + -0.0007999774 ], [ - -0.0012923438, - 0.013419649, - 0.03603258, - 0.046982195, - -0.008386184, - -0.012245008, - 0.017257063, - -0.014495833, - -0.06755615, - 0.013220825, - -0.071046636, - 0.022029007, - 0.04805814, - -0.06659013, - -0.030023778, - 0.014715108, - 0.04294596, - 0.031195298, - -0.06522679, - -0.07396746, - 0.017329818, - -0.0151756415, - -0.052758723, - 0.06344977, - 0.005364444, - 0.02631366, - 0.03665044, - 0.048812985, - -0.0044375616, - 0.0103826355, - -0.0089511005, - -0.07216287, - 0.05088121, - 0.017377803, - -0.061182447, - -0.010244597, - -0.06587784, - 0.069840916, - 0.028359821, - -0.037131228, - -0.052071016, - -0.07370394, - 0.0233667, - -0.02532014, - 0.06171828, - 0.11584273, - -0.08307468, - -0.08872316, - -0.04554565, - 0.02177065, - -0.12324151, - -0.023568366, - -0.0015541487, - -0.013532973, - -0.056209136, - 0.0880576, - 0.03321554, - 0.05171784, - 0.0074756956, - -0.025275769, - 0.023162214, - -0.15517598, - -0.010777206, - 0.016303454, - 0.034188252, - 0.020134093, - -0.022240352, - 0.050957076, - -0.005396301, - -0.04007687, - -0.020301744, - 0.10113998, - 0.002977471, - 0.06617704, - 0.040134214, - -0.02005319, - -0.059682623, - -0.06369068, - 0.08473604, - 0.023557685, - -0.017191878, - -0.005820709, - -0.026404407, - 0.09280466, - 0.04844145, - -0.06875489, - -0.022161635, - -0.015402431, - -0.0111024445, - -0.017707076, - 0.025355583, - -0.039296508, - -0.001362202, - -0.040884525, - -0.03204941, - 0.04150212, - 0.008948646, - -0.13776794, - 0.030302526, - 0.058231197, - 0.010572606, - 0.09247389, - -0.035872795, - -0.0036602807, - 0.056347203, - -0.003996722, - 0.035537403, - 0.014696888, - 0.10615937, - -0.13590123, - -0.05810754, - 0.04527657, - -0.06982519, - -0.049982276, - -0.041045085, - 0.01247287, - -0.040934183, - 0.028955987, - -0.02226216, - 0.08722953, - -0.009548719, - -0.025511682, - 0.0114325285, - 0.03363939, - 0.021809513, - -0.08675585, - -0.07089411, - 1.7909231e-33, - -0.04121751, - -0.1001688, - 0.006345352, - 0.0037210584, - 0.029166285, - -0.0872215, - -0.04271259, - -0.06566409, - 0.017946582, - 0.022238955, - -0.03249184, - -0.02349789, - 0.021466883, - 0.09511927, - 0.08346572, - 0.042806614, - 0.0038908664, - 0.037915263, - 0.020043708, - -0.033399176, - 0.10208849, - -0.014397545, - 0.021684645, - -0.021582458, - -0.0074115414, - 0.046073515, - 0.06664795, - 0.06434497, - -0.010910654, - 0.016172478, - 0.030913299, - 0.017434347, - -0.0762684, - 0.027927354, - 0.053165767, - -0.061656844, - 0.007082498, - 0.0057526245, - 0.055203717, - 0.069314696, - -0.027693065, - -0.045786254, - 0.094618365, - -0.02984729, - -0.045069296, - 0.01723317, - 0.016129777, - -0.06281533, - -0.045081936, - -0.045089465, - -0.0053253355, - -0.019320533, - -0.045810748, - -0.02639149, - 0.012412514, - 0.08566385, - -0.0034776065, - 0.0035142878, - -0.012017715, - 0.006649936, - 0.033606175, - -0.0012646043, - 0.042252455, - 0.055928096, - 0.017948387, - 0.07064788, - 0.10451079, - 0.062350754, - 0.04458121, - -0.0028225682, - 0.02566386, - -0.0021405003, - 0.040477417, - -0.012259745, - 0.052335545, - -0.0017080541, - 0.05346329, - -0.007733562, - -0.028276777, - 0.018282998, - -0.046343774, - -0.043290336, - -0.026471136, - -0.11104024, - 0.008576623, - 0.005548108, - -0.034847535, - -0.056416124, - -0.030293388, - 0.0053394907, - -0.09004081, - -0.03141982, - -0.062330373, - 0.09981983, - -0.032840475, - -3.3540373e-33, - -0.027300175, - 0.010525057, - -0.021980286, - 0.12664026, - 0.031588834, - 0.033247624, - -0.05148502, - -0.03101089, - -0.0465964, - -0.0022529345, - -0.056195565, - 0.007953736, - 0.064945616, - 0.03884713, - -0.06837888, - 0.077476665, - -0.06788635, - 0.0064428714, - -0.040736765, - 0.037416343, - -0.07232494, - 0.063321635, - 0.014398016, - -0.05871896, - 0.031005096, - -0.019561818, - -0.07452502, - 0.037396118, - -0.026255993, - 0.020780139, - -0.031075457, - 0.0058948854, - -0.047562398, - -0.010866235, - 0.0352409, - 0.0549852, - 0.07012556, - -0.056673322, - -0.017415406, - 0.07528239, - 0.05387259, - 0.0028653517, - -0.07284915, - -0.07543174, - -0.012900278, - 0.011457189, - -0.08563738, - -0.0015463261, - 0.036361244, - -0.062004283, - -0.0050084046, - 0.023846988, - -0.008083734, - -0.03593437, - -0.034260865, - 0.000298229, - -0.0578704, - 0.021156322, - 0.056237947, - 0.102285825, - -0.07694436, - -0.096381366, - 0.029115336, - 0.001019501, - -0.010235284, - 0.055199094, - -0.021333022, - 0.04801045, - -0.008948923, - 0.0043332377, - 0.002985581, - 0.049172573, - -0.049805593, - 0.07117998, - -0.04823976, - -0.072981454, - -0.026498413, - -0.06437876, - -0.0346269, - -0.0060303714, - 0.018713593, - -0.07784192, - -0.0046854415, - 0.04578587, - -0.043880597, - 0.012154217, - 0.024205454, - 0.0352363, - 0.0063410155, - -0.086736806, - -0.014489626, - 0.048670504, - -0.06944819, - 0.047556538, - -0.096405424, - -3.8881783e-08, - 0.020024363, - -0.0060733794, - 0.10675529, - -0.0072445725, - 0.11130468, - 0.0766799, - -0.089739904, - 0.10989663, - -0.060538583, - -0.061066266, - 0.046883732, - -0.016365182, - 0.016547771, - -0.012390388, - 0.0035057077, - 0.031388927, - 0.018324051, - 0.038030062, - -0.0005554988, - 0.019816065, - 0.110884875, - -0.023082083, - 0.049298774, - -0.049228016, - 0.03771876, - -0.10209589, - 0.021328293, - 0.0048561115, - -0.026669646, - 0.04161308, - -0.037887473, - 0.029118432, - 0.03738528, - -0.015714107, - 0.0959638, - 0.1434109, - 0.049922757, - -0.11274395, - -0.06264596, - -0.038560014, - -0.03071335, - 0.08555022, - -0.048136428, - 0.0401538, - 0.014374478, - -0.021280114, - 0.04872567, - -0.057720494, - 0.009963986, - 0.002822142, - 0.079809405, - 0.017903175, - 0.022365756, - 0.08987974, - 0.06651197, - 0.022014199, - 0.059419304, - -0.06117766, - 0.015350715, - 0.08376493, - -0.0017018274, - 0.08864588, - -0.027652979, - -0.060420066 + -0.0012985186, + 0.013428601, + 0.036027383, + 0.046960995, + -0.008376715, + -0.012273062, + 0.017215235, + -0.014517273, + -0.06755925, + 0.013262504, + -0.071071416, + 0.022008605, + 0.04802556, + -0.06656689, + -0.030001678, + 0.014703167, + 0.04293022, + 0.031151697, + -0.06519839, + -0.07397044, + 0.017323893, + -0.015189615, + -0.052739624, + 0.06344194, + 0.005378495, + 0.026332699, + 0.036680676, + 0.048806872, + -0.0044219326, + 0.010361781, + -0.008937124, + -0.07216964, + 0.050818473, + 0.017360602, + -0.061186902, + -0.010224321, + -0.06590306, + 0.06985154, + 0.028388679, + -0.037106816, + -0.052078426, + -0.07370584, + 0.023386989, + -0.025320385, + 0.06171919, + 0.11583571, + -0.08312255, + -0.08873915, + -0.04554808, + 0.021797463, + -0.12322211, + -0.02355109, + -0.0015547865, + -0.013524721, + -0.056223456, + 0.08805911, + 0.0332561, + 0.05172255, + 0.007508375, + -0.025260713, + 0.023153193, + -0.15516914, + -0.01075054, + 0.016285403, + 0.03417789, + 0.02007978, + -0.022245353, + 0.0509647, + -0.0054105176, + -0.040100772, + -0.020286275, + 0.10114523, + 0.0030004813, + 0.06618223, + 0.040104922, + -0.020045916, + -0.05968854, + -0.06369228, + 0.08476288, + 0.023561234, + -0.017190726, + -0.0057785655, + -0.02643019, + 0.09284292, + 0.048416004, + -0.068727545, + -0.02216159, + -0.015408143, + -0.011069366, + -0.017663702, + 0.025346316, + -0.03933665, + -0.0013904214, + -0.04090857, + -0.032088112, + 0.041472837, + 0.008925901, + -0.13771445, + 0.030238513, + 0.058210976, + 0.010595619, + 0.0924281, + -0.035886403, + -0.003660082, + 0.056327023, + -0.0040123863, + 0.035575725, + 0.014680677, + 0.10619057, + -0.13590562, + -0.05811401, + 0.04527551, + -0.06981517, + -0.049992837, + -0.041055493, + 0.012480766, + -0.04090579, + 0.02896762, + -0.022247234, + 0.087224506, + -0.009555419, + -0.025493871, + 0.0113851605, + 0.03364401, + 0.02181673, + -0.086783744, + -0.070866294, + 1.7912747e-33, + -0.04119901, + -0.100177474, + 0.006347325, + 0.0037340575, + 0.029203828, + -0.087258354, + -0.04274845, + -0.065680355, + 0.01794751, + 0.022274023, + -0.03245305, + -0.023515053, + 0.021447303, + 0.0950956, + 0.083449624, + 0.0428005, + 0.003910466, + 0.037924897, + 0.020041984, + -0.033424165, + 0.10205846, + -0.014394057, + 0.021688785, + -0.021577379, + -0.0074261655, + 0.04609739, + 0.06662811, + 0.06431144, + -0.010944364, + 0.016165929, + 0.030921511, + 0.017438315, + -0.07628473, + 0.027964544, + 0.05316952, + -0.06166001, + 0.00710056, + 0.0057538245, + 0.05521142, + 0.06931237, + -0.027706858, + -0.045808528, + 0.094666, + -0.02986965, + -0.04502887, + 0.017208695, + 0.016125973, + -0.0628507, + -0.045059443, + -0.045112878, + -0.005296992, + -0.019326933, + -0.045822155, + -0.02639405, + 0.01242909, + 0.08570191, + -0.003465873, + 0.003503288, + -0.012003436, + 0.006605807, + 0.03363934, + -0.001257058, + 0.04224235, + 0.055937544, + 0.017936032, + 0.07066278, + 0.1045465, + 0.062303454, + 0.044585444, + -0.002807214, + 0.02564102, + -0.002128406, + 0.040478833, + -0.01224923, + 0.052337434, + -0.0016797099, + 0.053469352, + -0.0077856537, + -0.028244767, + 0.018288352, + -0.046363432, + -0.04332065, + -0.026436778, + -0.11104876, + 0.008586205, + 0.0055207564, + -0.034841597, + -0.056425076, + -0.030253613, + 0.005325803, + -0.090041295, + -0.031432882, + -0.062356126, + 0.09982324, + -0.032827362, + -3.3549678e-33, + -0.027284035, + 0.010559345, + -0.021984268, + 0.12661384, + 0.0315912, + 0.033252638, + -0.051472977, + -0.030958762, + -0.04658957, + -0.0022805957, + -0.056222532, + 0.00796958, + 0.06494811, + 0.038894437, + -0.06838922, + 0.077499114, + -0.06790046, + 0.0064532245, + -0.040768467, + 0.037424307, + -0.072336495, + 0.06332956, + 0.014400053, + -0.05869224, + 0.031022472, + -0.019536898, + -0.07451289, + 0.03739678, + -0.02625108, + 0.02074715, + -0.031048505, + 0.0059261005, + -0.04759007, + -0.010896379, + 0.035239074, + 0.054979034, + 0.07011226, + -0.056623362, + -0.017411917, + 0.07528956, + 0.05387218, + 0.0028673257, + -0.07281712, + -0.07544035, + -0.012932695, + 0.011416252, + -0.08563262, + -0.0015282914, + 0.036346182, + -0.062029377, + -0.0050238175, + 0.02387278, + -0.008091779, + -0.035949487, + -0.034255754, + 0.0003292639, + -0.057821356, + 0.021184877, + 0.056231596, + 0.102305636, + -0.076927446, + -0.09633249, + 0.029132774, + 0.0010131018, + -0.010232655, + 0.055211753, + -0.021346482, + 0.048036017, + -0.008985098, + 0.0043310625, + 0.002983946, + 0.049164876, + -0.049816035, + 0.07115217, + -0.04826019, + -0.07298708, + -0.026493097, + -0.064357154, + -0.034591526, + -0.006029352, + 0.018753871, + -0.077848874, + -0.0046812696, + 0.04576945, + -0.043886483, + 0.012162078, + 0.02418125, + 0.035210256, + 0.0063425824, + -0.08672974, + -0.014485961, + 0.0486449, + -0.06944658, + 0.047546502, + -0.09639138, + -3.8882344e-08, + 0.020005174, + -0.0060803695, + 0.10673199, + -0.0072566518, + 0.11126952, + 0.07668037, + -0.0897575, + 0.109880716, + -0.060538035, + -0.061037064, + 0.046886686, + -0.016372517, + 0.01658076, + -0.012367154, + 0.0035005491, + 0.031382836, + 0.01833628, + 0.038030002, + -0.00055114034, + 0.019830866, + 0.11086577, + -0.02309543, + 0.04928018, + -0.049268693, + 0.037694186, + -0.10212397, + 0.021300899, + 0.004854364, + -0.026668059, + 0.04163984, + -0.037908267, + 0.029162008, + 0.03740134, + -0.015686596, + 0.09598688, + 0.14345205, + 0.04990253, + -0.11276881, + -0.062654205, + -0.038547758, + -0.030726157, + 0.08556472, + -0.048128515, + 0.04011241, + 0.014323266, + -0.021255655, + 0.048724912, + -0.057747725, + 0.009945408, + 0.0028096687, + 0.07980508, + 0.017901363, + 0.02239066, + 0.08985929, + 0.0665591, + 0.022021096, + 0.059401497, + -0.061183818, + 0.015351812, + 0.08374175, + -0.0016842537, + 0.08864498, + -0.027638372, + -0.06043769 ], [ - -0.019089537, - 0.08206227, - -0.031629756, - -0.037748322, - -0.013907723, - -0.15086435, - -0.054227855, - 0.013812081, - 0.022318492, - 0.025760967, - -0.018970305, - 0.0159997, - 0.046886247, - -0.008989786, - 0.042260803, - 0.01563633, - -0.08306234, - 0.018418225, - -0.016524842, - -0.033054315, - -0.021094276, - -0.04198475, - -0.108629815, - 0.019558346, - -0.021839257, - 0.14248955, - -0.0012803682, - -0.058087774, - 0.005395786, - -0.040014874, - 0.012412929, - -0.014448109, - 0.10412988, - 0.08678136, - -0.07392144, - 0.031378184, - 0.077501394, - -0.04197698, - -0.092644565, - 0.019878637, - -0.09584833, - 0.06355258, - 0.0034316017, - 0.03860985, - -0.022438047, - 0.04932071, - -0.026379092, - -0.049524873, - -0.013308545, - 0.012192514, - -0.11695286, - 0.04510036, - -0.029017858, - 0.025516428, - 0.04245081, - 0.070753604, - 0.07057494, - 0.003524953, - -0.06010962, - 0.041959174, - 0.016197778, - -0.07186037, - 0.014555853, - -0.006213116, - 0.030063417, - 0.047432736, - 0.011306432, - 0.013843393, - 0.0436187, - -0.021850524, - 0.022346757, - 0.047835413, - -0.04025223, - 0.09492459, - 0.03155159, - 0.013348888, - -0.039819352, - -0.021837216, - 0.028181475, - -0.03434981, - 0.019666592, - 0.043579087, - -0.042940862, - 0.054164745, - 0.02308801, - -0.056740467, - 0.016757911, - -0.02701336, - -0.039681926, - 0.022773864, - 0.074453875, - -0.01407503, - -0.008249863, - 0.008273288, - -0.024091411, - -0.020071099, - 0.024399305, - -0.025779521, - 0.1035294, - -0.016452465, - 0.05220051, - 0.043400586, - 0.024392875, - 0.0160118, - -0.050395392, - -0.11149879, - 0.05203916, - -0.017942373, - -0.03793447, - -0.06775703, - -0.01611577, - 0.05274979, - -0.08863033, - -0.085470706, - -0.076794446, - -0.09332248, - -0.1264284, - 0.013839316, - -0.030490262, - 0.009920159, - 0.03968685, - -0.01939706, - -0.028892461, - 0.008741198, - 0.017886965, - -0.117217556, - -0.1212998, - 1.35733635e-33, - -0.035622492, - -0.023267707, - -0.017018162, - 0.00010073695, - 0.007257954, - -0.029587401, - 0.022087794, - -0.010561547, - -0.06912062, - 0.04277785, - -0.034413584, - 0.041110493, - 0.017055655, - 0.038174715, - 0.13757399, - -0.008806284, - -0.0023235404, - 0.08372674, - -0.024748268, - -0.028528849, - 0.096861266, - -0.02111509, - 0.06039901, - -0.041284908, - 0.07366366, - 0.018533891, - -0.019621244, - 0.00789655, - -0.012412154, - -0.005184189, - -0.0202234, - -0.011487718, - 0.0026882978, - 0.036282968, - 0.12384692, - 0.029563135, - 0.02673901, - -0.06578298, - 0.02610267, - -0.062275145, - 0.036926493, - 0.030272253, - 0.034105044, - 0.03516919, - -0.06365454, - -0.016557874, - -0.020214476, - -0.007219471, - 0.004009068, - -0.07774858, - 0.06894675, - 0.012156706, - 0.024095584, - 0.07716194, - 0.027376112, - 0.03524163, - -0.046042208, - -0.061379924, - -0.026633548, - 0.08248479, - -0.06261388, - 0.009910456, - 0.034668844, - 0.023772387, - -0.005869554, - 0.02162769, - -0.026385942, - -0.02100117, - 0.11375441, - 0.03666832, - -0.008121711, - 0.0026215075, - -0.032531988, - 0.01391055, - -0.018540533, - -0.0059300573, - -0.012669122, - -0.04971856, - -0.048864197, - 0.027610987, - -0.08137648, - 0.012624587, - 0.045806322, - 0.01336533, - 0.002328637, - -0.050664812, - 0.041695803, - -0.015773693, - -0.07136885, - -0.016258836, - -0.018871423, - -0.0038626953, - 0.03402061, - -0.009335479, - 0.005747506, - -4.5611018e-33, - 0.023689948, - -0.02445775, - -0.00834689, - -0.00063168275, - -0.021578811, - 0.012567475, - -0.025760869, - -0.10368349, - -0.03997725, - 0.01210385, - -0.015231519, - 0.02017564, - 0.045654193, - -0.07050829, - 0.034459736, - 0.056491707, - -0.014989821, - -0.08433123, - -0.049400527, - -0.03832157, - -0.055948768, - 0.044390477, - -0.001941214, - -0.0763155, - 0.034730915, - -0.04243297, - -0.07322386, - -0.08912488, - 0.083965875, - 0.034240186, - -0.055734336, - -0.017151177, - -0.0023456868, - -0.019274496, - 0.03401833, - -0.006712739, - 0.070724845, - -0.013663151, - 0.035358265, - -0.011840785, - -0.011920096, - 0.081632204, - 0.011438198, - -0.04905726, - 0.04624871, - 0.029794158, - -0.035954632, - 0.1309978, - -0.0722, - -0.053626865, - 0.047662914, - -0.032893717, - 0.03320312, - -0.053293463, - 0.11909418, - -0.013308413, - -0.08026765, - 0.018056376, - 0.028816566, - 0.012597203, - -0.082487956, - -0.07992265, - 0.03653938, - 0.048042614, - -0.04597376, - -0.039927375, - -0.019282784, - -0.11115308, - -0.12229221, - -0.08222088, - 0.014523922, - 0.041549023, - -0.054067343, - 0.12032739, - -0.10513437, - -0.03352011, - -0.046141136, - 0.015660388, - 0.03162219, - 0.089564346, - 0.06229127, - 0.02344754, - 0.013432015, - 0.04364802, - 0.017062847, - 0.030911682, - 0.052861545, - -0.05597565, - 0.015810143, - -0.04374839, - -0.039106574, - -0.020592151, - -0.01868341, - 0.08352379, - -0.017375095, - -3.8713683e-08, - -0.052152414, - -0.09442023, - 0.009305927, - -0.024598995, - 0.04574071, - 0.0017779457, - -0.019384999, - 0.14307584, - -0.00092140987, - -0.018639628, - 0.06094085, - -0.022180414, - -0.06670714, - -0.042788457, - 0.07614433, - 0.052368972, - 0.08171796, - -0.13214965, - 0.015069824, - 0.07545052, - 0.016364794, - 0.0030805927, - -0.06188439, - 0.07879054, - 0.04179921, - -0.043787137, - 0.05729686, - 0.013950966, - -0.01580636, - 0.002741003, - -0.002896178, - -0.027976623, - 0.0352471, - 0.07360851, - 0.11537727, - 0.008016604, - 0.054790642, - 0.070841216, - -0.040544577, - -0.07585315, - 0.015317468, - -0.014144724, - -0.03884744, - 0.029432015, - 0.061295677, - 0.025552604, - -0.03950773, - 0.1131327, - -0.028318027, - 0.031907115, - -0.038748857, - 0.029967804, - -0.020923622, - -0.0045868345, - -0.060423743, - 0.01062511, - -0.006921613, - -0.046255972, - 0.04074385, - 0.039824147, - -0.016014125, - 0.025676023, - 0.03524506, - -0.0267346 + -0.019079557, + 0.0820648, + -0.031636775, + -0.037772615, + -0.013885996, + -0.1508895, + -0.054257914, + 0.01382107, + 0.022319643, + 0.025744708, + -0.019006949, + 0.01595819, + 0.046914633, + -0.00899574, + 0.042291548, + 0.015646506, + -0.08305796, + 0.018408896, + -0.016524782, + -0.033079498, + -0.02110188, + -0.0419632, + -0.10861823, + 0.019554872, + -0.021874238, + 0.14247465, + -0.0012422869, + -0.058081616, + 0.00540865, + -0.03999031, + 0.012399737, + -0.014456615, + 0.10413924, + 0.08677547, + -0.07393572, + 0.031389575, + 0.07748671, + -0.041946597, + -0.092635125, + 0.019878551, + -0.09585241, + 0.063563004, + 0.0034580587, + 0.038572513, + -0.022447942, + 0.049308285, + -0.02643344, + -0.049521465, + -0.013297457, + 0.012233744, + -0.11695251, + 0.045083124, + -0.029010503, + 0.025497276, + 0.042470127, + 0.0707831, + 0.07058064, + 0.0035199749, + -0.06013254, + 0.041935362, + 0.016181944, + -0.07186833, + 0.014542711, + -0.0062323804, + 0.030054, + 0.047468036, + 0.011281582, + 0.013848972, + 0.04363679, + -0.021843519, + 0.022379788, + 0.047847077, + -0.04025328, + 0.09494594, + 0.03154395, + 0.013367471, + -0.03980583, + -0.02183361, + 0.028191755, + -0.03431455, + 0.019671934, + 0.043623473, + -0.042967957, + 0.05416258, + 0.023089629, + -0.05675844, + 0.016767101, + -0.027033433, + -0.03967794, + 0.022832932, + 0.074487366, + -0.0140734995, + -0.008246596, + 0.008278476, + -0.024108624, + -0.020060774, + 0.024378806, + -0.025747048, + 0.103516266, + -0.016442155, + 0.05220777, + 0.043397434, + 0.02440455, + 0.015943957, + -0.050434876, + -0.11145781, + 0.052034505, + -0.017928654, + -0.037932526, + -0.06774673, + -0.016093384, + 0.052765142, + -0.088646345, + -0.085484, + -0.07681618, + -0.093297966, + -0.12641862, + 0.013837021, + -0.03048377, + 0.009924758, + 0.039679028, + -0.01936025, + -0.028867563, + 0.00871666, + 0.01787285, + -0.11724568, + -0.12129051, + 1.35681665e-33, + -0.035628006, + -0.02325887, + -0.017038958, + 6.923209e-05, + 0.0072679906, + -0.0295577, + 0.022121288, + -0.010553554, + -0.06914253, + 0.04274084, + -0.03442124, + 0.041121893, + 0.017030265, + 0.0381245, + 0.1375638, + -0.008848526, + -0.0022947441, + 0.08370864, + -0.024741588, + -0.028515331, + 0.096916184, + -0.02108659, + 0.060407557, + -0.04129938, + 0.07367577, + 0.01852983, + -0.019585919, + 0.00791101, + -0.012426415, + -0.0051718187, + -0.02018194, + -0.011494365, + 0.0027114314, + 0.036264967, + 0.12386286, + 0.029567113, + 0.026756234, + -0.065749444, + 0.02609893, + -0.06232083, + 0.036904484, + 0.03028667, + 0.03411426, + 0.03521002, + -0.06369096, + -0.016598077, + -0.02021809, + -0.007230074, + 0.0040345713, + -0.07773345, + 0.06900628, + 0.012128798, + 0.02410663, + 0.0771743, + 0.027342282, + 0.03522959, + -0.046029396, + -0.061365336, + -0.026628872, + 0.08244359, + -0.062566556, + 0.009933027, + 0.034682497, + 0.023791147, + -0.005842399, + 0.021625068, + -0.026427383, + -0.020991165, + 0.11373874, + 0.03665437, + -0.008091131, + 0.0026228908, + -0.03253574, + 0.013892951, + -0.018594475, + -0.0059351088, + -0.012646403, + -0.04972099, + -0.048871726, + 0.027652413, + -0.08134938, + 0.0126620745, + 0.045843933, + 0.013398319, + 0.0023260224, + -0.05067545, + 0.04169543, + -0.01574087, + -0.07133913, + -0.016233964, + -0.018855713, + -0.0039056542, + 0.03401857, + -0.0093123065, + 0.0057734908, + -4.560601e-33, + 0.023695195, + -0.024489691, + -0.008312362, + -0.00066975394, + -0.02158263, + 0.0125598665, + -0.025738584, + -0.103652894, + -0.04000462, + 0.012098888, + -0.015197609, + 0.02018357, + 0.045623176, + -0.07047928, + 0.034468062, + 0.056500535, + -0.014972724, + -0.08429199, + -0.04942398, + -0.038302135, + -0.055943407, + 0.044392228, + -0.0019404019, + -0.07631783, + 0.034751914, + -0.0424522, + -0.07319884, + -0.08912471, + 0.08396021, + 0.034198415, + -0.055730376, + -0.017105753, + -0.0023682339, + -0.019267518, + 0.034007754, + -0.0067198407, + 0.07068643, + -0.013686713, + 0.03535481, + -0.011829574, + -0.011924876, + 0.08163265, + 0.011458664, + -0.049093027, + 0.046278197, + 0.029842824, + -0.035928097, + 0.13096437, + -0.0722123, + -0.053622153, + 0.047652073, + -0.032896154, + 0.033168253, + -0.053275317, + 0.119145334, + -0.013329809, + -0.080296695, + 0.01806636, + 0.028828703, + 0.012575126, + -0.08250055, + -0.07993187, + 0.0365166, + 0.048019268, + -0.0459654, + -0.039913233, + -0.019308258, + -0.11114867, + -0.12229502, + -0.08222976, + 0.014503677, + 0.041564006, + -0.054101657, + 0.12031798, + -0.10518697, + -0.033531662, + -0.046120696, + 0.015669933, + 0.031650025, + 0.08953049, + 0.062307738, + 0.023478396, + 0.013392765, + 0.043648973, + 0.017074035, + 0.030888386, + 0.052875523, + -0.055972677, + 0.015790377, + -0.04368904, + -0.039097052, + -0.020597953, + -0.018675094, + 0.08349847, + -0.017391236, + -3.870914e-08, + -0.05217957, + -0.0943954, + 0.009313268, + -0.024596054, + 0.0457224, + 0.0017694158, + -0.0194238, + 0.14304265, + -0.00092139974, + -0.018642776, + 0.060916223, + -0.022210617, + -0.06669, + -0.042800087, + 0.076100215, + 0.05237621, + 0.08171605, + -0.13214897, + 0.015094836, + 0.075452864, + 0.01636198, + 0.0030703964, + -0.061852757, + 0.07880552, + 0.04179526, + -0.04381105, + 0.057303566, + 0.0139259575, + -0.015837422, + 0.0027170512, + -0.0029033618, + -0.02796994, + 0.035219938, + 0.07358342, + 0.115382664, + 0.008049736, + 0.054797564, + 0.070874535, + -0.04053772, + -0.07585998, + 0.015316053, + -0.014189948, + -0.038860295, + 0.029442793, + 0.061300512, + 0.025522308, + -0.039504033, + 0.11314281, + -0.028287454, + 0.031891253, + -0.038770907, + 0.029970054, + -0.020935897, + -0.004616352, + -0.06046541, + 0.010621891, + -0.0069159092, + -0.04626887, + 0.040723223, + 0.03980271, + -0.016016755, + 0.025667662, + 0.035244495, + -0.026702441 ], [ - -0.053171553, - -0.047855794, - 0.04959839, - -0.009352584, - -0.056259144, - -0.036997948, - 0.01525368, - 0.0033788579, - 0.04453428, - 0.016438372, - -0.065293424, - 0.04655176, - 0.012637792, - 0.025149647, - -0.11436081, - 0.027283441, - -0.052422393, - 0.060236752, - -0.046064522, - -0.022863738, - 0.016536511, - 0.014447978, - -0.07744467, - 0.016475804, - -0.067145765, - 0.120901324, - -0.0022643541, - -0.0005619333, - 0.03098974, - 0.03116176, - 0.10501578, - -0.06940328, - -0.013246061, - 0.029016647, - -0.08779694, - 0.055636257, - -0.09158273, - -0.018188708, - -0.024831342, - -0.020263424, - 0.013102336, - -0.0007477728, - 0.0018712403, - 0.0068353964, - 0.08601601, - 0.061896168, - -0.07733195, - -0.047134392, - -0.04994557, - -0.008955441, - -0.08808325, - 0.0011078792, - -0.015078675, - -0.007628681, - 0.08530312, - 0.059783977, - 0.024557464, - 0.037825108, - -0.05171798, - 0.03148071, - 0.11377193, - -0.04417297, - 0.009659848, - 0.0060449084, - 0.030134702, - 0.07118153, - -0.013864897, - 0.03624278, - 0.0049465275, - -0.07480586, - 0.09733932, - 0.071613275, - -0.009146446, - -0.009571701, - 0.042258315, - 0.011740325, - 0.032803785, - 0.018631615, - 0.012556345, - -0.009346388, - -0.03489368, - 0.01649207, - 0.005488214, - 0.03819102, - 0.09597803, - -0.002047146, - -0.020768773, - 0.018077927, - -0.032444023, - 0.012474241, - -0.014445184, - -0.0670006, - -0.095488854, - -0.10345397, - -0.0009862595, - -0.0030658073, - 0.027003448, - -0.033961065, - 0.0011482734, - -0.009025799, - -0.048620287, - 0.0029769312, - -0.04154341, - -0.0395945, - 0.07520094, - 0.031153427, - 0.030031031, - 0.03353441, - 0.11403943, - -0.082912125, - -0.109138384, - 0.030059446, - -0.041853014, - 0.042241115, - 0.033335667, - -0.038876496, - 0.02092849, - 0.028346559, - 0.054482125, - 0.09627962, - -0.0035115955, - -0.015083763, - -0.092599295, - -0.056257337, - -0.00332258, - -0.02934002, - -0.11417531, - 1.5075675e-33, - -0.04527847, - -0.07345357, - 0.034714583, - -0.067186035, - 0.023143126, - -0.05054431, - -0.017398916, - -0.0058387746, - 0.052131217, - -0.017985696, - -0.10168014, - 0.016505243, - -0.005961273, - 0.08834502, - 0.047341425, - -0.06262999, - -0.03724901, - -0.0490674, - 0.061806694, - -0.117662214, - 0.014966754, - -0.07085228, - 0.07317225, - -0.010064827, - -0.004601465, - 0.0014379362, - 0.0122654615, - -0.018565418, - 0.018996973, - -0.0076706754, - -0.0085447915, - 0.023833418, - -0.0074106916, - -0.04202295, - -0.008097604, - -0.0089935325, - 0.11068735, - -0.028457392, - 0.037548065, - 0.04710371, - 0.062597714, - -0.049594503, - 0.06267496, - 0.005339454, - 0.024064569, - 0.034303125, - -0.016984673, - -0.03375307, - 0.012577206, - -0.05741818, - -0.046267692, - -0.00036155691, - 0.02268587, - -0.109952465, - 0.09230675, - 0.048918508, - -0.044157643, - 0.05441931, - -0.0058244704, - 0.04833069, - 0.035635386, - -0.015495411, - -0.008146981, - 0.092891365, - 0.112310715, - 0.047900427, - -0.017513819, - -0.009520781, - 0.06212363, - -0.0040008924, - 0.00397841, - 0.09532846, - -0.05659656, - -0.058885954, - -0.013697212, - 0.009742546, - -0.04745855, - -0.061571207, - -0.085869245, - 0.05009574, - -0.027810305, - -0.007983068, - -0.06844095, - 0.032406274, - 0.015316275, - 0.0830624, - 0.063605405, - -0.005157704, - -0.011889667, - -0.05187598, - -0.0087124705, - -0.031850815, - 0.043204896, - 0.00032051498, - -0.0012597291, - -2.3328516e-33, - -0.08486178, - 0.023463517, - -0.05558325, - 0.028823433, - 0.0598007, - 0.044241305, - -0.06976774, - -0.08749109, - -0.023545535, - 0.0767821, - 0.015185076, - 0.019631226, - -0.058358442, - 0.018799065, - 0.0076146126, - -0.015977694, - -0.057259887, - -0.042667117, - 0.101026215, - -0.03983678, - -0.03180352, - 0.03177619, - -0.057957705, - -0.036778692, - 0.027305948, - -0.0069477605, - -0.0753, - 0.049428534, - 0.012732314, - 0.10010171, - -0.036260307, - -0.048061043, - 0.029081684, - 0.01795974, - 0.045303203, - 0.102590606, - 0.005036657, - -0.05526093, - 0.008327211, - -0.05970527, - 0.020131486, - 0.009408121, - -0.06648779, - -0.029893365, - 0.0434368, - -0.0683305, - -0.07649664, - 0.039999247, - -0.06477932, - 0.07227491, - 0.046653986, - -0.016773192, - -0.048649658, - -0.08454509, - -0.05255037, - 0.0319589, - 0.024662357, - 0.023793997, - 0.076360136, - -0.040995322, - -0.033935655, - -0.11416756, - 0.06787201, - 0.009610846, - -0.064101316, - 0.024561828, - 0.024906442, - -0.0041048713, - 0.018717252, - -0.038110614, - 0.0145301875, - 0.068478055, - 0.018691448, - 0.05943308, - 0.023695862, - -0.009747667, - -0.066519946, - 0.0209059, - -0.019389415, - 0.014860701, - 0.022718104, - -0.022605024, - 0.0105253365, - 0.05693715, - 0.07257885, - 0.06504599, - -0.010055237, - 0.07908256, - 0.035240322, - -0.02378674, - 0.017134566, - 0.0878081, - 0.005987074, - 0.007431842, - -0.10935983, - -2.8794002e-08, - -0.05234688, - -0.08765063, - 0.06662866, - 0.013907749, - 0.0999487, - -0.022422735, - 0.06214868, - 0.027856557, - -0.06424995, - -0.038701627, - 0.025059296, - 0.00807731, - -0.024077412, - 0.011949065, - 0.08715261, - 0.012486595, - 0.06470489, - -0.027933354, - 0.039985545, - -0.012295149, - 0.02333007, - -0.03250732, - -0.04260915, - 0.10736886, - 0.037696708, - -0.06628188, - -0.056817852, - -0.005238912, - -0.069547325, - 0.100934796, - -0.033363372, - 0.021774344, - 0.017414633, - 0.018075803, - 0.026276791, - 0.066073745, - 0.059642654, - -0.065390244, - -0.115749314, - -0.07125786, - -0.023382567, - 0.042660285, - 0.043636538, - -0.03665277, - 0.050204884, - 0.0030947176, - 0.057122562, - -0.034636553, - 0.025459053, - -0.046185397, - -0.067215376, - 0.06057241, - -0.041255984, - -0.019857686, - -0.013778329, - -0.06125949, - 0.014752149, - -0.07630465, - -0.056748062, - 0.0505062, - -0.036068004, - 0.12241577, - 0.06429002, - -0.038303368 + -0.053175602, + -0.047849268, + 0.049600203, + -0.009332594, + -0.05626027, + -0.03703611, + 0.015297836, + 0.0033727393, + 0.044511985, + 0.016425023, + -0.06529153, + 0.046528336, + 0.012637323, + 0.025194079, + -0.1143288, + 0.027321098, + -0.052430134, + 0.060264964, + -0.046056643, + -0.022868538, + 0.016518874, + 0.014427887, + -0.077468514, + 0.01650613, + -0.067144066, + 0.120887764, + -0.0022775852, + -0.0005598929, + 0.031006373, + 0.031167403, + 0.10499404, + -0.069391765, + -0.01322822, + 0.028970728, + -0.08779589, + 0.05563035, + -0.091597155, + -0.018200668, + -0.024829883, + -0.020258859, + 0.0131373005, + -0.0007341065, + 0.0018953033, + 0.006834895, + 0.08603948, + 0.06189398, + -0.07733514, + -0.047121815, + -0.04994335, + -0.0089659095, + -0.0880838, + 0.0011172506, + -0.015044709, + -0.0075995945, + 0.085313074, + 0.059796136, + 0.02457739, + 0.0378336, + -0.051707182, + 0.031467274, + 0.113771856, + -0.044192057, + 0.0096846735, + 0.006033161, + 0.030144352, + 0.07118354, + -0.013839908, + 0.036214717, + 0.004951509, + -0.07481083, + 0.09734058, + 0.07162632, + -0.009135306, + -0.009563247, + 0.042295255, + 0.0117468545, + 0.03281954, + 0.018608347, + 0.012542441, + -0.009309551, + -0.034870803, + 0.016498035, + 0.0054994198, + 0.038178287, + 0.09602082, + -0.0020852594, + -0.020779438, + 0.01808113, + -0.03249026, + 0.012480446, + -0.014463354, + -0.06702938, + -0.09548575, + -0.103447035, + -0.0009932001, + -0.0030760013, + 0.026984407, + -0.033981565, + 0.0011538514, + -0.009027189, + -0.048636526, + 0.0029721952, + -0.041503906, + -0.03960792, + 0.07517321, + 0.031135045, + 0.030046917, + 0.033542294, + 0.11397492, + -0.082903914, + -0.109131016, + 0.03003371, + -0.041856304, + 0.04223555, + 0.033319004, + -0.03889455, + 0.020930232, + 0.02838724, + 0.0545114, + 0.09626628, + -0.0035141057, + -0.015085271, + -0.09259153, + -0.056270823, + -0.0033157181, + -0.029304419, + -0.114175975, + 1.50678135e-33, + -0.0453055, + -0.07348326, + 0.034691177, + -0.0672167, + 0.023145972, + -0.050515983, + -0.017413607, + -0.0058405283, + 0.052108254, + -0.017992783, + -0.10167575, + 0.016488168, + -0.0059505017, + 0.08831343, + 0.047385737, + -0.06261416, + -0.03727668, + -0.049049053, + 0.061813977, + -0.11765181, + 0.014997916, + -0.07084365, + 0.07316741, + -0.010097435, + -0.0045747026, + 0.0014380639, + 0.0123074865, + -0.018593263, + 0.019023519, + -0.0076754233, + -0.008543783, + 0.023825979, + -0.0074089407, + -0.042009465, + -0.008104463, + -0.008959146, + 0.11069426, + -0.028461525, + 0.0375111, + 0.047092855, + 0.062606744, + -0.049568158, + 0.06266772, + 0.0053055165, + 0.024054594, + 0.034305595, + -0.017003167, + -0.033732932, + 0.012580805, + -0.057429112, + -0.046275277, + -0.0003945471, + 0.02263768, + -0.10997523, + 0.09229477, + 0.048902728, + -0.044187002, + 0.05441158, + -0.0057972632, + 0.04834593, + 0.035639632, + -0.015485863, + -0.008143862, + 0.092880696, + 0.11231507, + 0.047900956, + -0.017541546, + -0.009539733, + 0.06213859, + -0.0040546083, + 0.003987384, + 0.09531304, + -0.056603517, + -0.058908645, + -0.013667576, + 0.009745052, + -0.047453303, + -0.06157018, + -0.08587985, + 0.05011287, + -0.02779263, + -0.008005466, + -0.068401575, + 0.032416083, + 0.015329646, + 0.08306027, + 0.06357283, + -0.00512495, + -0.01188288, + -0.051893827, + -0.008702526, + -0.031820606, + 0.043191314, + 0.00033676252, + -0.0012971128, + -2.3314325e-33, + -0.084871486, + 0.023456383, + -0.05555233, + 0.028799664, + 0.059832368, + 0.044252343, + -0.069759004, + -0.08750932, + -0.023541803, + 0.076747485, + 0.015193914, + 0.01961009, + -0.05837612, + 0.01878715, + 0.007621002, + -0.015989477, + -0.057301812, + -0.0426483, + 0.10103607, + -0.03979966, + -0.03179959, + 0.031775456, + -0.05796451, + -0.036753736, + 0.02731803, + -0.0069522746, + -0.07528311, + 0.049413346, + 0.012717442, + 0.10011093, + -0.03626197, + -0.0480568, + 0.029068258, + 0.017971879, + 0.04527712, + 0.10260452, + 0.0050376365, + -0.05527294, + 0.008323474, + -0.05968206, + 0.020133188, + 0.009408143, + -0.06650717, + -0.029911388, + 0.0434493, + -0.068347804, + -0.076517664, + 0.040012714, + -0.064759254, + 0.07230589, + 0.04662111, + -0.016778024, + -0.048703287, + -0.08456952, + -0.052551, + 0.03198548, + 0.024643922, + 0.02381256, + 0.07633642, + -0.040978454, + -0.033941545, + -0.11415368, + 0.067884214, + 0.009646611, + -0.06406483, + 0.02458555, + 0.024917984, + -0.0041125035, + 0.018718159, + -0.03810467, + 0.014550252, + 0.06850764, + 0.018693756, + 0.059391443, + 0.023741595, + -0.00974202, + -0.06651425, + 0.020927029, + -0.019371133, + 0.01486253, + 0.022714352, + -0.022630502, + 0.010553403, + 0.056958556, + 0.072571084, + 0.06506972, + -0.010076679, + 0.079096675, + 0.035260018, + -0.023826087, + 0.017108874, + 0.087825984, + 0.0059526036, + 0.0074271723, + -0.109360956, + -2.8789334e-08, + -0.05233612, + -0.087671354, + 0.066617705, + 0.013912193, + 0.099948354, + -0.02244002, + 0.062119395, + 0.027858257, + -0.064296365, + -0.038687464, + 0.025052465, + 0.008087938, + -0.024082167, + 0.011928929, + 0.0871567, + 0.012509529, + 0.064730704, + -0.027875392, + 0.039984196, + -0.012320989, + 0.023347521, + -0.032504674, + -0.042588573, + 0.107389025, + 0.037681337, + -0.06630358, + -0.056843463, + -0.0052555962, + -0.069520734, + 0.100924, + -0.033373408, + 0.02178169, + 0.017423104, + 0.01809016, + 0.02630718, + 0.066061474, + 0.059622575, + -0.065362565, + -0.11576683, + -0.071220115, + -0.023386031, + 0.042642016, + 0.043645483, + -0.036648206, + 0.05023266, + 0.0031018173, + 0.057091165, + -0.03462122, + 0.025469558, + -0.046201944, + -0.06719312, + 0.06058484, + -0.041243985, + -0.019823411, + -0.013743429, + -0.061215486, + 0.014752095, + -0.07632035, + -0.056729525, + 0.050518394, + -0.0360576, + 0.12239626, + 0.06431157, + -0.038293842 ] ] } diff --git a/tests/integration/recordings/responses/b4cda53cd04f.json b/tests/integration/recordings/responses/b4cda53cd04f.json index d2fb387a8..9dcb18f8b 100644 --- a/tests/integration/recordings/responses/b4cda53cd04f.json +++ b/tests/integration/recordings/responses/b4cda53cd04f.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -20,14 +20,14 @@ "body": { "__type__": "openai.types.chat.chat_completion.ChatCompletion", "__data__": { - "id": "chatcmpl-4", + "id": "chatcmpl-11", "choices": [ { "finish_reason": "stop", "index": 0, "logprobs": null, "message": { - "content": "Humans live on Earth.", + "content": "Humans live on Earth, which is a terrestrial planet located in the solar system. Specifically, it's classified as a rocky planet and the third planet from the Sun, orbiting at an average distance of about 149.6 million kilometers (92.96 million miles).", "refusal": null, "role": "assistant", "annotations": null, @@ -37,15 +37,15 @@ } } ], - "created": 1753814880, + "created": 1753984660, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion", "service_tier": null, "system_fingerprint": "fp_ollama", "usage": { - "completion_tokens": 6, + "completion_tokens": 55, "prompt_tokens": 32, - "total_tokens": 38, + "total_tokens": 87, "completion_tokens_details": null, "prompt_tokens_details": null } diff --git a/tests/integration/recordings/responses/b91f1fb4aedb.json b/tests/integration/recordings/responses/b91f1fb4aedb.json index fad1dc8fd..37111cc2d 100644 --- a/tests/integration/recordings/responses/b91f1fb4aedb.json +++ b/tests/integration/recordings/responses/b91f1fb4aedb.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.891582Z", + "created_at": "2025-07-31T17:54:09.896924388Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.939133Z", + "created_at": "2025-07-31T17:54:10.082480973Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:31.985171Z", + "created_at": "2025-07-31T17:54:10.264445159Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.030448Z", + "created_at": "2025-07-31T17:54:10.445463678Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.075659Z", + "created_at": "2025-07-31T17:54:10.627295509Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.123939Z", + "created_at": "2025-07-31T17:54:10.81236738Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.169545Z", + "created_at": "2025-07-31T17:54:10.999420818Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.214044Z", + "created_at": "2025-07-31T17:54:11.188320899Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.259104Z", + "created_at": "2025-07-31T17:54:11.379926424Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.306215Z", + "created_at": "2025-07-31T17:54:11.574460052Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,15 +201,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.351121Z", + "created_at": "2025-07-31T17:54:11.771461946Z", "done": true, "done_reason": "stop", - "total_duration": 641307458, - "load_duration": 70513916, + "total_duration": 4947159618, + "load_duration": 43936066, "prompt_eval_count": 339, - "prompt_eval_duration": 106020875, + "prompt_eval_duration": 3027152411, "eval_count": 11, - "eval_duration": 464057250, + "eval_duration": 1875512881, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/bbd0637dce16.json b/tests/integration/recordings/responses/bbd0637dce16.json index c1746a279..1fa733e18 100644 --- a/tests/integration/recordings/responses/bbd0637dce16.json +++ b/tests/integration/recordings/responses/bbd0637dce16.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.734568Z", + "created_at": "2025-07-31T17:54:14.346282017Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,7 +39,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.780322Z", + "created_at": "2025-07-31T17:54:14.52939924Z", "done": false, "done_reason": null, "total_duration": null, @@ -57,7 +57,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.822494Z", + "created_at": "2025-07-31T17:54:14.714949794Z", "done": false, "done_reason": null, "total_duration": null, @@ -75,7 +75,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.864477Z", + "created_at": "2025-07-31T17:54:14.894736492Z", "done": false, "done_reason": null, "total_duration": null, @@ -93,7 +93,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.905567Z", + "created_at": "2025-07-31T17:54:15.081363009Z", "done": false, "done_reason": null, "total_duration": null, @@ -111,7 +111,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.946526Z", + "created_at": "2025-07-31T17:54:15.264275377Z", "done": false, "done_reason": null, "total_duration": null, @@ -129,7 +129,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:32.987333Z", + "created_at": "2025-07-31T17:54:15.444600756Z", "done": false, "done_reason": null, "total_duration": null, @@ -147,7 +147,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.028636Z", + "created_at": "2025-07-31T17:54:15.624829048Z", "done": false, "done_reason": null, "total_duration": null, @@ -165,7 +165,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.069659Z", + "created_at": "2025-07-31T17:54:15.805880499Z", "done": false, "done_reason": null, "total_duration": null, @@ -183,7 +183,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.111852Z", + "created_at": "2025-07-31T17:54:15.987285938Z", "done": false, "done_reason": null, "total_duration": null, @@ -201,7 +201,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.154357Z", + "created_at": "2025-07-31T17:54:16.17196429Z", "done": false, "done_reason": null, "total_duration": null, @@ -219,7 +219,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.196576Z", + "created_at": "2025-07-31T17:54:16.359146297Z", "done": false, "done_reason": null, "total_duration": null, @@ -237,7 +237,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.241275Z", + "created_at": "2025-07-31T17:54:16.555039Z", "done": false, "done_reason": null, "total_duration": null, @@ -255,7 +255,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.288617Z", + "created_at": "2025-07-31T17:54:16.747367383Z", "done": false, "done_reason": null, "total_duration": null, @@ -273,7 +273,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.334713Z", + "created_at": "2025-07-31T17:54:16.935061953Z", "done": false, "done_reason": null, "total_duration": null, @@ -291,7 +291,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.379281Z", + "created_at": "2025-07-31T17:54:17.118660094Z", "done": false, "done_reason": null, "total_duration": null, @@ -309,7 +309,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.422844Z", + "created_at": "2025-07-31T17:54:17.30199003Z", "done": false, "done_reason": null, "total_duration": null, @@ -327,7 +327,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.465411Z", + "created_at": "2025-07-31T17:54:17.482657228Z", "done": false, "done_reason": null, "total_duration": null, @@ -345,7 +345,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.506968Z", + "created_at": "2025-07-31T17:54:17.661680024Z", "done": false, "done_reason": null, "total_duration": null, @@ -363,7 +363,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.548473Z", + "created_at": "2025-07-31T17:54:17.84105973Z", "done": false, "done_reason": null, "total_duration": null, @@ -381,7 +381,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.589558Z", + "created_at": "2025-07-31T17:54:18.021351615Z", "done": false, "done_reason": null, "total_duration": null, @@ -399,7 +399,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.630602Z", + "created_at": "2025-07-31T17:54:18.203546164Z", "done": false, "done_reason": null, "total_duration": null, @@ -417,7 +417,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.672127Z", + "created_at": "2025-07-31T17:54:18.386621093Z", "done": false, "done_reason": null, "total_duration": null, @@ -435,7 +435,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.713946Z", + "created_at": "2025-07-31T17:54:18.566868241Z", "done": false, "done_reason": null, "total_duration": null, @@ -453,7 +453,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.755302Z", + "created_at": "2025-07-31T17:54:18.747882123Z", "done": false, "done_reason": null, "total_duration": null, @@ -471,7 +471,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.796803Z", + "created_at": "2025-07-31T17:54:18.929988724Z", "done": false, "done_reason": null, "total_duration": null, @@ -489,7 +489,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.837979Z", + "created_at": "2025-07-31T17:54:19.113741944Z", "done": false, "done_reason": null, "total_duration": null, @@ -507,7 +507,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.879103Z", + "created_at": "2025-07-31T17:54:19.309719802Z", "done": false, "done_reason": null, "total_duration": null, @@ -525,7 +525,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.920442Z", + "created_at": "2025-07-31T17:54:19.496484724Z", "done": false, "done_reason": null, "total_duration": null, @@ -543,7 +543,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:33.961679Z", + "created_at": "2025-07-31T17:54:19.681877423Z", "done": false, "done_reason": null, "total_duration": null, @@ -561,7 +561,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.003538Z", + "created_at": "2025-07-31T17:54:19.86470071Z", "done": false, "done_reason": null, "total_duration": null, @@ -579,7 +579,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.047067Z", + "created_at": "2025-07-31T17:54:20.046320492Z", "done": false, "done_reason": null, "total_duration": null, @@ -597,7 +597,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.092011Z", + "created_at": "2025-07-31T17:54:20.231195326Z", "done": false, "done_reason": null, "total_duration": null, @@ -615,7 +615,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.14147Z", + "created_at": "2025-07-31T17:54:20.416953957Z", "done": false, "done_reason": null, "total_duration": null, @@ -633,7 +633,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.188688Z", + "created_at": "2025-07-31T17:54:20.606717706Z", "done": false, "done_reason": null, "total_duration": null, @@ -651,7 +651,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.23099Z", + "created_at": "2025-07-31T17:54:20.796684899Z", "done": false, "done_reason": null, "total_duration": null, @@ -669,7 +669,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.271917Z", + "created_at": "2025-07-31T17:54:20.988605781Z", "done": false, "done_reason": null, "total_duration": null, @@ -687,7 +687,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.313106Z", + "created_at": "2025-07-31T17:54:21.176611596Z", "done": false, "done_reason": null, "total_duration": null, @@ -705,7 +705,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.354534Z", + "created_at": "2025-07-31T17:54:21.364624459Z", "done": false, "done_reason": null, "total_duration": null, @@ -723,7 +723,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.396694Z", + "created_at": "2025-07-31T17:54:21.551710639Z", "done": false, "done_reason": null, "total_duration": null, @@ -741,7 +741,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.438042Z", + "created_at": "2025-07-31T17:54:21.74247772Z", "done": false, "done_reason": null, "total_duration": null, @@ -759,7 +759,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.479347Z", + "created_at": "2025-07-31T17:54:21.926758183Z", "done": false, "done_reason": null, "total_duration": null, @@ -777,7 +777,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.520112Z", + "created_at": "2025-07-31T17:54:22.112755801Z", "done": false, "done_reason": null, "total_duration": null, @@ -795,7 +795,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.56141Z", + "created_at": "2025-07-31T17:54:22.300570707Z", "done": false, "done_reason": null, "total_duration": null, @@ -813,7 +813,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.602445Z", + "created_at": "2025-07-31T17:54:22.487035456Z", "done": false, "done_reason": null, "total_duration": null, @@ -831,7 +831,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.64327Z", + "created_at": "2025-07-31T17:54:22.673888899Z", "done": false, "done_reason": null, "total_duration": null, @@ -849,7 +849,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.685056Z", + "created_at": "2025-07-31T17:54:22.85859447Z", "done": false, "done_reason": null, "total_duration": null, @@ -867,7 +867,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.726668Z", + "created_at": "2025-07-31T17:54:23.050738337Z", "done": false, "done_reason": null, "total_duration": null, @@ -885,7 +885,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.768538Z", + "created_at": "2025-07-31T17:54:23.243461441Z", "done": false, "done_reason": null, "total_duration": null, @@ -903,7 +903,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.810414Z", + "created_at": "2025-07-31T17:54:23.428547012Z", "done": false, "done_reason": null, "total_duration": null, @@ -921,7 +921,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.851436Z", + "created_at": "2025-07-31T17:54:23.614149313Z", "done": false, "done_reason": null, "total_duration": null, @@ -939,7 +939,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.893488Z", + "created_at": "2025-07-31T17:54:23.79871292Z", "done": false, "done_reason": null, "total_duration": null, @@ -957,7 +957,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.935748Z", + "created_at": "2025-07-31T17:54:23.993971096Z", "done": false, "done_reason": null, "total_duration": null, @@ -975,7 +975,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:34.976678Z", + "created_at": "2025-07-31T17:54:24.182818036Z", "done": false, "done_reason": null, "total_duration": null, @@ -993,7 +993,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.017535Z", + "created_at": "2025-07-31T17:54:24.368311299Z", "done": false, "done_reason": null, "total_duration": null, @@ -1011,7 +1011,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.058735Z", + "created_at": "2025-07-31T17:54:24.554855056Z", "done": false, "done_reason": null, "total_duration": null, @@ -1029,7 +1029,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.099818Z", + "created_at": "2025-07-31T17:54:24.739783206Z", "done": false, "done_reason": null, "total_duration": null, @@ -1047,7 +1047,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.141235Z", + "created_at": "2025-07-31T17:54:24.919169901Z", "done": false, "done_reason": null, "total_duration": null, @@ -1065,7 +1065,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.182196Z", + "created_at": "2025-07-31T17:54:25.099824662Z", "done": false, "done_reason": null, "total_duration": null, @@ -1083,7 +1083,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.224652Z", + "created_at": "2025-07-31T17:54:25.282939961Z", "done": false, "done_reason": null, "total_duration": null, @@ -1101,7 +1101,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.266777Z", + "created_at": "2025-07-31T17:54:25.466347174Z", "done": false, "done_reason": null, "total_duration": null, @@ -1119,7 +1119,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.30776Z", + "created_at": "2025-07-31T17:54:25.651265406Z", "done": false, "done_reason": null, "total_duration": null, @@ -1137,7 +1137,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.348575Z", + "created_at": "2025-07-31T17:54:25.834990001Z", "done": false, "done_reason": null, "total_duration": null, @@ -1155,7 +1155,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.389571Z", + "created_at": "2025-07-31T17:54:26.019818141Z", "done": false, "done_reason": null, "total_duration": null, @@ -1173,7 +1173,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.431018Z", + "created_at": "2025-07-31T17:54:26.205188289Z", "done": false, "done_reason": null, "total_duration": null, @@ -1191,7 +1191,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.47221Z", + "created_at": "2025-07-31T17:54:26.388926617Z", "done": false, "done_reason": null, "total_duration": null, @@ -1209,7 +1209,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.513378Z", + "created_at": "2025-07-31T17:54:26.572416679Z", "done": false, "done_reason": null, "total_duration": null, @@ -1227,7 +1227,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.554268Z", + "created_at": "2025-07-31T17:54:26.757831343Z", "done": false, "done_reason": null, "total_duration": null, @@ -1245,7 +1245,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.595227Z", + "created_at": "2025-07-31T17:54:26.939127101Z", "done": false, "done_reason": null, "total_duration": null, @@ -1263,7 +1263,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.636167Z", + "created_at": "2025-07-31T17:54:27.125717929Z", "done": false, "done_reason": null, "total_duration": null, @@ -1281,7 +1281,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.677032Z", + "created_at": "2025-07-31T17:54:27.306735967Z", "done": false, "done_reason": null, "total_duration": null, @@ -1299,7 +1299,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.717956Z", + "created_at": "2025-07-31T17:54:27.486472357Z", "done": false, "done_reason": null, "total_duration": null, @@ -1317,7 +1317,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.759034Z", + "created_at": "2025-07-31T17:54:27.667914623Z", "done": false, "done_reason": null, "total_duration": null, @@ -1335,7 +1335,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.800669Z", + "created_at": "2025-07-31T17:54:27.858779966Z", "done": false, "done_reason": null, "total_duration": null, @@ -1353,7 +1353,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.847116Z", + "created_at": "2025-07-31T17:54:28.039403127Z", "done": false, "done_reason": null, "total_duration": null, @@ -1371,7 +1371,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.890369Z", + "created_at": "2025-07-31T17:54:28.224070121Z", "done": false, "done_reason": null, "total_duration": null, @@ -1389,7 +1389,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.932566Z", + "created_at": "2025-07-31T17:54:28.415498584Z", "done": false, "done_reason": null, "total_duration": null, @@ -1407,7 +1407,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:35.973941Z", + "created_at": "2025-07-31T17:54:28.597971605Z", "done": false, "done_reason": null, "total_duration": null, @@ -1425,7 +1425,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.015033Z", + "created_at": "2025-07-31T17:54:28.777776103Z", "done": false, "done_reason": null, "total_duration": null, @@ -1443,7 +1443,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.05598Z", + "created_at": "2025-07-31T17:54:28.959922934Z", "done": false, "done_reason": null, "total_duration": null, @@ -1461,7 +1461,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.09692Z", + "created_at": "2025-07-31T17:54:29.145546229Z", "done": false, "done_reason": null, "total_duration": null, @@ -1479,7 +1479,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.138631Z", + "created_at": "2025-07-31T17:54:29.325162626Z", "done": false, "done_reason": null, "total_duration": null, @@ -1497,7 +1497,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.179745Z", + "created_at": "2025-07-31T17:54:29.506611664Z", "done": false, "done_reason": null, "total_duration": null, @@ -1515,7 +1515,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.220859Z", + "created_at": "2025-07-31T17:54:29.691331228Z", "done": false, "done_reason": null, "total_duration": null, @@ -1533,7 +1533,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.261698Z", + "created_at": "2025-07-31T17:54:29.877068274Z", "done": false, "done_reason": null, "total_duration": null, @@ -1551,7 +1551,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.30394Z", + "created_at": "2025-07-31T17:54:30.06009055Z", "done": false, "done_reason": null, "total_duration": null, @@ -1569,7 +1569,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.34674Z", + "created_at": "2025-07-31T17:54:30.242033122Z", "done": false, "done_reason": null, "total_duration": null, @@ -1587,7 +1587,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.386587Z", + "created_at": "2025-07-31T17:54:30.423487413Z", "done": false, "done_reason": null, "total_duration": null, @@ -1605,7 +1605,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.428548Z", + "created_at": "2025-07-31T17:54:30.609599401Z", "done": false, "done_reason": null, "total_duration": null, @@ -1623,7 +1623,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.471219Z", + "created_at": "2025-07-31T17:54:30.793301575Z", "done": false, "done_reason": null, "total_duration": null, @@ -1641,7 +1641,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.513168Z", + "created_at": "2025-07-31T17:54:30.972545984Z", "done": false, "done_reason": null, "total_duration": null, @@ -1659,7 +1659,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.55472Z", + "created_at": "2025-07-31T17:54:31.155886698Z", "done": false, "done_reason": null, "total_duration": null, @@ -1677,7 +1677,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.596569Z", + "created_at": "2025-07-31T17:54:31.34101472Z", "done": false, "done_reason": null, "total_duration": null, @@ -1695,7 +1695,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.638905Z", + "created_at": "2025-07-31T17:54:31.526108593Z", "done": false, "done_reason": null, "total_duration": null, @@ -1713,7 +1713,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.680526Z", + "created_at": "2025-07-31T17:54:31.713602609Z", "done": false, "done_reason": null, "total_duration": null, @@ -1731,7 +1731,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.722286Z", + "created_at": "2025-07-31T17:54:31.894945451Z", "done": false, "done_reason": null, "total_duration": null, @@ -1749,7 +1749,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.764106Z", + "created_at": "2025-07-31T17:54:32.081440199Z", "done": false, "done_reason": null, "total_duration": null, @@ -1767,7 +1767,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.806926Z", + "created_at": "2025-07-31T17:54:32.267787985Z", "done": false, "done_reason": null, "total_duration": null, @@ -1785,7 +1785,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.85187Z", + "created_at": "2025-07-31T17:54:32.448188568Z", "done": false, "done_reason": null, "total_duration": null, @@ -1803,7 +1803,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.893859Z", + "created_at": "2025-07-31T17:54:32.642930659Z", "done": false, "done_reason": null, "total_duration": null, @@ -1821,7 +1821,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.935921Z", + "created_at": "2025-07-31T17:54:32.825982387Z", "done": false, "done_reason": null, "total_duration": null, @@ -1839,7 +1839,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:36.977807Z", + "created_at": "2025-07-31T17:54:33.010425637Z", "done": false, "done_reason": null, "total_duration": null, @@ -1857,7 +1857,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.019661Z", + "created_at": "2025-07-31T17:54:33.194160049Z", "done": false, "done_reason": null, "total_duration": null, @@ -1875,7 +1875,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.061525Z", + "created_at": "2025-07-31T17:54:33.379248083Z", "done": false, "done_reason": null, "total_duration": null, @@ -1893,7 +1893,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.103815Z", + "created_at": "2025-07-31T17:54:33.563743684Z", "done": false, "done_reason": null, "total_duration": null, @@ -1911,7 +1911,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.14614Z", + "created_at": "2025-07-31T17:54:33.748061333Z", "done": false, "done_reason": null, "total_duration": null, @@ -1929,7 +1929,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.187851Z", + "created_at": "2025-07-31T17:54:33.928760223Z", "done": false, "done_reason": null, "total_duration": null, @@ -1947,7 +1947,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.232245Z", + "created_at": "2025-07-31T17:54:34.109239225Z", "done": false, "done_reason": null, "total_duration": null, @@ -1965,7 +1965,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.274382Z", + "created_at": "2025-07-31T17:54:34.293493176Z", "done": false, "done_reason": null, "total_duration": null, @@ -1983,7 +1983,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.318911Z", + "created_at": "2025-07-31T17:54:34.473228178Z", "done": false, "done_reason": null, "total_duration": null, @@ -2001,7 +2001,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.364488Z", + "created_at": "2025-07-31T17:54:34.655305546Z", "done": false, "done_reason": null, "total_duration": null, @@ -2019,7 +2019,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.406333Z", + "created_at": "2025-07-31T17:54:34.835624787Z", "done": false, "done_reason": null, "total_duration": null, @@ -2037,7 +2037,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.449795Z", + "created_at": "2025-07-31T17:54:35.012982088Z", "done": false, "done_reason": null, "total_duration": null, @@ -2055,7 +2055,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.492084Z", + "created_at": "2025-07-31T17:54:35.193157278Z", "done": false, "done_reason": null, "total_duration": null, @@ -2073,7 +2073,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.533935Z", + "created_at": "2025-07-31T17:54:35.369403767Z", "done": false, "done_reason": null, "total_duration": null, @@ -2091,7 +2091,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.575962Z", + "created_at": "2025-07-31T17:54:35.547587382Z", "done": false, "done_reason": null, "total_duration": null, @@ -2109,7 +2109,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.618139Z", + "created_at": "2025-07-31T17:54:35.726121535Z", "done": false, "done_reason": null, "total_duration": null, @@ -2127,7 +2127,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.660485Z", + "created_at": "2025-07-31T17:54:35.905829777Z", "done": false, "done_reason": null, "total_duration": null, @@ -2145,7 +2145,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.702762Z", + "created_at": "2025-07-31T17:54:36.09129003Z", "done": false, "done_reason": null, "total_duration": null, @@ -2163,7 +2163,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.744045Z", + "created_at": "2025-07-31T17:54:36.273703965Z", "done": false, "done_reason": null, "total_duration": null, @@ -2181,7 +2181,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.786164Z", + "created_at": "2025-07-31T17:54:36.450332934Z", "done": false, "done_reason": null, "total_duration": null, @@ -2199,7 +2199,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.828222Z", + "created_at": "2025-07-31T17:54:36.629217924Z", "done": false, "done_reason": null, "total_duration": null, @@ -2217,7 +2217,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.869943Z", + "created_at": "2025-07-31T17:54:36.806365658Z", "done": false, "done_reason": null, "total_duration": null, @@ -2235,7 +2235,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.912776Z", + "created_at": "2025-07-31T17:54:36.984327623Z", "done": false, "done_reason": null, "total_duration": null, @@ -2253,7 +2253,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.954933Z", + "created_at": "2025-07-31T17:54:37.166911308Z", "done": false, "done_reason": null, "total_duration": null, @@ -2271,7 +2271,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:37.996971Z", + "created_at": "2025-07-31T17:54:37.346878818Z", "done": false, "done_reason": null, "total_duration": null, @@ -2289,7 +2289,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.038907Z", + "created_at": "2025-07-31T17:54:37.527252601Z", "done": false, "done_reason": null, "total_duration": null, @@ -2307,7 +2307,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.081044Z", + "created_at": "2025-07-31T17:54:37.705701214Z", "done": false, "done_reason": null, "total_duration": null, @@ -2325,7 +2325,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.123049Z", + "created_at": "2025-07-31T17:54:37.885325419Z", "done": false, "done_reason": null, "total_duration": null, @@ -2343,7 +2343,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.165272Z", + "created_at": "2025-07-31T17:54:38.067474944Z", "done": false, "done_reason": null, "total_duration": null, @@ -2361,7 +2361,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.2065Z", + "created_at": "2025-07-31T17:54:38.251569266Z", "done": false, "done_reason": null, "total_duration": null, @@ -2379,7 +2379,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.248971Z", + "created_at": "2025-07-31T17:54:38.435335346Z", "done": false, "done_reason": null, "total_duration": null, @@ -2397,7 +2397,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.290632Z", + "created_at": "2025-07-31T17:54:38.616959483Z", "done": false, "done_reason": null, "total_duration": null, @@ -2415,7 +2415,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.332334Z", + "created_at": "2025-07-31T17:54:38.793826986Z", "done": false, "done_reason": null, "total_duration": null, @@ -2433,7 +2433,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.374467Z", + "created_at": "2025-07-31T17:54:38.971019847Z", "done": false, "done_reason": null, "total_duration": null, @@ -2451,7 +2451,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.415399Z", + "created_at": "2025-07-31T17:54:39.148209672Z", "done": false, "done_reason": null, "total_duration": null, @@ -2469,7 +2469,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.456774Z", + "created_at": "2025-07-31T17:54:39.32452659Z", "done": false, "done_reason": null, "total_duration": null, @@ -2487,7 +2487,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.498137Z", + "created_at": "2025-07-31T17:54:39.505731737Z", "done": false, "done_reason": null, "total_duration": null, @@ -2505,7 +2505,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.539292Z", + "created_at": "2025-07-31T17:54:39.68324544Z", "done": false, "done_reason": null, "total_duration": null, @@ -2523,7 +2523,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.580484Z", + "created_at": "2025-07-31T17:54:39.861527605Z", "done": false, "done_reason": null, "total_duration": null, @@ -2541,7 +2541,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.621541Z", + "created_at": "2025-07-31T17:54:40.03950865Z", "done": false, "done_reason": null, "total_duration": null, @@ -2559,7 +2559,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.662543Z", + "created_at": "2025-07-31T17:54:40.220135819Z", "done": false, "done_reason": null, "total_duration": null, @@ -2577,7 +2577,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.703884Z", + "created_at": "2025-07-31T17:54:40.397433129Z", "done": false, "done_reason": null, "total_duration": null, @@ -2595,7 +2595,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.744981Z", + "created_at": "2025-07-31T17:54:40.576831646Z", "done": false, "done_reason": null, "total_duration": null, @@ -2613,7 +2613,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.786144Z", + "created_at": "2025-07-31T17:54:40.755712976Z", "done": false, "done_reason": null, "total_duration": null, @@ -2631,7 +2631,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.827281Z", + "created_at": "2025-07-31T17:54:40.934839621Z", "done": false, "done_reason": null, "total_duration": null, @@ -2649,7 +2649,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.868476Z", + "created_at": "2025-07-31T17:54:41.114951105Z", "done": false, "done_reason": null, "total_duration": null, @@ -2667,7 +2667,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.910121Z", + "created_at": "2025-07-31T17:54:41.292251634Z", "done": false, "done_reason": null, "total_duration": null, @@ -2685,7 +2685,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.952457Z", + "created_at": "2025-07-31T17:54:41.472339804Z", "done": false, "done_reason": null, "total_duration": null, @@ -2703,7 +2703,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:38.99402Z", + "created_at": "2025-07-31T17:54:41.652743941Z", "done": false, "done_reason": null, "total_duration": null, @@ -2721,7 +2721,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.035663Z", + "created_at": "2025-07-31T17:54:41.831983649Z", "done": false, "done_reason": null, "total_duration": null, @@ -2739,7 +2739,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.077114Z", + "created_at": "2025-07-31T17:54:42.011794018Z", "done": false, "done_reason": null, "total_duration": null, @@ -2757,7 +2757,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.119566Z", + "created_at": "2025-07-31T17:54:42.192167124Z", "done": false, "done_reason": null, "total_duration": null, @@ -2775,7 +2775,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.161288Z", + "created_at": "2025-07-31T17:54:42.372655402Z", "done": false, "done_reason": null, "total_duration": null, @@ -2793,7 +2793,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.202588Z", + "created_at": "2025-07-31T17:54:42.552895703Z", "done": false, "done_reason": null, "total_duration": null, @@ -2811,7 +2811,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.243626Z", + "created_at": "2025-07-31T17:54:42.736452553Z", "done": false, "done_reason": null, "total_duration": null, @@ -2829,7 +2829,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.285009Z", + "created_at": "2025-07-31T17:54:42.920965653Z", "done": false, "done_reason": null, "total_duration": null, @@ -2847,7 +2847,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.326201Z", + "created_at": "2025-07-31T17:54:43.107973824Z", "done": false, "done_reason": null, "total_duration": null, @@ -2865,7 +2865,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.367613Z", + "created_at": "2025-07-31T17:54:43.295841598Z", "done": false, "done_reason": null, "total_duration": null, @@ -2883,7 +2883,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.409895Z", + "created_at": "2025-07-31T17:54:43.487677068Z", "done": false, "done_reason": null, "total_duration": null, @@ -2901,7 +2901,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.453475Z", + "created_at": "2025-07-31T17:54:43.670981299Z", "done": false, "done_reason": null, "total_duration": null, @@ -2919,7 +2919,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.494355Z", + "created_at": "2025-07-31T17:54:43.859031006Z", "done": false, "done_reason": null, "total_duration": null, @@ -2937,7 +2937,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.536621Z", + "created_at": "2025-07-31T17:54:44.058854438Z", "done": false, "done_reason": null, "total_duration": null, @@ -2955,7 +2955,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.577862Z", + "created_at": "2025-07-31T17:54:44.247165646Z", "done": false, "done_reason": null, "total_duration": null, @@ -2973,7 +2973,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.619208Z", + "created_at": "2025-07-31T17:54:44.443919945Z", "done": false, "done_reason": null, "total_duration": null, @@ -2991,7 +2991,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.66022Z", + "created_at": "2025-07-31T17:54:44.640345509Z", "done": false, "done_reason": null, "total_duration": null, @@ -3009,7 +3009,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.701241Z", + "created_at": "2025-07-31T17:54:44.826366647Z", "done": false, "done_reason": null, "total_duration": null, @@ -3027,7 +3027,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.74224Z", + "created_at": "2025-07-31T17:54:45.018273616Z", "done": false, "done_reason": null, "total_duration": null, @@ -3045,7 +3045,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.783122Z", + "created_at": "2025-07-31T17:54:45.214270158Z", "done": false, "done_reason": null, "total_duration": null, @@ -3063,7 +3063,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.824227Z", + "created_at": "2025-07-31T17:54:45.407600479Z", "done": false, "done_reason": null, "total_duration": null, @@ -3081,7 +3081,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.86602Z", + "created_at": "2025-07-31T17:54:45.587463306Z", "done": false, "done_reason": null, "total_duration": null, @@ -3099,7 +3099,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.908909Z", + "created_at": "2025-07-31T17:54:45.763700034Z", "done": false, "done_reason": null, "total_duration": null, @@ -3117,7 +3117,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.951188Z", + "created_at": "2025-07-31T17:54:45.942709439Z", "done": false, "done_reason": null, "total_duration": null, @@ -3135,7 +3135,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:39.992916Z", + "created_at": "2025-07-31T17:54:46.120338544Z", "done": false, "done_reason": null, "total_duration": null, @@ -3153,7 +3153,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.034549Z", + "created_at": "2025-07-31T17:54:46.300553317Z", "done": false, "done_reason": null, "total_duration": null, @@ -3171,7 +3171,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.076372Z", + "created_at": "2025-07-31T17:54:46.479050541Z", "done": false, "done_reason": null, "total_duration": null, @@ -3189,7 +3189,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.11828Z", + "created_at": "2025-07-31T17:54:46.658713024Z", "done": false, "done_reason": null, "total_duration": null, @@ -3207,7 +3207,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.160065Z", + "created_at": "2025-07-31T17:54:46.839290118Z", "done": false, "done_reason": null, "total_duration": null, @@ -3225,7 +3225,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.201565Z", + "created_at": "2025-07-31T17:54:47.016963671Z", "done": false, "done_reason": null, "total_duration": null, @@ -3243,7 +3243,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.242918Z", + "created_at": "2025-07-31T17:54:47.194067179Z", "done": false, "done_reason": null, "total_duration": null, @@ -3261,7 +3261,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.284165Z", + "created_at": "2025-07-31T17:54:47.373898514Z", "done": false, "done_reason": null, "total_duration": null, @@ -3279,7 +3279,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.325345Z", + "created_at": "2025-07-31T17:54:47.552058297Z", "done": false, "done_reason": null, "total_duration": null, @@ -3297,7 +3297,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.36659Z", + "created_at": "2025-07-31T17:54:47.728696074Z", "done": false, "done_reason": null, "total_duration": null, @@ -3315,7 +3315,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.407669Z", + "created_at": "2025-07-31T17:54:47.903507242Z", "done": false, "done_reason": null, "total_duration": null, @@ -3333,7 +3333,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.448949Z", + "created_at": "2025-07-31T17:54:48.082542103Z", "done": false, "done_reason": null, "total_duration": null, @@ -3351,7 +3351,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.490103Z", + "created_at": "2025-07-31T17:54:48.257942426Z", "done": false, "done_reason": null, "total_duration": null, @@ -3369,7 +3369,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.531327Z", + "created_at": "2025-07-31T17:54:48.432408495Z", "done": false, "done_reason": null, "total_duration": null, @@ -3387,7 +3387,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.572437Z", + "created_at": "2025-07-31T17:54:48.611052957Z", "done": false, "done_reason": null, "total_duration": null, @@ -3405,7 +3405,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.613651Z", + "created_at": "2025-07-31T17:54:48.787480718Z", "done": false, "done_reason": null, "total_duration": null, @@ -3423,7 +3423,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.655535Z", + "created_at": "2025-07-31T17:54:48.968500481Z", "done": false, "done_reason": null, "total_duration": null, @@ -3441,7 +3441,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.697005Z", + "created_at": "2025-07-31T17:54:49.149465089Z", "done": false, "done_reason": null, "total_duration": null, @@ -3459,7 +3459,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.738543Z", + "created_at": "2025-07-31T17:54:49.330734142Z", "done": false, "done_reason": null, "total_duration": null, @@ -3477,7 +3477,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.780037Z", + "created_at": "2025-07-31T17:54:49.508892429Z", "done": false, "done_reason": null, "total_duration": null, @@ -3495,7 +3495,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.82152Z", + "created_at": "2025-07-31T17:54:49.684465885Z", "done": false, "done_reason": null, "total_duration": null, @@ -3513,7 +3513,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.863659Z", + "created_at": "2025-07-31T17:54:49.862974656Z", "done": false, "done_reason": null, "total_duration": null, @@ -3531,7 +3531,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.905889Z", + "created_at": "2025-07-31T17:54:50.038079229Z", "done": false, "done_reason": null, "total_duration": null, @@ -3549,7 +3549,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.948062Z", + "created_at": "2025-07-31T17:54:50.215836607Z", "done": false, "done_reason": null, "total_duration": null, @@ -3567,7 +3567,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:40.989336Z", + "created_at": "2025-07-31T17:54:50.394935584Z", "done": false, "done_reason": null, "total_duration": null, @@ -3585,7 +3585,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.030953Z", + "created_at": "2025-07-31T17:54:50.570525491Z", "done": false, "done_reason": null, "total_duration": null, @@ -3603,7 +3603,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.072459Z", + "created_at": "2025-07-31T17:54:50.74798129Z", "done": false, "done_reason": null, "total_duration": null, @@ -3621,7 +3621,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.114443Z", + "created_at": "2025-07-31T17:54:50.927182397Z", "done": false, "done_reason": null, "total_duration": null, @@ -3639,7 +3639,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.156211Z", + "created_at": "2025-07-31T17:54:51.104970726Z", "done": false, "done_reason": null, "total_duration": null, @@ -3657,7 +3657,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.197985Z", + "created_at": "2025-07-31T17:54:51.28314893Z", "done": false, "done_reason": null, "total_duration": null, @@ -3675,7 +3675,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.239925Z", + "created_at": "2025-07-31T17:54:51.45834561Z", "done": false, "done_reason": null, "total_duration": null, @@ -3693,7 +3693,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.282Z", + "created_at": "2025-07-31T17:54:51.635708096Z", "done": false, "done_reason": null, "total_duration": null, @@ -3711,7 +3711,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.323767Z", + "created_at": "2025-07-31T17:54:51.814425693Z", "done": false, "done_reason": null, "total_duration": null, @@ -3729,7 +3729,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.365403Z", + "created_at": "2025-07-31T17:54:51.992397527Z", "done": false, "done_reason": null, "total_duration": null, @@ -3747,7 +3747,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.407109Z", + "created_at": "2025-07-31T17:54:52.171570528Z", "done": false, "done_reason": null, "total_duration": null, @@ -3765,7 +3765,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.449495Z", + "created_at": "2025-07-31T17:54:52.349997427Z", "done": false, "done_reason": null, "total_duration": null, @@ -3783,7 +3783,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.491013Z", + "created_at": "2025-07-31T17:54:52.525371439Z", "done": false, "done_reason": null, "total_duration": null, @@ -3801,7 +3801,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.532665Z", + "created_at": "2025-07-31T17:54:52.705645938Z", "done": false, "done_reason": null, "total_duration": null, @@ -3819,7 +3819,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.573844Z", + "created_at": "2025-07-31T17:54:52.883992223Z", "done": false, "done_reason": null, "total_duration": null, @@ -3837,7 +3837,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.615535Z", + "created_at": "2025-07-31T17:54:53.064097937Z", "done": false, "done_reason": null, "total_duration": null, @@ -3855,7 +3855,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.658551Z", + "created_at": "2025-07-31T17:54:53.240460062Z", "done": false, "done_reason": null, "total_duration": null, @@ -3873,7 +3873,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.700054Z", + "created_at": "2025-07-31T17:54:53.419193116Z", "done": false, "done_reason": null, "total_duration": null, @@ -3891,7 +3891,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.741849Z", + "created_at": "2025-07-31T17:54:53.599456757Z", "done": false, "done_reason": null, "total_duration": null, @@ -3909,7 +3909,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.783403Z", + "created_at": "2025-07-31T17:54:53.78307921Z", "done": false, "done_reason": null, "total_duration": null, @@ -3927,7 +3927,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.824851Z", + "created_at": "2025-07-31T17:54:53.965372901Z", "done": false, "done_reason": null, "total_duration": null, @@ -3945,7 +3945,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.866422Z", + "created_at": "2025-07-31T17:54:54.146391033Z", "done": false, "done_reason": null, "total_duration": null, @@ -3963,7 +3963,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.908779Z", + "created_at": "2025-07-31T17:54:54.326644865Z", "done": false, "done_reason": null, "total_duration": null, @@ -3981,7 +3981,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.950867Z", + "created_at": "2025-07-31T17:54:54.506866496Z", "done": false, "done_reason": null, "total_duration": null, @@ -3999,7 +3999,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:41.992421Z", + "created_at": "2025-07-31T17:54:54.683701545Z", "done": false, "done_reason": null, "total_duration": null, @@ -4017,7 +4017,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.034025Z", + "created_at": "2025-07-31T17:54:54.864896708Z", "done": false, "done_reason": null, "total_duration": null, @@ -4035,7 +4035,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.076122Z", + "created_at": "2025-07-31T17:54:55.04477992Z", "done": false, "done_reason": null, "total_duration": null, @@ -4053,7 +4053,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.118733Z", + "created_at": "2025-07-31T17:54:55.220726112Z", "done": false, "done_reason": null, "total_duration": null, @@ -4071,7 +4071,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.160604Z", + "created_at": "2025-07-31T17:54:55.39955031Z", "done": false, "done_reason": null, "total_duration": null, @@ -4089,7 +4089,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.202442Z", + "created_at": "2025-07-31T17:54:55.574953894Z", "done": false, "done_reason": null, "total_duration": null, @@ -4107,7 +4107,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.244163Z", + "created_at": "2025-07-31T17:54:55.751159487Z", "done": false, "done_reason": null, "total_duration": null, @@ -4125,15 +4125,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:42.285586Z", + "created_at": "2025-07-31T17:54:55.930109804Z", "done": true, "done_reason": "stop", - "total_duration": 9738300667, - "load_duration": 115362042, + "total_duration": 44012354454, + "load_duration": 42929539, "prompt_eval_count": 34, - "prompt_eval_duration": 70657583, + "prompt_eval_duration": 2383587559, "eval_count": 229, - "eval_duration": 9551589208, + "eval_duration": 41585294334, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/c9cba6f3ee38.json b/tests/integration/recordings/responses/c9cba6f3ee38.json index d1595fedb..4240fe90b 100644 --- a/tests/integration/recordings/responses/c9cba6f3ee38.json +++ b/tests/integration/recordings/responses/c9cba6f3ee38.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:25.579594Z", + "created_at": "2025-07-31T17:46:15.281412692Z", "done": true, "done_reason": "stop", - "total_duration": 268778000, - "load_duration": 52132709, + "total_duration": 2848280419, + "load_duration": 40304146, "prompt_eval_count": 219, - "prompt_eval_duration": 203828500, + "prompt_eval_duration": 2765376163, "eval_count": 2, - "eval_duration": 12057875, + "eval_duration": 42099357, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/d0ac68cbde69.json b/tests/integration/recordings/responses/d0ac68cbde69.json index b37962fb6..63b352b44 100644 --- a/tests/integration/recordings/responses/d0ac68cbde69.json +++ b/tests/integration/recordings/responses/d0ac68cbde69.json @@ -11,7 +11,62 @@ "body": { "__type__": "ollama._types.ProcessResponse", "__data__": { - "models": [] + "models": [ + { + "model": "all-minilm:l6-v2", + "name": "all-minilm:l6-v2", + "digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef", + "expires_at": "2025-07-31T18:05:06.618179Z", + "size": 53334016, + "size_vram": 0, + "details": { + "parent_model": "", + "format": "gguf", + "family": "bert", + "families": [ + "bert" + ], + "parameter_size": "23M", + "quantization_level": "F16" + } + }, + { + "model": "llama3.2:3b-instruct-fp16", + "name": "llama3.2:3b-instruct-fp16", + "digest": "195a8c01d91ec3cb1e0aad4624a51f2602c51fa7d96110f8ab5a20c84081804d", + "expires_at": "2025-07-31T18:04:42.166739Z", + "size": 7382700032, + "size_vram": 0, + "details": { + "parent_model": "", + "format": "gguf", + "family": "llama", + "families": [ + "llama" + ], + "parameter_size": "3.2B", + "quantization_level": "F16" + } + }, + { + "model": "llama-guard3:1b", + "name": "llama-guard3:1b", + "digest": "494147e06bf99e10dbe67b63a07ac81c162f18ef3341aa3390007ac828571b3b", + "expires_at": "2025-07-31T18:04:06.788447Z", + "size": 1814095872, + "size_vram": 0, + "details": { + "parent_model": "", + "format": "gguf", + "family": "llama", + "families": [ + "llama" + ], + "parameter_size": "1.5B", + "quantization_level": "Q8_0" + } + } + ] } }, "is_streaming": false diff --git a/tests/integration/recordings/responses/d4c86ac355fb.json b/tests/integration/recordings/responses/d4c86ac355fb.json index 4a1b70dc2..418ccb9c7 100644 --- a/tests/integration/recordings/responses/d4c86ac355fb.json +++ b/tests/integration/recordings/responses/d4c86ac355fb.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:56.580734Z", + "created_at": "2025-07-31T17:59:05.729198288Z", "done": true, "done_reason": "stop", - "total_duration": 145215666, - "load_duration": 72557916, + "total_duration": 1097329407, + "load_duration": 48350112, "prompt_eval_count": 220, - "prompt_eval_duration": 60363125, + "prompt_eval_duration": 1004630210, "eval_count": 2, - "eval_duration": 11629750, + "eval_duration": 43843552, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/dac7a32e5db9.json b/tests/integration/recordings/responses/dac7a32e5db9.json index 675eef3b0..a28144442 100644 --- a/tests/integration/recordings/responses/dac7a32e5db9.json +++ b/tests/integration/recordings/responses/dac7a32e5db9.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:08.784695Z", + "created_at": "2025-07-31T17:50:00.921192644Z", "done": true, "done_reason": "stop", - "total_duration": 420724000, - "load_duration": 57238084, + "total_duration": 2073152067, + "load_duration": 42902450, "prompt_eval_count": 23, - "prompt_eval_duration": 72133167, + "prompt_eval_duration": 795517987, "eval_count": 8, - "eval_duration": 290696708, + "eval_duration": 1234259942, "response": "The capital of France is Paris.", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/dd226d71f844.json b/tests/integration/recordings/responses/dd226d71f844.json index 7d9cadf40..1083750d2 100644 --- a/tests/integration/recordings/responses/dd226d71f844.json +++ b/tests/integration/recordings/responses/dd226d71f844.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.329935Z", + "created_at": "2025-07-31T17:46:34.048227816Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.37326Z", + "created_at": "2025-07-31T17:46:34.229102748Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.415761Z", + "created_at": "2025-07-31T17:46:34.413533748Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.458843Z", + "created_at": "2025-07-31T17:46:34.59977971Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.501468Z", + "created_at": "2025-07-31T17:46:34.781710387Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.543451Z", + "created_at": "2025-07-31T17:46:34.962137136Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.586683Z", + "created_at": "2025-07-31T17:46:35.142960328Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.629666Z", + "created_at": "2025-07-31T17:46:35.322879215Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.672199Z", + "created_at": "2025-07-31T17:46:35.504172249Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.71471Z", + "created_at": "2025-07-31T17:46:35.685767411Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.757321Z", + "created_at": "2025-07-31T17:46:35.865854085Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.801345Z", + "created_at": "2025-07-31T17:46:36.046336351Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,15 +238,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:29.844187Z", + "created_at": "2025-07-31T17:46:36.228600193Z", "done": true, "done_reason": "stop", - "total_duration": 691818542, - "load_duration": 102634584, + "total_duration": 4246630545, + "load_duration": 46591467, "prompt_eval_count": 402, - "prompt_eval_duration": 72389458, + "prompt_eval_duration": 2017607358, "eval_count": 13, - "eval_duration": 516194167, + "eval_duration": 2181831592, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/dd9e7d5913e9.json b/tests/integration/recordings/responses/dd9e7d5913e9.json index af89f9076..03c673405 100644 --- a/tests/integration/recordings/responses/dd9e7d5913e9.json +++ b/tests/integration/recordings/responses/dd9e7d5913e9.json @@ -21,7 +21,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:48.95435Z", + "created_at": "2025-07-31T17:56:02.787875238Z", "done": false, "done_reason": null, "total_duration": null, @@ -39,15 +39,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T18:47:48.996247Z", + "created_at": "2025-07-31T17:56:02.972585441Z", "done": true, "done_reason": "stop", - "total_duration": 667274458, - "load_duration": 80712750, + "total_duration": 32348917084, + "load_duration": 39040993, "prompt_eval_count": 386, - "prompt_eval_duration": 543388792, + "prompt_eval_duration": 32123680358, "eval_count": 2, - "eval_duration": 42471125, + "eval_duration": 185666617, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/e08848bfcd28.json b/tests/integration/recordings/responses/e08848bfcd28.json index 8dd27869c..94096d224 100644 --- a/tests/integration/recordings/responses/e08848bfcd28.json +++ b/tests/integration/recordings/responses/e08848bfcd28.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 91772250, - "load_duration": 72008875, + "total_duration": 19433209, + "load_duration": 8285331, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.07473014, - 0.08137506, - -0.06463602, - 0.011821943, - -0.07454815, - 0.021821007, - 0.077573344, - 0.012804661, - 0.05853777, - -0.014141324, - 0.053993534, - -0.026554074, - -0.018055506, - -0.060447972, - -0.019253474, - -0.006501444, - -0.047272332, - -0.048944764, - -0.090516366, - -0.06656194, - 0.09287066, - 0.02129739, - -0.013401809, - -0.006629013, - 0.0079892, - 0.016818035, - 0.03971694, - 0.021875564, - 0.014873574, - -0.039426163, - 0.025255844, - -0.036836684, - 0.016627828, - 0.008789532, - -0.053503897, - 0.03616121, - -0.034633957, - -0.009877797, - 0.064843215, - -0.01517806, - 0.020897496, - -0.07135096, - -0.008519908, - 0.05118655, - -0.062102985, - 0.059486073, - -0.047937352, - 0.07045817, - -0.024867272, - -0.010756205, - 0.06538509, - -0.03693754, - -0.08240387, - 0.08169191, - 0.017090658, - 0.012944557, - -0.047139525, - 0.0025796075, - 0.008701712, - 0.099866174, - 0.04969699, - -0.025922626, - -0.017354922, - 0.03395182, - 0.038391408, - -0.054247838, - 0.008610521, - -0.04077977, - 0.0265637, - -0.07186012, - -0.019953186, - -0.041191205, - -0.07246228, - 0.00041248833, - 0.018758524, - 0.023036895, - 0.01662864, - -0.06335885, - 0.03495032, - 0.050063577, - 0.00043262896, - -0.06176693, - 0.0062733325, - 0.11142063, - 0.0040838965, - 0.085737824, - 0.023284689, - 0.05699812, - -0.03149832, - -0.013344509, - -0.045138564, - -0.117300816, - 0.016063986, - -0.016894838, - -0.028934335, - 0.03575864, - -0.05156192, - 0.032958068, - -0.11266628, - 0.06640015, - 0.037839692, - 0.022948038, - 0.058071073, - -0.039643735, - -0.03247236, - 0.017690921, - -0.005001274, - 0.019046135, - 0.07745316, - -0.020402163, - -0.020310633, - -0.009519755, - 0.0031459313, - -0.0045639877, - -0.029116316, - 0.033835515, - 0.00050839526, - 0.06419946, - 0.010721198, - 0.124151744, - -0.0053820186, - 0.00491648, - -0.059696514, - 0.029483523, - -0.13409872, - 0.016187217, - -0.048092023, - -6.6084764e-33, - 0.012305612, - 0.060384244, - 0.036461998, - -0.035974216, - -0.04197416, - 0.012333701, - -0.084805995, - 0.012502633, - 0.02794982, - 0.0861082, - -0.030791838, - -0.061355945, - -0.0009604986, - -0.0252044, - 0.045444816, - -0.027590565, - -0.009594973, - 0.006712001, - 0.043692384, - -0.021483036, - 0.003300438, - 0.11860881, - 0.047044385, - -0.1348901, - 0.025469579, - -0.01029819, - 0.0022393467, - -0.061863262, - 0.10386513, - 0.018658707, - -0.0017492755, - -0.051914047, - 0.046442248, - 0.03761067, - 0.033752125, - 0.006650237, - 0.022015076, - -0.07834835, - -0.008209136, - 0.027432231, - 0.017393896, - -0.07524756, - 0.006497012, - 0.027272953, - 0.0005804994, - -0.010941825, - -0.020050043, - -0.00012092298, - 0.013705002, - 0.004699541, - 0.022770848, - 0.015477994, - -0.0142482165, - -0.013953546, - 0.015865315, - -0.023075614, - 0.03379947, - -0.039221376, - -0.043229815, - 0.02998769, - -0.01652291, - 0.06981088, - 0.04606923, - 0.05332633, - -0.055300076, - 0.02511626, - 0.014049543, - -0.09398743, - 0.03590562, - 0.029452223, - -0.13200304, - -0.005059034, - -0.03784268, - -0.03180819, - -0.095502876, - -0.027853556, - 0.0024331037, - -0.007881495, - 0.058296, - -0.031999517, - -0.06077097, - -0.023381822, - -0.00048603877, - 0.13765746, - -0.060579, - -0.008109843, - -0.034873307, - -0.1024547, - -0.009072849, - -0.018931676, - -0.0016711762, - -0.07710289, - -0.043332253, - -0.03619527, - 0.03958017, - 3.0217083e-33, - 0.0050329794, - 0.00016030145, - -0.063078895, - 0.012225751, - 0.10637338, - 0.015972024, - 0.006653195, - 0.01880781, - -0.04708357, - 0.045863643, - 0.0076015075, - 0.03243478, - 0.032097474, - -0.020893326, - 0.10697852, - 0.0075498912, - 0.036074348, - 0.1462344, - 0.03779065, - -0.043190572, - -0.02176097, - -0.009340132, - -0.06983617, - 0.015578788, - 0.021121953, - 0.030661412, - 0.08434581, - -0.09288574, - 0.008169474, - 0.078080945, - -0.081626564, - 0.011895231, - 0.017099649, - 0.0040119104, - -0.14145434, - 0.0040375097, - 0.046316408, - 0.008959473, - -0.0056506568, - -0.055587813, - 0.028007837, - 0.055937108, - 0.062269785, - 0.08602392, - -0.12157818, - 0.021943888, - -0.0050934856, - 0.029819332, - -0.012127162, - 0.048801802, - 0.06409215, - -0.041438665, - 0.01809265, - -0.028214281, - -0.0213588, - 0.05564267, - -0.1547868, - 0.027465124, - 0.018855799, - 0.04327939, - 0.011500479, - 0.017364705, - -0.023216385, - 0.051007293, - 0.02946264, - 0.012533944, - -0.04542834, - -0.002238765, - -0.05611544, - -0.0789272, - 0.07960444, - -0.020431034, - -0.0762138, - 0.011588508, - -0.035614885, - -0.04803985, - -0.06607436, - -0.057365946, - -0.040188126, - 0.07176218, - 0.03135825, - 0.02303279, - -0.023997622, - 0.023614945, - 0.09607302, - -0.06843066, - 0.014260722, - 0.08802569, - -0.037736766, - 0.029445928, - -0.028643936, - 0.10217973, - -0.0660917, - 0.022864237, - 0.042151757, - -1.4814046e-08, - 0.030838449, - 0.043877687, - -0.0245681, - -0.09818859, - 0.056659035, - 0.0929652, - -0.010337853, - -0.0983916, - 0.018008571, - -0.0131424805, - 0.026400762, - 0.008793538, - -0.05285605, - -0.042175982, - 0.030133193, - 0.01710666, - -0.06242493, - -0.018753909, - -0.015986755, - -0.018400662, - -0.026477808, - 0.010281372, - -0.030476814, - -0.084556945, - -0.05402664, - 0.010030052, - 0.029531356, - 0.13555466, - 0.033426728, - 0.12098221, - 0.040777553, - 0.008206964, - -0.018235989, - -0.0568263, - -0.1289943, - 0.12416113, - -0.053454727, - -0.038151894, - 0.030221034, - 0.019807614, - 0.047819767, - 0.029434063, - 0.0015704447, - 0.0611775, - -0.05557245, - -0.030236417, - 0.10799873, - -0.07073352, - -0.08215229, - 0.004518122, - -0.015573616, - -0.013696145, - -0.0023438279, - 0.026377691, - -0.015769389, - 0.016251203, - -0.04062322, - -0.013962793, - -0.08309221, - 0.031991288, - 0.049991824, - -0.0038595141, - 0.07031122, - 0.0049263495 + -0.07474477, + 0.08135398, + -0.064634696, + 0.011828698, + -0.07453829, + 0.02182288, + 0.07757148, + 0.012819247, + 0.058521118, + -0.014155038, + 0.05401544, + -0.026546013, + -0.0180307, + -0.060459055, + -0.019297317, + -0.0065130494, + -0.047251288, + -0.048953593, + -0.09050955, + -0.066574454, + 0.09286769, + 0.021295147, + -0.013418025, + -0.0066250013, + 0.008012828, + 0.016831141, + 0.03971127, + 0.021881687, + 0.014906969, + -0.03941482, + 0.025240429, + -0.036830667, + 0.016622525, + 0.0087695215, + -0.053536925, + 0.03617892, + -0.03462396, + -0.009888095, + 0.064840525, + -0.015164931, + 0.02093279, + -0.07134467, + -0.008522566, + 0.051180046, + -0.062126577, + 0.059486803, + -0.047944583, + 0.070436746, + -0.02486217, + -0.0107475445, + 0.06539541, + -0.036926474, + -0.08239496, + 0.081696264, + 0.0170991, + 0.012957423, + -0.0471463, + 0.0025890933, + 0.008701811, + 0.099854566, + 0.04971971, + -0.025905404, + -0.017348187, + 0.033950463, + 0.038388666, + -0.05427885, + 0.008590445, + -0.04073761, + 0.02659113, + -0.071868345, + -0.01996084, + -0.04121524, + -0.0725133, + 0.00041767213, + 0.018783698, + 0.023019075, + 0.016632775, + -0.063352264, + 0.0349477, + 0.05009186, + 0.00040466923, + -0.061769173, + 0.0062912344, + 0.111402325, + 0.004091874, + 0.085738786, + 0.023293864, + 0.056976296, + -0.03147358, + -0.013350156, + -0.045120586, + -0.1172993, + 0.016056577, + -0.016870052, + -0.028950103, + 0.035740018, + -0.05160049, + 0.03299421, + -0.112658225, + 0.06639703, + 0.037844308, + 0.022936413, + 0.058079578, + -0.03964461, + -0.032472335, + 0.017667979, + -0.0050076996, + 0.019045876, + 0.07745966, + -0.020397017, + -0.0203175, + -0.009526668, + 0.0031668684, + -0.0045527318, + -0.029142363, + 0.03386657, + 0.000497873, + 0.0641948, + 0.010712736, + 0.124138065, + -0.005370307, + 0.004939277, + -0.059704892, + 0.029501028, + -0.13407569, + 0.016168706, + -0.048107002, + -6.609533e-33, + 0.012297678, + 0.060365792, + 0.036464352, + -0.035990164, + -0.041963857, + 0.012336972, + -0.08480322, + 0.012513757, + 0.027956247, + 0.08611167, + -0.030789725, + -0.061359033, + -0.0009869405, + -0.025196983, + 0.045444682, + -0.027597353, + -0.009597403, + 0.0067014666, + 0.043697767, + -0.021481235, + 0.0032871987, + 0.11861492, + 0.047057446, + -0.13488701, + 0.025470478, + -0.010290927, + 0.0022252023, + -0.061861552, + 0.10386814, + 0.018669913, + -0.0017687998, + -0.05192628, + 0.046416067, + 0.03761, + 0.03375052, + 0.0066689374, + 0.022028843, + -0.07836795, + -0.008201764, + 0.027449533, + 0.017385295, + -0.07525137, + 0.0064914557, + 0.027283166, + 0.00058765704, + -0.010939965, + -0.02004952, + -0.00010343878, + 0.013708183, + 0.0047120173, + 0.02279417, + 0.015444727, + -0.014262697, + -0.013951755, + 0.015873758, + -0.023056472, + 0.033782665, + -0.03921495, + -0.04322829, + 0.02998795, + -0.016523762, + 0.06979977, + 0.046053544, + 0.053350296, + -0.055315897, + 0.025097992, + 0.014044151, + -0.09400391, + 0.035926215, + 0.029474363, + -0.13199449, + -0.0050542126, + -0.037856363, + -0.031812496, + -0.09550558, + -0.02784898, + 0.0024248678, + -0.00787009, + 0.058306027, + -0.031987824, + -0.060772236, + -0.023396304, + -0.00048196205, + 0.13763544, + -0.060605947, + -0.008114516, + -0.03488438, + -0.10245707, + -0.009050287, + -0.018934958, + -0.0016597, + -0.07709882, + -0.04334038, + -0.0361865, + 0.039601848, + 3.0219162e-33, + 0.005053912, + 0.00015431564, + -0.06306949, + 0.012218069, + 0.1063659, + 0.015990963, + 0.0066400464, + 0.018770201, + -0.0471047, + 0.045863718, + 0.0075768502, + 0.032443933, + 0.032128394, + -0.020898193, + 0.10698109, + 0.007532993, + 0.036073696, + 0.14622816, + 0.037774064, + -0.043192174, + -0.02174738, + -0.00935259, + -0.06981355, + 0.015573663, + 0.02109968, + 0.030653212, + 0.084332876, + -0.09286219, + 0.008142902, + 0.07806401, + -0.08162771, + 0.011892468, + 0.017087212, + 0.00398037, + -0.14143594, + 0.0040631783, + 0.046298824, + 0.008982074, + -0.005657945, + -0.055603515, + 0.0280101, + 0.05592923, + 0.062290046, + 0.08601, + -0.12158096, + 0.021957703, + -0.005100348, + 0.02980473, + -0.012115537, + 0.048780393, + 0.06410944, + -0.041410886, + 0.018104166, + -0.02820519, + -0.021403184, + 0.055625826, + -0.15481973, + 0.027487248, + 0.018844942, + 0.04328508, + 0.011498004, + 0.017362038, + -0.023222718, + 0.050983228, + 0.029452797, + 0.01254276, + -0.04543061, + -0.0022227901, + -0.056091864, + -0.078929745, + 0.07960116, + -0.020429092, + -0.07619497, + 0.011574957, + -0.035614606, + -0.04804958, + -0.06608312, + -0.057374183, + -0.040188707, + 0.07175473, + 0.03131696, + 0.0230157, + -0.023998452, + 0.023608608, + 0.09605228, + -0.068432696, + 0.014267485, + 0.088031985, + -0.03773173, + 0.029476892, + -0.028656572, + 0.102167346, + -0.06608837, + 0.022874568, + 0.042159338, + -1.4813683e-08, + 0.03084451, + 0.04388152, + -0.024583708, + -0.09819144, + 0.056637153, + 0.09295194, + -0.010354905, + -0.098406926, + 0.018038867, + -0.013122614, + 0.026393658, + 0.008811199, + -0.052838095, + -0.042186745, + 0.030139562, + 0.017119711, + -0.062443938, + -0.01874104, + -0.015988652, + -0.0183798, + -0.02648776, + 0.010297134, + -0.030490613, + -0.084531814, + -0.05403832, + 0.010054051, + 0.029555269, + 0.13556379, + 0.03345113, + 0.120973155, + 0.040772952, + 0.008201573, + -0.018219272, + -0.056843966, + -0.12900989, + 0.124168746, + -0.05343351, + -0.03815373, + 0.030203871, + 0.019795638, + 0.04782522, + 0.029415023, + 0.0015728051, + 0.06116331, + -0.05558004, + -0.03020239, + 0.10799468, + -0.07070225, + -0.08214713, + 0.004543827, + -0.015572335, + -0.013686601, + -0.002344595, + 0.026393257, + -0.0158105, + 0.01623434, + -0.040631585, + -0.013981304, + -0.08312135, + 0.03198948, + 0.049973406, + -0.0038489774, + 0.07033907, + 0.004923655 ] ] } diff --git a/tests/integration/recordings/responses/e29300494763.json b/tests/integration/recordings/responses/e29300494763.json index 9b159fb50..03bd0afdd 100644 --- a/tests/integration/recordings/responses/e29300494763.json +++ b/tests/integration/recordings/responses/e29300494763.json @@ -1,7 +1,7 @@ { "request": { "method": "POST", - "url": "http://localhost:11434/v1/v1/completions", + "url": "http://0.0.0.0:11434/v1/v1/completions", "headers": {}, "body": { "model": "llama3.2:3b-instruct-fp16", @@ -43,7 +43,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-919", + "id": "chatcmpl-92", "choices": [ { "delta": { @@ -54,7 +54,7 @@ "tool_calls": [ { "index": 0, - "id": "call_fsjgzlg2", + "id": "call_ygt6yxij", "function": { "arguments": "{\"city\":\"Tokyo\"}", "name": "get_weather" @@ -68,7 +68,7 @@ "logprobs": null } ], - "created": 1753819501, + "created": 1753984138, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, @@ -79,7 +79,7 @@ { "__type__": "openai.types.chat.chat_completion_chunk.ChatCompletionChunk", "__data__": { - "id": "chatcmpl-919", + "id": "chatcmpl-92", "choices": [ { "delta": { @@ -94,7 +94,7 @@ "logprobs": null } ], - "created": 1753819502, + "created": 1753984138, "model": "llama3.2:3b-instruct-fp16", "object": "chat.completion.chunk", "service_tier": null, diff --git a/tests/integration/recordings/responses/e96152610712.json b/tests/integration/recordings/responses/e96152610712.json index fafaa751b..176499280 100644 --- a/tests/integration/recordings/responses/e96152610712.json +++ b/tests/integration/recordings/responses/e96152610712.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:54.687705Z", + "created_at": "2025-07-31T17:58:56.109631764Z", "done": true, "done_reason": "stop", - "total_duration": 186928041, - "load_duration": 92885041, + "total_duration": 1062584022, + "load_duration": 43445939, "prompt_eval_count": 207, - "prompt_eval_duration": 48226375, + "prompt_eval_duration": 845723813, "eval_count": 5, - "eval_duration": 45085542, + "eval_duration": 172894900, "response": "unsafe\nS2", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/ed9e9b34008d.json b/tests/integration/recordings/responses/ed9e9b34008d.json index 36dda5ef0..ae46f481a 100644 --- a/tests/integration/recordings/responses/ed9e9b34008d.json +++ b/tests/integration/recordings/responses/ed9e9b34008d.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T21:56:19.673198Z", + "created_at": "2025-07-31T17:50:48.719062652Z", "done": true, "done_reason": "stop", - "total_duration": 9704897750, - "load_duration": 74034167, + "total_duration": 42572007410, + "load_duration": 42791399, "prompt_eval_count": 26, - "prompt_eval_duration": 64475958, + "prompt_eval_duration": 1301967184, "eval_count": 232, - "eval_duration": 9565656334, + "eval_duration": 41226696354, "response": "The largest planet in our solar system is Jupiter. It is a gas giant, meaning it is primarily composed of hydrogen and helium gases. Jupiter has a diameter of approximately 142,984 kilometers (88,846 miles), which is more than 11 times the diameter of Earth.\n\nJupiter is not only the largest planet in terms of size, but also the most massive planet in our solar system, with a mass that is more than 318 times that of Earth. It has a thick atmosphere and a strong magnetic field, and is known for its distinctive banded appearance, which is caused by strong winds in the upper atmosphere.\n\nJupiter's massive size and gravitational pull have a significant impact on the surrounding space, including the orbits of nearby planets and asteroids. Its moons are also notable, with four large ones: Io, Europa, Ganymede, and Callisto, which are known as the Galilean moons due to their discovery by Galileo Galilei in 1610.\n\nJupiter is a fascinating planet that continues to be studied by astronomers and space agencies around the world, offering insights into the formation and evolution of our solar system.", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/eee47930e3ae.json b/tests/integration/recordings/responses/eee47930e3ae.json index d7580097e..ca7f88c74 100644 --- a/tests/integration/recordings/responses/eee47930e3ae.json +++ b/tests/integration/recordings/responses/eee47930e3ae.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:27.950916Z", + "created_at": "2025-07-31T17:46:28.464372035Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:27.997401Z", + "created_at": "2025-07-31T17:46:28.646456238Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.050139Z", + "created_at": "2025-07-31T17:46:28.825659141Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.096439Z", + "created_at": "2025-07-31T17:46:29.007157531Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.142877Z", + "created_at": "2025-07-31T17:46:29.188522934Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.190578Z", + "created_at": "2025-07-31T17:46:29.370620046Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.237602Z", + "created_at": "2025-07-31T17:46:29.556327466Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.288616Z", + "created_at": "2025-07-31T17:46:29.738894573Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.33735Z", + "created_at": "2025-07-31T17:46:29.925972025Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.383775Z", + "created_at": "2025-07-31T17:46:30.108143385Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.431402Z", + "created_at": "2025-07-31T17:46:30.289255107Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.47837Z", + "created_at": "2025-07-31T17:46:30.467730564Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.528806Z", + "created_at": "2025-07-31T17:46:30.648982445Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.576646Z", + "created_at": "2025-07-31T17:46:30.830340138Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.626251Z", + "created_at": "2025-07-31T17:46:31.010188241Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.67358Z", + "created_at": "2025-07-31T17:46:31.193929127Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.722215Z", + "created_at": "2025-07-31T17:46:31.376791155Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.770512Z", + "created_at": "2025-07-31T17:46:31.556745128Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,15 +346,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:28.815207Z", + "created_at": "2025-07-31T17:46:31.736710894Z", "done": true, "done_reason": "stop", - "total_duration": 1022732667, - "load_duration": 72059667, + "total_duration": 5753329414, + "load_duration": 42032646, "prompt_eval_count": 371, - "prompt_eval_duration": 83482875, + "prompt_eval_duration": 2437350540, "eval_count": 19, - "eval_duration": 866202458, + "eval_duration": 3273363452, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/eee6a163b837.json b/tests/integration/recordings/responses/eee6a163b837.json index 3100da886..d99322948 100644 --- a/tests/integration/recordings/responses/eee6a163b837.json +++ b/tests/integration/recordings/responses/eee6a163b837.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 37544375, - "load_duration": 27636125, + "total_duration": 19830334, + "load_duration": 7418054, "prompt_eval_count": 9, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.060630284, - 0.06372823, - -0.059383437, - -0.010313639, - -0.11985778, - 0.033409074, - 0.056847293, - -0.0064553, - 0.029896382, - -0.05037607, - 0.015193001, - -0.0634204, - 0.015119892, - -0.08354324, - 0.0092577925, - 0.044272587, - -0.024397198, - -0.05100177, - -0.028086444, - -0.07390362, - 0.07088186, - 0.08101153, - 0.006050408, - -0.043090094, - 0.010714593, - -0.01581376, - 0.0351736, - 0.06538307, - 0.03639655, - -0.05625738, - 0.073681176, - 0.04730274, - 0.067169026, - -0.01207242, - -0.018193275, - 0.0042488067, - 0.029168725, - 0.0067459582, - 0.037927665, - 0.0024767139, - 0.014044963, - 0.022671249, - -0.090508185, - 0.041952047, - -0.07933115, - 0.031992197, - -0.038355146, - 0.037013844, - -0.0036946274, - -0.016986867, - 0.03696087, - -0.07697335, - -0.020080294, - 0.07733012, - 0.04521822, - -0.007816803, - -0.0058926586, - 0.009962128, - 0.033492323, - 0.09000152, - 0.016161384, - 0.036999356, - -0.039193578, - -0.010969346, - 0.023929566, - -0.03698458, - -0.008227196, - 0.018780757, - -0.0006967325, - -0.062018193, - -0.030388007, - -0.037649162, - -0.04654288, - 0.038450293, - -0.010377299, - -0.032971557, - 0.013547814, - -0.059036925, - 0.0630603, - 0.0159564, - -0.04845087, - -0.069917254, - -0.022502322, - 0.04408022, - 0.03618941, - 0.060470726, - -0.04313285, - 0.028797466, - 0.0062393937, - 0.01027349, - -0.078714885, - -0.091531575, - 0.04391341, - 0.013202597, - -0.0037814155, - 0.0102497, - 0.020225797, - 0.05634384, - -0.09700619, - 0.06577961, - 0.047118917, - 0.01876648, - 0.12445029, - -0.06447121, - -0.012632697, - 0.016056264, - 0.08604982, - 0.024878234, - 0.10627678, - -0.043176394, - -0.046339765, - -0.03149599, - -0.001784808, - -0.023469802, - -0.05079461, - 0.0046657966, - 0.043237828, - 0.057146583, - -0.065833576, - 0.032975562, - -0.028763266, - 0.037831448, - 0.00017829033, - 0.043322463, - -0.13265091, - 0.0263673, - -0.04247752, - -3.3340873e-33, - -0.0022191573, - 0.050657377, - 0.028066125, - -0.033898965, - -0.0045730886, - -0.034653578, - -0.08628417, - 0.043108672, - 0.01022734, - 0.044009056, - -0.03020062, - -0.0936044, - -0.06522928, - -0.059762992, - 0.037560984, - -0.025942331, - -0.06655938, - 0.0043691625, - 0.018846871, - -0.035582166, - 0.02240012, - 0.08943218, - 0.033568345, - -0.11379316, - 0.03822112, - -0.044403847, - 0.10261262, - -0.07330182, - 0.089390896, - 0.056668896, - -0.009407597, - -0.0646505, - 0.016652016, - 0.007326742, - 0.005187682, - 0.0051324354, - -0.013595071, - -0.04918112, - -0.06672084, - 0.010838405, - 0.04638185, - -0.11490209, - -0.055054087, - 0.040443793, - -0.032746885, - 0.03498173, - -0.023567867, - -0.012213799, - 0.048050664, - 0.01159698, - 0.007860181, - 0.03801084, - -0.027765153, - 0.003296162, - -0.0033349432, - 0.006083357, - 0.03200884, - 0.048306234, - 0.013800832, - 0.036165927, - -0.022672432, - 0.09197581, - 0.029846204, - 0.08112345, - -0.08677228, - -0.028041098, - 0.0556574, - -0.030357547, - -0.016538681, - 0.031826265, - -0.07586954, - -0.009915978, - 0.028101236, - 0.002207158, - -0.10496646, - -0.023673821, - -0.024204832, - -0.0003132271, - 0.0016462951, - -0.037603874, - 0.025533162, - -0.05221861, - 0.021656586, - 0.099111386, - -0.06896361, - -0.018568028, - 0.07245527, - -0.10582686, - -0.08505038, - -0.029969748, - -0.015717981, - -0.056855034, - -0.02698479, - -0.06410572, - 0.0057078917, - 1.2902391e-33, - 0.05490771, - -0.036417797, - -0.0023541928, - -0.03591478, - 0.106852315, - -0.04931468, - 0.037884213, - 0.050633065, - -0.083874516, - -0.018756155, - 0.0036251817, - 0.028974183, - -0.0027879397, - -0.036439158, - 0.11148004, - 0.051007163, - 0.040258586, - 0.09245398, - -0.01367112, - -0.070999645, - -0.043213032, - -0.060117763, - -0.03019449, - 0.009107182, - -0.044254936, - 0.04843456, - 0.117205575, - -0.009833911, - 0.0023962231, - 0.09339494, - -0.059902366, - 0.0101377955, - -0.03777244, - -0.04344207, - -0.14677393, - -0.022666233, - -0.008934328, - -0.02157697, - -0.021902358, - -0.06611372, - 0.016243221, - 0.062620856, - 0.01056146, - 0.04721975, - -0.087221384, - 0.009420561, - -0.017691165, - -0.03847053, - 0.010398396, - 0.022942957, - 0.099518456, - -0.021421565, - 0.0016765085, - -0.039359514, - 0.01641369, - 0.039669517, - -0.119695365, - 0.009885617, - 0.003855461, - 0.018273395, - -0.0454586, - 0.0020496584, - 0.024263415, - 0.016978405, - 0.06884217, - -0.027432522, - -0.01813802, - 0.053840507, - -0.028815664, - -0.045221787, - 0.11472852, - 0.019796453, - -0.05785514, - 0.016556906, - -0.07362942, - 0.04025756, - -0.01510899, - 0.0067040483, - -0.049666926, - 0.045941774, - 0.077951804, - -0.042951427, - 0.021852365, - 0.063826546, - 0.08110754, - -0.070652775, - -0.03245094, - 0.09259784, - -0.020451743, - 0.0701599, - -0.020740295, - 0.09339449, - -0.051164806, - 0.039440546, - 0.02560772, - -1.6767814e-08, - 0.001529873, - 0.0080792755, - -0.017666567, - -0.034070052, - 0.06805411, - 0.07387949, - -0.07592055, - -0.11369049, - -0.022008128, - 0.009088418, - 0.03108134, - -0.0056734695, - -0.0462051, - 0.0037219985, - 0.013269294, - -0.03213892, - -0.05557376, - -0.010602884, - 0.006751397, - -0.025462827, - -0.0836812, - 0.08886153, - 0.005159859, - -0.051621262, - -0.051873572, - 0.039706588, - -0.042155124, - 0.057125967, - 0.088910565, - 0.049736783, - 0.04144574, - 0.094677895, - -0.037107926, - -0.06845684, - -0.061673928, - 0.09891817, - -0.05952751, - -0.0331722, - -0.026014913, - 0.077612035, - 0.056150436, - 0.010709955, - 0.018974187, - 0.056079865, - -0.041700333, - -0.02731697, - 0.10184176, - -0.036189064, - -0.029914921, - -0.043333948, - 0.043660097, - 0.018800316, - -0.0042763646, - 0.055898346, - -0.0034344571, - 0.060258396, - -0.1337251, - 0.008184424, - -0.031549457, - 0.022398692, - 0.037932154, - 0.024529235, - 0.068037644, - 0.07021777 + -0.060639773, + 0.06372565, + -0.05936291, + -0.010300041, + -0.11982569, + 0.033425555, + 0.056833234, + -0.0064585046, + 0.029884538, + -0.050377, + 0.015235641, + -0.06343532, + 0.015153007, + -0.08355496, + 0.009256003, + 0.044258773, + -0.024394983, + -0.050998896, + -0.028089164, + -0.073907696, + 0.0708678, + 0.081041634, + 0.0060154344, + -0.043091357, + 0.010724255, + -0.0157963, + 0.03515573, + 0.06535516, + 0.036405835, + -0.056207117, + 0.073691264, + 0.047304515, + 0.06716326, + -0.012080835, + -0.018201472, + 0.004228335, + 0.029158425, + 0.0067739366, + 0.037928075, + 0.0024743075, + 0.014024643, + 0.022656543, + -0.090546414, + 0.04194943, + -0.0793169, + 0.03200897, + -0.03836494, + 0.036980182, + -0.0037388564, + -0.016997939, + 0.03697645, + -0.07700678, + -0.020121705, + 0.077319205, + 0.04524197, + -0.007828367, + -0.005898434, + 0.009954127, + 0.033502743, + 0.08998491, + 0.01617543, + 0.036984075, + -0.039189536, + -0.010982272, + 0.023962036, + -0.036975376, + -0.008243782, + 0.018792296, + -0.0006868037, + -0.06198398, + -0.030387625, + -0.03764361, + -0.0465636, + 0.038446005, + -0.010364095, + -0.03298246, + 0.013548113, + -0.059034206, + 0.06306018, + 0.015928335, + -0.048439246, + -0.06992423, + -0.02249995, + 0.044059318, + 0.036188874, + 0.0604999, + -0.043173794, + 0.02878105, + 0.00625082, + 0.010277735, + -0.07871832, + -0.09155406, + 0.043908454, + 0.013187137, + -0.0037733258, + 0.010227577, + 0.020180272, + 0.056316458, + -0.097017966, + 0.06576817, + 0.04714018, + 0.018830564, + 0.124457024, + -0.064480856, + -0.012605383, + 0.016036145, + 0.08604121, + 0.024867544, + 0.10626747, + -0.04315638, + -0.046307754, + -0.031482875, + -0.0017896753, + -0.023491126, + -0.05078794, + 0.0046478175, + 0.043237753, + 0.05715731, + -0.06581743, + 0.032969296, + -0.028760076, + 0.037811063, + 0.00016727838, + 0.043302964, + -0.13265245, + 0.0263716, + -0.042466283, + -3.3353205e-33, + -0.002227074, + 0.05063068, + 0.028064998, + -0.033884156, + -0.0045258533, + -0.03464052, + -0.086286224, + 0.04311723, + 0.01025313, + 0.044037264, + -0.03019185, + -0.09359287, + -0.065240555, + -0.059729856, + 0.037555248, + -0.025923185, + -0.06656174, + 0.0043645045, + 0.018837698, + -0.035559133, + 0.022410586, + 0.089450255, + 0.033591606, + -0.11377698, + 0.03821906, + -0.044418316, + 0.10258541, + -0.07330748, + 0.089376666, + 0.056672454, + -0.009439043, + -0.064637296, + 0.016651757, + 0.007316741, + 0.0052164183, + 0.0051133917, + -0.013578286, + -0.04917494, + -0.06671976, + 0.010861998, + 0.04635771, + -0.114912644, + -0.055052113, + 0.040456027, + -0.032764934, + 0.03497564, + -0.023565749, + -0.0122220665, + 0.048061397, + 0.011574431, + 0.0078836065, + 0.038066708, + -0.02776296, + 0.0033052622, + -0.0033062513, + 0.0060893223, + 0.032002907, + 0.048337206, + 0.013765614, + 0.0361784, + -0.022666454, + 0.09200785, + 0.029846026, + 0.08113161, + -0.08676324, + -0.028000059, + 0.05564783, + -0.030363169, + -0.01651962, + 0.031817682, + -0.07585742, + -0.009924711, + 0.028093029, + 0.0022190511, + -0.10494398, + -0.023684397, + -0.02419872, + -0.00029041493, + 0.0016061965, + -0.037576895, + 0.025535172, + -0.052210968, + 0.021654254, + 0.099135235, + -0.06895632, + -0.018578324, + 0.0724588, + -0.105865955, + -0.08502584, + -0.030003337, + -0.015741726, + -0.05684276, + -0.027024021, + -0.06408883, + 0.005700278, + 1.2909128e-33, + 0.054930277, + -0.036410004, + -0.0023171217, + -0.035894036, + 0.1068399, + -0.049308524, + 0.037888233, + 0.050643876, + -0.08384806, + -0.018762833, + 0.0036076272, + 0.028958267, + -0.0028438307, + -0.036437068, + 0.11147226, + 0.051020138, + 0.04023225, + 0.09245051, + -0.01367069, + -0.07095874, + -0.043207154, + -0.060137726, + -0.030198809, + 0.009124354, + -0.04424328, + 0.04844442, + 0.117211945, + -0.009819735, + 0.002429452, + 0.09333068, + -0.05990318, + 0.010140399, + -0.03776785, + -0.043452136, + -0.14679956, + -0.022661027, + -0.008955439, + -0.021544594, + -0.021905722, + -0.066112846, + 0.016239611, + 0.062623896, + 0.010557838, + 0.047209036, + -0.087235674, + 0.0094077485, + -0.01769091, + -0.038465302, + 0.010402301, + 0.022951292, + 0.09949292, + -0.021435224, + 0.0016363884, + -0.039346013, + 0.016425362, + 0.039651647, + -0.11970253, + 0.009909306, + 0.003858687, + 0.0182787, + -0.04543836, + 0.0020555719, + 0.024242649, + 0.016967652, + 0.06883535, + -0.027423857, + -0.0181242, + 0.053814657, + -0.02883445, + -0.045210328, + 0.11474227, + 0.019766666, + -0.057896398, + 0.016537277, + -0.0736607, + 0.040268008, + -0.015094288, + 0.006702388, + -0.049662724, + 0.045974977, + 0.077959426, + -0.04295526, + 0.02186354, + 0.063874446, + 0.08110985, + -0.07061161, + -0.032482613, + 0.09260675, + -0.020442914, + 0.07014854, + -0.020733764, + 0.093375094, + -0.05119609, + 0.039442588, + 0.025606012, + -1.6766675e-08, + 0.0015268949, + 0.008090479, + -0.017678792, + -0.03410393, + 0.068066135, + 0.07387084, + -0.07586656, + -0.113715105, + -0.022030246, + 0.0091101825, + 0.031090891, + -0.005711195, + -0.04621522, + 0.003679675, + 0.013247459, + -0.032131832, + -0.05559878, + -0.010590206, + 0.006735521, + -0.025462449, + -0.08367791, + 0.08888197, + 0.0051492383, + -0.051671468, + -0.051902205, + 0.039699014, + -0.042151738, + 0.05715088, + 0.088929184, + 0.049737025, + 0.041459583, + 0.0946667, + -0.037086494, + -0.06848715, + -0.06164879, + 0.09891768, + -0.05950153, + -0.03321193, + -0.025980823, + 0.07762438, + 0.05616985, + 0.010686045, + 0.01899881, + 0.056079634, + -0.04168257, + -0.027310291, + 0.10182688, + -0.036167838, + -0.029922716, + -0.04332533, + 0.0436739, + 0.01878726, + -0.0042941514, + 0.055926275, + -0.0034490179, + 0.060258176, + -0.13373649, + 0.008205703, + -0.03154335, + 0.02240309, + 0.03797733, + 0.02454498, + 0.06804169, + 0.07020173 ] ] } diff --git a/tests/integration/recordings/responses/ef59cbff54d0.json b/tests/integration/recordings/responses/ef59cbff54d0.json index 5cc8cce14..bf9c40660 100644 --- a/tests/integration/recordings/responses/ef59cbff54d0.json +++ b/tests/integration/recordings/responses/ef59cbff54d0.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:32:56.400171Z", + "created_at": "2025-07-31T17:59:04.621241272Z", "done": true, "done_reason": "stop", - "total_duration": 194586042, - "load_duration": 117270208, + "total_duration": 1020488139, + "load_duration": 48171516, "prompt_eval_count": 213, - "prompt_eval_duration": 63001709, + "prompt_eval_duration": 924613213, "eval_count": 2, - "eval_duration": 11829541, + "eval_duration": 47179414, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/f477c2fe1332.json b/tests/integration/recordings/responses/f477c2fe1332.json index f8ea8b315..36db1f056 100644 --- a/tests/integration/recordings/responses/f477c2fe1332.json +++ b/tests/integration/recordings/responses/f477c2fe1332.json @@ -22,7 +22,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.066885Z", + "created_at": "2025-07-31T17:48:02.228613069Z", "done": false, "done_reason": null, "total_duration": null, @@ -40,7 +40,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.111127Z", + "created_at": "2025-07-31T17:48:02.416516088Z", "done": false, "done_reason": null, "total_duration": null, @@ -58,7 +58,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.154415Z", + "created_at": "2025-07-31T17:48:02.601455913Z", "done": false, "done_reason": null, "total_duration": null, @@ -76,7 +76,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.199308Z", + "created_at": "2025-07-31T17:48:02.789779738Z", "done": false, "done_reason": null, "total_duration": null, @@ -94,7 +94,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.242681Z", + "created_at": "2025-07-31T17:48:02.979797092Z", "done": false, "done_reason": null, "total_duration": null, @@ -112,7 +112,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.285299Z", + "created_at": "2025-07-31T17:48:03.171273909Z", "done": false, "done_reason": null, "total_duration": null, @@ -130,7 +130,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.329456Z", + "created_at": "2025-07-31T17:48:03.366307208Z", "done": false, "done_reason": null, "total_duration": null, @@ -148,7 +148,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.37332Z", + "created_at": "2025-07-31T17:48:03.553860717Z", "done": false, "done_reason": null, "total_duration": null, @@ -166,7 +166,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.417505Z", + "created_at": "2025-07-31T17:48:03.741050708Z", "done": false, "done_reason": null, "total_duration": null, @@ -184,7 +184,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.459524Z", + "created_at": "2025-07-31T17:48:03.928660324Z", "done": false, "done_reason": null, "total_duration": null, @@ -202,7 +202,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.502376Z", + "created_at": "2025-07-31T17:48:04.114285473Z", "done": false, "done_reason": null, "total_duration": null, @@ -220,7 +220,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.545028Z", + "created_at": "2025-07-31T17:48:04.297051984Z", "done": false, "done_reason": null, "total_duration": null, @@ -238,7 +238,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.587118Z", + "created_at": "2025-07-31T17:48:04.481247453Z", "done": false, "done_reason": null, "total_duration": null, @@ -256,7 +256,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.6295Z", + "created_at": "2025-07-31T17:48:04.663502116Z", "done": false, "done_reason": null, "total_duration": null, @@ -274,7 +274,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.671588Z", + "created_at": "2025-07-31T17:48:04.847068051Z", "done": false, "done_reason": null, "total_duration": null, @@ -292,7 +292,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.712532Z", + "created_at": "2025-07-31T17:48:05.029468907Z", "done": false, "done_reason": null, "total_duration": null, @@ -310,7 +310,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.754386Z", + "created_at": "2025-07-31T17:48:05.210665493Z", "done": false, "done_reason": null, "total_duration": null, @@ -328,7 +328,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.796521Z", + "created_at": "2025-07-31T17:48:05.394445374Z", "done": false, "done_reason": null, "total_duration": null, @@ -346,7 +346,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.838452Z", + "created_at": "2025-07-31T17:48:05.577557339Z", "done": false, "done_reason": null, "total_duration": null, @@ -364,7 +364,7 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.880192Z", + "created_at": "2025-07-31T17:48:05.759740241Z", "done": false, "done_reason": null, "total_duration": null, @@ -382,15 +382,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama3.2:3b-instruct-fp16", - "created_at": "2025-07-29T20:04:45.923175Z", + "created_at": "2025-07-31T17:48:05.941978534Z", "done": true, "done_reason": "stop", - "total_duration": 1469644334, - "load_duration": 62107584, + "total_duration": 35821787819, + "load_duration": 41469279, "prompt_eval_count": 375, - "prompt_eval_duration": 546489083, + "prompt_eval_duration": 32065473198, "eval_count": 21, - "eval_duration": 860370875, + "eval_duration": 3714298846, "response": "", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/fcdef245da95.json b/tests/integration/recordings/responses/fcdef245da95.json index 832eecb2a..1aa18e60f 100644 --- a/tests/integration/recordings/responses/fcdef245da95.json +++ b/tests/integration/recordings/responses/fcdef245da95.json @@ -20,15 +20,15 @@ "__type__": "ollama._types.GenerateResponse", "__data__": { "model": "llama-guard3:1b", - "created_at": "2025-07-29T20:04:05.027857Z", + "created_at": "2025-07-31T17:43:20.195368079Z", "done": true, "done_reason": "stop", - "total_duration": 317769083, - "load_duration": 68092000, + "total_duration": 4029876004, + "load_duration": 1111568374, "prompt_eval_count": 212, - "prompt_eval_duration": 237798125, + "prompt_eval_duration": 2868613054, "eval_count": 2, - "eval_duration": 11373291, + "eval_duration": 48866061, "response": "safe", "thinking": null, "context": null diff --git a/tests/integration/recordings/responses/fe140befeba4.json b/tests/integration/recordings/responses/fe140befeba4.json index 02e06429b..428827790 100644 --- a/tests/integration/recordings/responses/fe140befeba4.json +++ b/tests/integration/recordings/responses/fe140befeba4.json @@ -20,398 +20,398 @@ "created_at": null, "done": null, "done_reason": null, - "total_duration": 46540167, - "load_duration": 29688959, + "total_duration": 17148955, + "load_duration": 6651167, "prompt_eval_count": 6, "prompt_eval_duration": null, "eval_count": null, "eval_duration": null, "embeddings": [ [ - -0.08566708, - -0.09559047, - 0.044014607, - -0.015974598, - 0.029406257, - 0.07229597, - -0.010901963, - -0.023829829, - 0.07381301, - -0.05698464, - -0.033780586, - 0.051200844, - 0.0050912783, - 0.014317088, - -0.07878143, - -0.012908666, - -0.041628323, - 0.06881713, - -0.10783476, - -0.04042705, - 0.026262026, - -0.0019893218, - -0.011008084, - -0.0019646112, - 0.004033132, - 0.08881656, - 0.014049165, - -0.018416086, - 0.032621212, - -0.034692146, - 0.07614942, - -0.014122101, - -0.024901746, - 0.03755059, - -0.10197354, - 0.054705318, - -0.022539826, - 0.024209768, - 0.011698194, - -0.008956377, - -0.050146304, - 0.0026327297, - 0.055942897, - 0.009974366, - 0.12796965, - -0.025006283, - 0.024338534, - -0.024487961, - -0.0022703854, - -0.024687177, - -0.10482094, - -0.05994297, - -0.055200897, - 0.0152664175, - 0.03496896, - 0.052624088, - -0.0006445885, - 0.06637695, - -0.031790398, - -0.007308742, - -0.0050764186, - -0.042508755, - -0.04089097, - 0.020062948, - 0.038683955, - 0.022463562, - -0.02866933, - 0.053370677, - 0.022435635, - 0.01934692, - 0.12264713, - 0.023911418, - -0.037264284, - 0.0059156846, - 0.05235448, - 0.054004095, - 0.08022169, - -0.010992806, - 0.029295033, - -0.0672064, - -0.00021147476, - -0.050584126, - -0.0095251575, - 0.04616498, - 0.078677796, - 0.01416309, - -0.033226117, - 0.0018380182, - -0.06667651, - -0.020977372, - -0.017116925, - -0.04396714, - -0.05969979, - -0.07344942, - -0.03985366, - -0.030863814, - -0.019918729, - -0.1075161, - -0.026654154, - 0.0689854, - -0.0049292273, - 0.026645623, - 0.018879393, - 0.022113768, - 0.064208575, - -0.053153764, - 0.06160797, - 0.014026719, - 0.11772326, - -0.051769163, - -0.07634968, - 0.03090975, - -0.038558383, - -0.025260162, - 0.039262023, - -0.061449137, - 0.008389126, - 0.016175874, - 0.032293033, - 0.06679397, - -0.06503257, - 0.014676881, - -0.038542666, - 0.018718671, - -0.030111106, - -0.028481327, - -0.14707623, - -3.455443e-33, - -0.048577547, - -0.024983348, - 0.071679614, - 0.035652317, - 0.07931413, - -0.07811974, - 0.023085583, - -0.047467884, - 0.08872273, - -0.0010074769, - -0.11320135, - 0.091322996, - 0.023978539, - 0.11368158, - 0.042203873, - -0.05773289, - -0.074543044, - -0.0021036167, - -0.051522236, - -0.050925426, - -0.0016557347, - 0.030671587, - 0.045119714, - -0.03974729, - -0.05871358, - -0.030611658, - 0.0017253247, - 0.009114429, - -0.013763352, - 0.023424039, - 0.0017495834, - 0.046633217, - -0.07230643, - -0.027882291, - 0.016182518, - 0.044456217, - -0.004326421, - -0.061798126, - 0.0697968, - 0.031249145, - -0.013697079, - -0.007417679, - 0.031665757, - -0.02367961, - 0.07153089, - 0.023938214, - 0.009729952, - 0.0071919435, - -0.03235391, - -0.04955071, - -0.050248373, - 0.02151118, - 0.015327139, - -0.0674203, - 0.06544387, - -0.025547959, - 0.03207046, - 0.02038825, - 0.0112230005, - 0.00019493286, - -0.023462659, - -0.004949742, - -0.014066955, - 0.0014178518, - 0.059315395, - 0.039931085, - -0.032498423, - -0.023698896, - 0.05445033, - 0.064231694, - -0.034013335, - 0.08745776, - -0.080473825, - -0.090545714, - -0.065398656, - -8.2386265e-05, - -0.021441188, - -0.0684535, - -0.029121745, - 0.034134887, - -0.07799698, - -0.05388711, - -0.035591345, - 0.044826802, - -0.040090464, - 0.07972004, - 0.026058797, - -0.08184859, - 0.0018106091, - -0.027676936, - -0.04312832, - -0.042090744, - 0.08336437, - -0.049453646, - -0.0902778, - 2.6716498e-33, - -0.091911495, - 0.02641473, - -0.07022486, - 0.075562105, - 0.03900905, - 0.027913846, - -0.05444872, - -0.036666486, - -0.048225258, - 0.07551892, - 0.046452336, - 0.025874302, - 0.052248206, - -0.00018527219, - 0.010575236, - -0.040591337, - -0.028484622, - -0.020559357, - 0.08882296, - -0.06755767, - 0.04941752, - 0.13231009, - -0.06998129, - -0.040112328, - 0.044030365, - 0.034218542, - -0.08650528, - 0.05746921, - -0.0075130556, - 0.049070083, - -0.0148686, - -0.018103259, - -0.020280316, - 0.038828347, - 0.022253176, - 0.13486238, - 0.06899369, - -0.002589861, - -0.016430879, - 0.0033818923, - 0.017275693, - 0.013614936, - 0.044220798, - 0.049155377, - -0.008259856, - -0.046575654, - -0.043921605, - 0.04156687, - -0.035468902, - 0.042837795, - 0.03131579, - 0.017961076, - -0.026213305, - -0.05458616, - -0.04259084, - -0.004110002, - 0.029035388, - 0.0010451805, - 0.09044077, - 0.014110149, - -0.068820216, - -0.07098938, - 0.020328037, - 0.00433692, - -0.046977337, - 0.016492791, - -0.028396707, - 0.104340956, - 0.002814702, - -0.08339559, - 0.037326302, - 0.058929898, - 0.0376423, - 0.09580634, - -0.12376848, - -0.054060236, - -0.014485116, - 0.0013106487, - -0.04537336, - -0.0899294, - 0.001730278, - -0.05520831, - 0.000568523, - 0.00053380145, - 0.07856981, - 0.104590714, - 0.00355283, - 0.008365939, - 0.04291482, - 0.010064388, - 0.025177509, - 0.05732803, - -0.023061136, - 0.054399785, - -0.049828697, - -1.3290186e-08, - -0.0539168, - 0.08074109, - 0.03397028, - 0.024365881, - 0.0906225, - -0.07162824, - 0.07550329, - 0.017278913, - -0.061226364, - -0.03298407, - 0.07829606, - 0.03967995, - -0.036696997, - 0.02665964, - 0.1000655, - -0.014426734, - 0.020708792, - -0.039230846, - 0.0085029, - -0.0012509917, - 0.06740856, - 0.013992665, - -0.054007422, - -0.016785627, - 0.07651403, - -0.035508703, - -0.050085396, - 0.08382383, - -0.009957674, - 0.08140875, - 0.019287178, - 0.049911316, - 0.0022236605, - -0.07807412, - 0.019454133, - 0.111560374, - -0.01269702, - -0.06466137, - -0.09346588, - -0.050038446, - -0.042178612, - 0.0599713, - 0.034831088, - -0.014957726, - 0.014484159, - -0.022619838, - 0.06916277, - -0.088544875, - 0.021478733, - 0.01378541, - -0.0075770007, - 0.027888266, - 0.015526889, - 0.0052174823, - 0.010616002, - -0.022908956, - -0.02535865, - -0.04139556, - -0.08375561, - 0.092626974, - 0.051755503, - 0.09296614, - 0.011223383, - -0.016759252 + -0.085681304, + -0.095581464, + 0.043999746, + -0.015962392, + 0.029400628, + 0.07226456, + -0.010901451, + -0.023813674, + 0.07378506, + -0.05697427, + -0.033765975, + 0.05121266, + 0.0050979555, + 0.014309963, + -0.07880483, + -0.012905849, + -0.041623827, + 0.0687718, + -0.107817836, + -0.040402025, + 0.02626439, + -0.0019839199, + -0.011018193, + -0.0019673458, + 0.0040422864, + 0.08879382, + 0.01402507, + -0.018432936, + 0.03259668, + -0.03467639, + 0.07614212, + -0.014108018, + -0.02491903, + 0.037580706, + -0.10195762, + 0.054681186, + -0.022511847, + 0.024178648, + 0.011709601, + -0.008936471, + -0.05014084, + 0.002622049, + 0.05592362, + 0.0099558495, + 0.1279459, + -0.024969373, + 0.024331603, + -0.024513736, + -0.0022768104, + -0.024678802, + -0.10482663, + -0.059968818, + -0.05517871, + 0.015272669, + 0.03499024, + 0.052664176, + -0.0006159338, + 0.066367485, + -0.031800635, + -0.0073153223, + -0.0050571254, + -0.042501386, + -0.040885996, + 0.020060783, + 0.03869819, + 0.02252177, + -0.028677218, + 0.053372957, + 0.022434458, + 0.01932899, + 0.122685425, + 0.02391069, + -0.037310783, + 0.0058996845, + 0.052371886, + 0.054006398, + 0.08022549, + -0.0109860925, + 0.029316097, + -0.06720931, + -0.00020536478, + -0.050579377, + -0.009566371, + 0.04615162, + 0.07864397, + 0.0141741, + -0.033241987, + 0.0018572828, + -0.066625245, + -0.021015463, + -0.017156467, + -0.043931913, + -0.0596939, + -0.07345105, + -0.0398444, + -0.030874873, + -0.019920446, + -0.10750585, + -0.026640767, + 0.06898602, + -0.0049453005, + 0.02661913, + 0.018901112, + 0.022115147, + 0.064215325, + -0.05311056, + 0.061604578, + 0.01401751, + 0.11773954, + -0.051754657, + -0.0763492, + 0.03089861, + -0.038562253, + -0.025287481, + 0.039286032, + -0.061460134, + 0.0083680395, + 0.016174005, + 0.032318905, + 0.066826284, + -0.06503665, + 0.014697402, + -0.038522907, + 0.018723847, + -0.030096343, + -0.028485447, + -0.14709295, + -3.4554582e-33, + -0.048594434, + -0.02497814, + 0.07171551, + 0.035668552, + 0.07933306, + -0.078095146, + 0.023089593, + -0.0474716, + 0.08870901, + -0.0010188936, + -0.11324424, + 0.09134128, + 0.023990165, + 0.11367867, + 0.04216775, + -0.0577585, + -0.07449343, + -0.0021416158, + -0.051485773, + -0.050966885, + -0.001616003, + 0.030676797, + 0.04510477, + -0.039778266, + -0.058712397, + -0.030624842, + 0.001723139, + 0.009131228, + -0.013748993, + 0.023429343, + 0.0017690851, + 0.046615023, + -0.07229696, + -0.027881058, + 0.016162753, + 0.04445625, + -0.0043517617, + -0.061766528, + 0.06977974, + 0.031259496, + -0.013689306, + -0.007429736, + 0.03165361, + -0.023703648, + 0.07151753, + 0.023939254, + 0.00974393, + 0.0071852654, + -0.032335706, + -0.049569704, + -0.05028296, + 0.021495296, + 0.015300162, + -0.06743517, + 0.06547497, + -0.025525656, + 0.03209374, + 0.020391569, + 0.011251086, + 0.00022983168, + -0.023480719, + -0.004975026, + -0.014066877, + 0.0013942543, + 0.059347607, + 0.03995627, + -0.032481793, + -0.02370644, + 0.054430693, + 0.06420778, + -0.03403204, + 0.08743523, + -0.08050841, + -0.090586886, + -0.06534727, + -9.977549e-05, + -0.021465743, + -0.068421625, + -0.029139552, + 0.034137648, + -0.078038886, + -0.05387686, + -0.035576295, + 0.04482086, + -0.040064223, + 0.07973023, + 0.025987988, + -0.08184961, + 0.0018208853, + -0.027658636, + -0.043111984, + -0.04208418, + 0.083367765, + -0.049456984, + -0.090272054, + 2.6719462e-33, + -0.0919105, + 0.026422026, + -0.0702218, + 0.07556679, + 0.039019883, + 0.027918797, + -0.054435197, + -0.03665134, + -0.048217386, + 0.07555299, + 0.04645861, + 0.025837917, + 0.052232184, + -0.00019511126, + 0.010594723, + -0.04060704, + -0.028484648, + -0.020562524, + 0.08883098, + -0.0675747, + 0.049423475, + 0.13226803, + -0.06998591, + -0.040101074, + 0.04402345, + 0.034215808, + -0.086508036, + 0.05748614, + -0.007484621, + 0.04904636, + -0.014905731, + -0.018129041, + -0.020285906, + 0.03886023, + 0.022301883, + 0.13486, + 0.068968445, + -0.002601102, + -0.016437862, + 0.0033667886, + 0.017248135, + 0.013624591, + 0.044167783, + 0.049155388, + -0.008251536, + -0.046576213, + -0.043907613, + 0.041560307, + -0.035454296, + 0.042816173, + 0.03130763, + 0.01795219, + -0.026223853, + -0.054602407, + -0.042619392, + -0.0041024024, + 0.02905227, + 0.0010594176, + 0.09045834, + 0.01407639, + -0.06884184, + -0.071025275, + 0.02034249, + 0.004330081, + -0.046988014, + 0.016498951, + -0.028359761, + 0.104389384, + 0.0028101353, + -0.08335455, + 0.037342783, + 0.058927905, + 0.037627388, + 0.09579175, + -0.12379578, + -0.05406347, + -0.014487729, + 0.00132331, + -0.04533123, + -0.08991201, + 0.0017578854, + -0.05520286, + 0.0005905125, + 0.00057696534, + 0.07861307, + 0.104590856, + 0.0035720735, + 0.008376108, + 0.04295002, + 0.010083029, + 0.025200473, + 0.05733988, + -0.023040904, + 0.054405548, + -0.049814496, + -1.32881794e-08, + -0.053953137, + 0.080773644, + 0.03396369, + 0.024386076, + 0.09063602, + -0.071622826, + 0.075512506, + 0.017275201, + -0.061214693, + -0.032958332, + 0.0783013, + 0.039668255, + -0.03671256, + 0.02667037, + 0.100062974, + -0.014423388, + 0.020694766, + -0.03923435, + 0.008530012, + -0.0012191309, + 0.06742351, + 0.013990705, + -0.053969115, + -0.016794305, + 0.07650657, + -0.03551458, + -0.050098673, + 0.08379899, + -0.009979221, + 0.081417635, + 0.019264791, + 0.049911313, + 0.0022199217, + -0.07809246, + 0.019478371, + 0.11155656, + -0.0127101755, + -0.0646376, + -0.09345448, + -0.050051365, + -0.04220205, + 0.059940416, + 0.034817107, + -0.014977587, + 0.0144794, + -0.022621814, + 0.06914498, + -0.088525705, + 0.021482527, + 0.013780395, + -0.007574969, + 0.027911302, + 0.015498402, + 0.005222799, + 0.010627239, + -0.02296286, + -0.02536398, + -0.041442133, + -0.083752654, + 0.09260617, + 0.051770985, + 0.09294301, + 0.011217839, + -0.01673822 ] ] } diff --git a/tests/integration/recordings/test.txt b/tests/integration/recordings/test.txt new file mode 100644 index 000000000..a0b97cc4f --- /dev/null +++ b/tests/integration/recordings/test.txt @@ -0,0 +1 @@ +Testing testing diff --git a/tests/integration/recordings/vision/index.sqlite b/tests/integration/recordings/vision/index.sqlite new file mode 100644 index 0000000000000000000000000000000000000000..a2df92cfbe913ed93e0316b8598c48e91699b8c2 GIT binary patch literal 12288 zcmeI$O=}ZD7y#f+{fd?B77=>z;P$3q+xecIoU9T=tY5VUJxG|@nXxPRGP@~WD}oe% zgC~E6M}Lj~z@w9B$%3)ciWhkYmdtMEo%hYNxy<|Y$#zk>)DOpfU8OWyi5iXQR+>gp zm`2f=a2<_Cn4CSj37^yZKaG`WWB;&~e2bRi!zeyXzWgZwz6Svi009sH0T2KI5C8!X z009vA&jsGf#)Z|jwZ=PDY14CKmk-Cb7`!a!v!y#bovqzYy1R9IyOYkJr`MO$1!L@9 zPh44b_jI|J?slH-rjK{-J=ofLp5E^~zxkVxax@&2uIr1Qo5d_%=LU8(EC$svOZr3W zddGZU75z}0?vIXHQ)e6_S*83>_m|RIdwBE;;r%z#k00@8p2!H?xfB*=900@8p z2>c^~gJ#@3Xl-7j^rH9D8iWxsB*Q%OK4U~uZ3IPvAfmb9D&quOEhHz@62q9p3VUyv z$StbJZ@empgH1guHuF7Q^?JIm=>}QfFUn##Xk%=`hI}2#b&6A*wGnGGe$Y&QmeTZ? zD`K6K*l}(oQQYK&D~mmrOd_teMoburoy;Ys#0qOH_0kJUGEMV({7FmUZ6YVlStTvn zq?wd1>}PU8oXr#>#0%`S)soOmDZy}#F>;|Y2mTpIv2l)lFeiie~mwdDGU$*0T2KI5C8!X009sH0T2KI5IC7Y=j=hN zQSV2brk@f2*yS%<5y@ Date: Thu, 31 Jul 2025 11:54:51 -0700 Subject: [PATCH 11/19] fix(ci): ensure workflow runs when manually run or scheduled --- .github/workflows/integration-tests.yml | 3 +++ .github/workflows/integration-vision-tests.yml | 3 +++ 2 files changed, 6 insertions(+) diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index d80fbc525..df9a83596 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -53,6 +53,9 @@ concurrency: jobs: discover-tests: if: | + github.event_name == "push" || + github.event_name == "schedule" || + github.event_name == "workflow_dispatch" || github.event.action == 'opened' || github.event.action == 'synchronize' || (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) diff --git a/.github/workflows/integration-vision-tests.yml b/.github/workflows/integration-vision-tests.yml index 38da333c2..c24b33a75 100644 --- a/.github/workflows/integration-vision-tests.yml +++ b/.github/workflows/integration-vision-tests.yml @@ -45,6 +45,9 @@ concurrency: jobs: discover-tests: if: | + github.event_name == "push" || + github.event_name == "schedule" || + github.event_name == "workflow_dispatch" || github.event.action == 'opened' || github.event.action == 'synchronize' || (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) From 752fd3b1c182eb17de24401c2db192ba40ec8bc1 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 11:56:25 -0700 Subject: [PATCH 12/19] fix(ci): use single quotes please --- .github/workflows/integration-tests.yml | 6 +++--- .github/workflows/integration-vision-tests.yml | 6 +++--- 2 files changed, 6 insertions(+), 6 deletions(-) diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index df9a83596..e64d71f98 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -53,9 +53,9 @@ concurrency: jobs: discover-tests: if: | - github.event_name == "push" || - github.event_name == "schedule" || - github.event_name == "workflow_dispatch" || + github.event_name == 'push' || + github.event_name == 'schedule' || + github.event_name == 'workflow_dispatch' || github.event.action == 'opened' || github.event.action == 'synchronize' || (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) diff --git a/.github/workflows/integration-vision-tests.yml b/.github/workflows/integration-vision-tests.yml index c24b33a75..28f11fe94 100644 --- a/.github/workflows/integration-vision-tests.yml +++ b/.github/workflows/integration-vision-tests.yml @@ -45,9 +45,9 @@ concurrency: jobs: discover-tests: if: | - github.event_name == "push" || - github.event_name == "schedule" || - github.event_name == "workflow_dispatch" || + github.event_name == 'push' || + github.event_name == 'schedule' || + github.event_name == 'workflow_dispatch' || github.event.action == 'opened' || github.event.action == 'synchronize' || (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) From d1b300ead9156c1e1e09135cf82d7ef7439eb5ee Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 12:31:40 -0700 Subject: [PATCH 13/19] fix(ci, nvidia): do not use module level pytest skip for now --- tests/integration/providers/nvidia/__init__.py | 11 +++++++++++ tests/integration/providers/nvidia/conftest.py | 14 -------------- .../integration/providers/nvidia/test_datastore.py | 3 +++ 3 files changed, 14 insertions(+), 14 deletions(-) delete mode 100644 tests/integration/providers/nvidia/conftest.py diff --git a/tests/integration/providers/nvidia/__init__.py b/tests/integration/providers/nvidia/__init__.py index 756f351d8..ec98e3b7a 100644 --- a/tests/integration/providers/nvidia/__init__.py +++ b/tests/integration/providers/nvidia/__init__.py @@ -3,3 +3,14 @@ # # This source code is licensed under the terms described in the LICENSE file in # the root directory of this source tree. + +import os + +import pytest + +# Reusable skip decorator for NVIDIA tests in GitHub Actions +# Adding this in conftest.py as a module level skip statement causes pytest to error +# out in certain cases. +skip_in_github_actions = pytest.mark.skipif( + os.environ.get("GITHUB_ACTIONS") == "true", reason="Skipping NVIDIA tests in GitHub Actions environment" +) diff --git a/tests/integration/providers/nvidia/conftest.py b/tests/integration/providers/nvidia/conftest.py deleted file mode 100644 index 8beb113b0..000000000 --- a/tests/integration/providers/nvidia/conftest.py +++ /dev/null @@ -1,14 +0,0 @@ -# Copyright (c) Meta Platforms, Inc. and affiliates. -# All rights reserved. -# -# This source code is licensed under the terms described in the LICENSE file in -# the root directory of this source tree. - -import os - -import pytest - -# Skip all tests in this directory when running in GitHub Actions -in_github_actions = os.environ.get("GITHUB_ACTIONS") == "true" -if in_github_actions: - pytest.skip("Skipping NVIDIA tests in GitHub Actions environment", allow_module_level=True) diff --git a/tests/integration/providers/nvidia/test_datastore.py b/tests/integration/providers/nvidia/test_datastore.py index 5f96dee9f..5cddd7781 100644 --- a/tests/integration/providers/nvidia/test_datastore.py +++ b/tests/integration/providers/nvidia/test_datastore.py @@ -7,12 +7,15 @@ import pytest +from . import skip_in_github_actions + # How to run this test: # # LLAMA_STACK_CONFIG="nvidia" pytest -v tests/integration/providers/nvidia/test_datastore.py # nvidia provider only +@skip_in_github_actions @pytest.mark.parametrize( "provider_id", [ From 18576349ca2983c698919560d185a4781dfca956 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 13:11:04 -0700 Subject: [PATCH 14/19] fix(ci): simplified concurrency and job eligibility criteria --- .../actions/run-and-record-tests/action.yml | 2 +- .github/workflows/integration-tests.yml | 20 ++++++++----------- .../workflows/integration-vision-tests.yml | 20 ++++++++----------- 3 files changed, 17 insertions(+), 25 deletions(-) diff --git a/.github/actions/run-and-record-tests/action.yml b/.github/actions/run-and-record-tests/action.yml index 469049b6a..333b6b99f 100644 --- a/.github/actions/run-and-record-tests/action.yml +++ b/.github/actions/run-and-record-tests/action.yml @@ -192,7 +192,7 @@ runs: if: ${{ always() }} uses: actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 # v4.6.2 with: - name: ${{ inputs.inference-mode }}-logs-${{ github.run_id }}-${{ github.run_attempt || '' }}-${{ inputs.provider }} + name: ${{ inputs.inference-mode }}-logs-${{ github.run_id }}-${{ github.run_attempt || '' }}-${{ inputs.provider }}-${{ inputs.run-vision-tests }}-${{ inputs.stack-config }} path: | *.log retention-days: 1 diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index e64d71f98..12e3af862 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -38,27 +38,23 @@ on: concurrency: # This creates three concurrency groups: - # ${{ github.workflow }}-${{ github.ref }}-rerecord (for valid triggers with re-record-tests label) - # ${{ github.workflow }}-${{ github.ref }}-replay (for valid triggers without re-record-tests label) - # ${{ github.workflow }}-${{ github.ref }}-no-run (for invalid triggers that will be skipped) + # ${{ github.workflow }}-${{ github.ref }}-rerecord (for labeled events with re-record-tests label) + # ${{ github.workflow }}-${{ github.ref }}-replay (for all non-labeled events) + # ${{ github.workflow }}-${{ github.ref }}-no-run (for labeled events without re-record-tests label) # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. group: >- ${{ github.workflow }}-${{ github.ref }}-${{ - ((github.event.action == 'opened' || github.event.action == 'synchronize') && 'replay') || - ((github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) && 'rerecord' || - 'no-run') + github.event.action == 'labeled' && ( + contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'no-run' + ) || 'replay' }} cancel-in-progress: true jobs: discover-tests: if: | - github.event_name == 'push' || - github.event_name == 'schedule' || - github.event_name == 'workflow_dispatch' || - github.event.action == 'opened' || - github.event.action == 'synchronize' || - (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) + github.event.action != 'labeled' || + contains(github.event.pull_request.labels.*.name, 're-record-tests') runs-on: ubuntu-latest outputs: test-types: ${{ steps.generate-test-types.outputs.test-types }} diff --git a/.github/workflows/integration-vision-tests.yml b/.github/workflows/integration-vision-tests.yml index 28f11fe94..cad945045 100644 --- a/.github/workflows/integration-vision-tests.yml +++ b/.github/workflows/integration-vision-tests.yml @@ -30,27 +30,23 @@ on: concurrency: # This creates three concurrency groups: - # ${{ github.workflow }}-${{ github.ref }}-rerecord (for valid triggers with re-record-tests label) - # ${{ github.workflow }}-${{ github.ref }}-replay (for valid triggers without re-record-tests label) - # ${{ github.workflow }}-${{ github.ref }}-no-run (for invalid triggers that will be skipped) + # ${{ github.workflow }}-${{ github.ref }}-rerecord (for labeled events with re-record-tests label) + # ${{ github.workflow }}-${{ github.ref }}-replay (for all non-labeled events) + # ${{ github.workflow }}-${{ github.ref }}-no-run (for labeled events without re-record-tests label) # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. group: >- ${{ github.workflow }}-${{ github.ref }}-${{ - ((github.event.action == 'opened' || github.event.action == 'synchronize') && 'replay') || - ((github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) && 'rerecord' || - 'no-run') + github.event.action == 'labeled' && ( + contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'no-run' + ) || 'replay' }} cancel-in-progress: true jobs: discover-tests: if: | - github.event_name == 'push' || - github.event_name == 'schedule' || - github.event_name == 'workflow_dispatch' || - github.event.action == 'opened' || - github.event.action == 'synchronize' || - (github.event.action == 'labeled' && contains(github.event.pull_request.labels.*.name, 're-record-tests')) + github.event.action != 'labeled' || + contains(github.event.pull_request.labels.*.name, 're-record-tests') runs-on: ubuntu-latest outputs: rerecord-tests: ${{ steps.check-rerecord-tests.outputs.rerecord-tests }} From 22f79bdb9eab5474d6644e66036db3d3daa79679 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 13:22:24 -0700 Subject: [PATCH 15/19] fix(ci): lets attempt another fix for concurrency --- .github/workflows/integration-tests.yml | 5 +++-- .github/workflows/integration-vision-tests.yml | 5 +++-- 2 files changed, 6 insertions(+), 4 deletions(-) diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index 12e3af862..33e80d49c 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -37,13 +37,14 @@ on: default: '' concurrency: - # This creates three concurrency groups: + # Skip concurrency for pushes to main - each commit should be tested independently + # For other events, create concurrency groups: # ${{ github.workflow }}-${{ github.ref }}-rerecord (for labeled events with re-record-tests label) # ${{ github.workflow }}-${{ github.ref }}-replay (for all non-labeled events) # ${{ github.workflow }}-${{ github.ref }}-no-run (for labeled events without re-record-tests label) # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. group: >- - ${{ github.workflow }}-${{ github.ref }}-${{ + ${{ github.workflow }}-${{ github.ref == 'refs/heads/main' && github.run_id || github.ref }}-${{ github.event.action == 'labeled' && ( contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'no-run' ) || 'replay' diff --git a/.github/workflows/integration-vision-tests.yml b/.github/workflows/integration-vision-tests.yml index cad945045..759cde07e 100644 --- a/.github/workflows/integration-vision-tests.yml +++ b/.github/workflows/integration-vision-tests.yml @@ -29,13 +29,14 @@ on: default: '' concurrency: - # This creates three concurrency groups: + # Skip concurrency for pushes to main - each commit should be tested independently + # For other events, create concurrency groups: # ${{ github.workflow }}-${{ github.ref }}-rerecord (for labeled events with re-record-tests label) # ${{ github.workflow }}-${{ github.ref }}-replay (for all non-labeled events) # ${{ github.workflow }}-${{ github.ref }}-no-run (for labeled events without re-record-tests label) # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. group: >- - ${{ github.workflow }}-${{ github.ref }}-${{ + ${{ github.workflow }}-${{ github.ref == 'refs/heads/main' && github.run_id || github.ref }}-${{ github.event.action == 'labeled' && ( contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'no-run' ) || 'replay' From 218c89fff11a5e81b38d3c10ea2ed7d0c164f2ea Mon Sep 17 00:00:00 2001 From: Matthew Farrellee Date: Thu, 31 Jul 2025 16:33:16 -0400 Subject: [PATCH 16/19] feat: Add clear error message when API key is missing (#2992) # What does this PR do? Improve user experience by providing specific guidance when no API key is available, showing both provider data header and config options with the correct field name for each provider. Also adds comprehensive test coverage for API key resolution scenarios. addresses #2990 for providers using litellm openai mixin ## Test Plan `./scripts/unit-tests.sh tests/unit/providers/inference/test_litellm_openai_mixin.py` --- .../utils/inference/litellm_openai_mixin.py | 6 + .../inference/test_litellm_openai_mixin.py | 112 ++++++++++++++++++ 2 files changed, 118 insertions(+) create mode 100644 tests/unit/providers/inference/test_litellm_openai_mixin.py diff --git a/llama_stack/providers/utils/inference/litellm_openai_mixin.py b/llama_stack/providers/utils/inference/litellm_openai_mixin.py index abba62f87..e9a41fcf3 100644 --- a/llama_stack/providers/utils/inference/litellm_openai_mixin.py +++ b/llama_stack/providers/utils/inference/litellm_openai_mixin.py @@ -254,6 +254,12 @@ class LiteLLMOpenAIMixin( api_key = getattr(provider_data, key_field) else: api_key = self.api_key_from_config + if not api_key: + raise ValueError( + "API key is not set. Please provide a valid API key in the " + "provider data header, e.g. x-llamastack-provider-data: " + f'{{"{key_field}": ""}}, or in the provider config.' + ) return api_key async def embeddings( diff --git a/tests/unit/providers/inference/test_litellm_openai_mixin.py b/tests/unit/providers/inference/test_litellm_openai_mixin.py new file mode 100644 index 000000000..bbc437edf --- /dev/null +++ b/tests/unit/providers/inference/test_litellm_openai_mixin.py @@ -0,0 +1,112 @@ +# Copyright (c) Meta Platforms, Inc. and affiliates. +# All rights reserved. +# +# This source code is licensed under the terms described in the LICENSE file in +# the root directory of this source tree. + +import json +from unittest.mock import MagicMock + +import pytest +from pydantic import BaseModel, Field + +from llama_stack.core.request_headers import request_provider_data_context +from llama_stack.providers.utils.inference.litellm_openai_mixin import LiteLLMOpenAIMixin + + +# Test fixtures and helper classes +class TestConfig(BaseModel): + api_key: str | None = Field(default=None) + + +class TestProviderDataValidator(BaseModel): + test_api_key: str | None = Field(default=None) + + +class TestLiteLLMAdapter(LiteLLMOpenAIMixin): + def __init__(self, config: TestConfig): + super().__init__( + model_entries=[], + litellm_provider_name="test", + api_key_from_config=config.api_key, + provider_data_api_key_field="test_api_key", + openai_compat_api_base=None, + ) + + +@pytest.fixture +def adapter_with_config_key(): + """Fixture to create adapter with API key in config""" + config = TestConfig(api_key="config-api-key") + adapter = TestLiteLLMAdapter(config) + adapter.__provider_spec__ = MagicMock() + adapter.__provider_spec__.provider_data_validator = ( + "tests.unit.providers.inference.test_litellm_openai_mixin.TestProviderDataValidator" + ) + return adapter + + +@pytest.fixture +def adapter_without_config_key(): + """Fixture to create adapter without API key in config""" + config = TestConfig(api_key=None) + adapter = TestLiteLLMAdapter(config) + adapter.__provider_spec__ = MagicMock() + adapter.__provider_spec__.provider_data_validator = ( + "tests.unit.providers.inference.test_litellm_openai_mixin.TestProviderDataValidator" + ) + return adapter + + +def test_api_key_from_config_when_no_provider_data(adapter_with_config_key): + """Test that adapter uses config API key when no provider data is available""" + api_key = adapter_with_config_key.get_api_key() + assert api_key == "config-api-key" + + +def test_provider_data_takes_priority_over_config(adapter_with_config_key): + """Test that provider data API key overrides config API key""" + with request_provider_data_context( + {"x-llamastack-provider-data": json.dumps({"test_api_key": "provider-data-key"})} + ): + api_key = adapter_with_config_key.get_api_key() + assert api_key == "provider-data-key" + + +def test_fallback_to_config_when_provider_data_missing_key(adapter_with_config_key): + """Test fallback to config when provider data doesn't have the required key""" + with request_provider_data_context({"x-llamastack-provider-data": json.dumps({"wrong_key": "some-value"})}): + api_key = adapter_with_config_key.get_api_key() + assert api_key == "config-api-key" + + +def test_error_when_no_api_key_available(adapter_without_config_key): + """Test that ValueError is raised when neither config nor provider data have API key""" + with pytest.raises(ValueError, match="API key is not set"): + adapter_without_config_key.get_api_key() + + +def test_error_when_provider_data_has_wrong_key(adapter_without_config_key): + """Test that ValueError is raised when provider data exists but doesn't have required key""" + with request_provider_data_context({"x-llamastack-provider-data": json.dumps({"wrong_key": "some-value"})}): + with pytest.raises(ValueError, match="API key is not set"): + adapter_without_config_key.get_api_key() + + +def test_provider_data_works_when_config_is_none(adapter_without_config_key): + """Test that provider data works even when config has no API key""" + with request_provider_data_context( + {"x-llamastack-provider-data": json.dumps({"test_api_key": "provider-only-key"})} + ): + api_key = adapter_without_config_key.get_api_key() + assert api_key == "provider-only-key" + + +def test_error_message_includes_correct_field_names(adapter_without_config_key): + """Test that error message includes correct field name and header information""" + try: + adapter_without_config_key.get_api_key() + raise AssertionError("Should have raised ValueError") + except ValueError as e: + assert "test_api_key" in str(e) # Should mention the correct field name + assert "x-llamastack-provider-data" in str(e) # Should mention header name From f4489eeb83735ba7f81d6d92f9392a9b15fe881f Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 15:18:18 -0700 Subject: [PATCH 17/19] fix(ci): simplify integration tests replay mode (#2997) We are going to split record and replay workflows completely to simplify the concurrency key design. We can add vision tests by just adding to our matrix. --- .../actions/run-and-record-tests/action.yml | 2 +- .../actions/setup-test-environment/action.yml | 7 +- .github/workflows/README.md | 3 +- .github/workflows/integration-tests.yml | 86 ++--------- .../workflows/integration-vision-tests.yml | 141 ------------------ 5 files changed, 19 insertions(+), 220 deletions(-) delete mode 100644 .github/workflows/integration-vision-tests.yml diff --git a/.github/actions/run-and-record-tests/action.yml b/.github/actions/run-and-record-tests/action.yml index 333b6b99f..a6acc5ce6 100644 --- a/.github/actions/run-and-record-tests/action.yml +++ b/.github/actions/run-and-record-tests/action.yml @@ -192,7 +192,7 @@ runs: if: ${{ always() }} uses: actions/upload-artifact@ea165f8d65b6e75b540449e92b4886f43607fa02 # v4.6.2 with: - name: ${{ inputs.inference-mode }}-logs-${{ github.run_id }}-${{ github.run_attempt || '' }}-${{ inputs.provider }}-${{ inputs.run-vision-tests }}-${{ inputs.stack-config }} + name: logs-${{ github.run_id }}-${{ github.run_attempt || '' }}-${{ strategy.job-index }} path: | *.log retention-days: 1 diff --git a/.github/actions/setup-test-environment/action.yml b/.github/actions/setup-test-environment/action.yml index 7296e9bfd..30b9b0130 100644 --- a/.github/actions/setup-test-environment/action.yml +++ b/.github/actions/setup-test-environment/action.yml @@ -16,6 +16,9 @@ inputs: description: 'Whether to setup provider for vision tests' required: false default: 'false' + inference-mode: + description: 'Inference mode (record or replay)' + required: true runs: using: 'composite' @@ -27,13 +30,13 @@ runs: client-version: ${{ inputs.client-version }} - name: Setup ollama - if: ${{ inputs.provider == 'ollama' }} + if: ${{ inputs.provider == 'ollama' && inputs.inference-mode == 'record' }} uses: ./.github/actions/setup-ollama with: run-vision-tests: ${{ inputs.run-vision-tests }} - name: Setup vllm - if: ${{ inputs.provider == 'vllm' }} + if: ${{ inputs.provider == 'vllm' && inputs.inference-mode == 'record' }} uses: ./.github/actions/setup-vllm - name: Build Llama Stack diff --git a/.github/workflows/README.md b/.github/workflows/README.md index 875b7f99f..ef591a51d 100644 --- a/.github/workflows/README.md +++ b/.github/workflows/README.md @@ -8,9 +8,8 @@ Llama Stack uses GitHub Actions for Continuous Integration (CI). Below is a tabl | Installer CI | [install-script-ci.yml](install-script-ci.yml) | Test the installation script | | Integration Auth Tests | [integration-auth-tests.yml](integration-auth-tests.yml) | Run the integration test suite with Kubernetes authentication | | SqlStore Integration Tests | [integration-sql-store-tests.yml](integration-sql-store-tests.yml) | Run the integration test suite with SqlStore | -| Integration Tests | [integration-tests.yml](integration-tests.yml) | Run the integration test suite from tests/integration | +| Integration Tests (Replay) | [integration-tests.yml](integration-tests.yml) | Run the integration test suite from tests/integration in replay mode | | Vector IO Integration Tests | [integration-vector-io-tests.yml](integration-vector-io-tests.yml) | Run the integration test suite with various VectorIO providers | -| Vision Inference Integration Tests | [integration-vision-tests.yml](integration-vision-tests.yml) | Run vision inference integration test suite from tests/integration/inference | | Pre-commit | [pre-commit.yml](pre-commit.yml) | Run pre-commit checks | | Test Llama Stack Build | [providers-build.yml](providers-build.yml) | Test llama stack build | | Python Package Build Test | [python-build-test.yml](python-build-test.yml) | Test building the llama-stack PyPI project | diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index 33e80d49c..513b7c861 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -1,13 +1,13 @@ -name: Integration Tests +name: Integration Tests (Replay) -run-name: Run the integration test suite from tests/integration +run-name: Run the integration test suite from tests/integration in replay mode on: push: branches: [ main ] - pull_request_target: + pull_request: branches: [ main ] - types: [opened, synchronize, labeled] + types: [opened, synchronize, reopened] paths: - 'llama_stack/**' - 'tests/**' @@ -31,35 +31,17 @@ on: description: 'Test against a specific provider' type: string default: 'ollama' - force-inference-mode: - description: 'Force inference mode (record or replay)' - type: string - default: '' concurrency: # Skip concurrency for pushes to main - each commit should be tested independently - # For other events, create concurrency groups: - # ${{ github.workflow }}-${{ github.ref }}-rerecord (for labeled events with re-record-tests label) - # ${{ github.workflow }}-${{ github.ref }}-replay (for all non-labeled events) - # ${{ github.workflow }}-${{ github.ref }}-no-run (for labeled events without re-record-tests label) - # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. - group: >- - ${{ github.workflow }}-${{ github.ref == 'refs/heads/main' && github.run_id || github.ref }}-${{ - github.event.action == 'labeled' && ( - contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'no-run' - ) || 'replay' - }} + group: ${{ github.workflow }}-${{ github.ref == 'refs/heads/main' && github.run_id || github.ref }} cancel-in-progress: true jobs: discover-tests: - if: | - github.event.action != 'labeled' || - contains(github.event.pull_request.labels.*.name, 're-record-tests') runs-on: ubuntu-latest outputs: test-types: ${{ steps.generate-test-types.outputs.test-types }} - rerecord-tests: ${{ steps.check-rerecord-tests.outputs.rerecord-tests }} steps: - name: Checkout repository @@ -69,61 +51,13 @@ jobs: id: generate-test-types run: | # Get test directories dynamically, excluding non-test directories + # NOTE: we are excluding post_training since the tests take too long TEST_TYPES=$(find tests/integration -maxdepth 1 -mindepth 1 -type d -printf "%f\n" | - grep -Ev "^(__pycache__|fixtures|test_cases|recordings)$" | + grep -Ev "^(__pycache__|fixtures|test_cases|recordings|post_training)$" | sort | jq -R -s -c 'split("\n")[:-1]') echo "test-types=$TEST_TYPES" >> $GITHUB_OUTPUT - - name: Check if re-record-tests label exists - id: check-rerecord-tests - run: | - if [[ "${{ inputs.force-inference-mode }}" == "record" ]]; then - echo "rerecord-tests=true" >> $GITHUB_OUTPUT - elif [[ "${{ inputs.force-inference-mode }}" == "replay" ]]; then - echo "rerecord-tests=false" >> $GITHUB_OUTPUT - else - if [[ "${{ contains(github.event.pull_request.labels.*.name, 're-record-tests') }}" == "true" ]]; then - echo "rerecord-tests=true" >> $GITHUB_OUTPUT - else - echo "rerecord-tests=false" >> $GITHUB_OUTPUT - fi - fi - - record-tests: - # Sequential job for recording to avoid SQLite conflicts - if: ${{ needs.discover-tests.outputs.rerecord-tests == 'true' }} - needs: discover-tests - runs-on: ubuntu-latest - - permissions: - contents: write - pull-requests: write - - steps: - - name: Checkout repository - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 - with: - ref: ${{ github.event.pull_request.head.ref }} - fetch-depth: 0 - - - name: Setup test environment - uses: ./.github/actions/setup-test-environment - with: - python-version: "3.12" # Use single Python version for recording - client-version: "latest" - provider: ${{ inputs.test-provider || 'ollama' }} - - - name: Run and record tests - uses: ./.github/actions/run-and-record-tests - with: - test-types: ${{ needs.discover-tests.outputs.test-types }} - stack-config: 'server:ci-tests' # recording must be done with server since more tests are run - provider: ${{ inputs.test-provider || 'ollama' }} - inference-mode: 'record' - run-replay-mode-tests: - # Skip this job if we're in recording mode (handled by record-tests job) - if: ${{ needs.discover-tests.outputs.rerecord-tests != 'true' }} needs: discover-tests runs-on: ubuntu-latest @@ -135,6 +69,7 @@ jobs: provider: ${{ (github.event.schedule == '1 0 * * 0') && fromJSON('["vllm"]') || fromJSON(format('["{0}"]', github.event.inputs.test-provider || 'ollama')) }} python-version: ["3.12", "3.13"] client-version: ${{ (github.event.schedule == '0 0 * * 0' || github.event.inputs.test-all-client-versions == 'true') && fromJSON('["published", "latest"]') || fromJSON('["latest"]') }} + run-vision-tests: ['true', 'false'] steps: - name: Checkout repository @@ -146,11 +81,14 @@ jobs: python-version: ${{ matrix.python-version }} client-version: ${{ matrix.client-version }} provider: ${{ matrix.provider }} + run-vision-tests: ${{ matrix.run-vision-tests }} + inference-mode: 'replay' - - name: Run and record tests + - name: Run tests uses: ./.github/actions/run-and-record-tests with: test-types: ${{ needs.discover-tests.outputs.test-types }} stack-config: ${{ matrix.client-type == 'library' && 'ci-tests' || 'server:ci-tests' }} provider: ${{ matrix.provider }} inference-mode: 'replay' + run-vision-tests: ${{ matrix.run-vision-tests }} diff --git a/.github/workflows/integration-vision-tests.yml b/.github/workflows/integration-vision-tests.yml deleted file mode 100644 index 759cde07e..000000000 --- a/.github/workflows/integration-vision-tests.yml +++ /dev/null @@ -1,141 +0,0 @@ -name: Vision Inference Integration Tests - -run-name: Run vision inference integration test suite from tests/integration/inference - -on: - push: - branches: [ main ] - pull_request_target: - branches: [ main ] - types: [opened, synchronize, labeled] - paths: - - 'llama_stack/**' - - 'tests/**' - - 'uv.lock' - - 'pyproject.toml' - - '.github/workflows/integration-vision-tests.yml' # This workflow - - '.github/actions/setup-ollama/action.yml' - - '.github/actions/setup-test-environment/action.yml' - - '.github/actions/run-and-record-tests/action.yml' - workflow_dispatch: - inputs: - test-all-client-versions: - description: 'Test against both the latest and published versions' - type: boolean - default: false - force-inference-mode: - description: 'Force inference mode (record or replay)' - type: string - default: '' - -concurrency: - # Skip concurrency for pushes to main - each commit should be tested independently - # For other events, create concurrency groups: - # ${{ github.workflow }}-${{ github.ref }}-rerecord (for labeled events with re-record-tests label) - # ${{ github.workflow }}-${{ github.ref }}-replay (for all non-labeled events) - # ${{ github.workflow }}-${{ github.ref }}-no-run (for labeled events without re-record-tests label) - # The "no-run" group ensures that irrelevant label events don't interfere with the real workflows. - group: >- - ${{ github.workflow }}-${{ github.ref == 'refs/heads/main' && github.run_id || github.ref }}-${{ - github.event.action == 'labeled' && ( - contains(github.event.pull_request.labels.*.name, 're-record-tests') && 'rerecord' || 'no-run' - ) || 'replay' - }} - cancel-in-progress: true - -jobs: - discover-tests: - if: | - github.event.action != 'labeled' || - contains(github.event.pull_request.labels.*.name, 're-record-tests') - runs-on: ubuntu-latest - outputs: - rerecord-tests: ${{ steps.check-rerecord-tests.outputs.rerecord-tests }} - - steps: - - name: Checkout repository - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 - - - name: Check if re-record-tests label exists - id: check-rerecord-tests - run: | - if [[ "${{ inputs.force-inference-mode }}" == "record" ]]; then - echo "rerecord-tests=true" >> $GITHUB_OUTPUT - elif [[ "${{ inputs.force-inference-mode }}" == "replay" ]]; then - echo "rerecord-tests=false" >> $GITHUB_OUTPUT - else - if [[ "${{ contains(github.event.pull_request.labels.*.name, 're-record-tests') }}" == "true" ]]; then - echo "rerecord-tests=true" >> $GITHUB_OUTPUT - else - echo "rerecord-tests=false" >> $GITHUB_OUTPUT - fi - fi - - record-tests: - # Sequential job for recording to avoid SQLite conflicts - if: ${{ needs.discover-tests.outputs.rerecord-tests == 'true' }} - needs: discover-tests - runs-on: ubuntu-latest - - permissions: - contents: write - pull-requests: write - - steps: - - name: Checkout repository - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 - with: - ref: ${{ github.event.pull_request.head.ref }} - fetch-depth: 0 - - - name: Setup test environment - uses: ./.github/actions/setup-test-environment - with: - python-version: "3.12" # Use single Python version for recording - client-version: "latest" - provider: 'ollama' - run-vision-tests: 'true' - - - name: Run and record tests - uses: ./.github/actions/run-and-record-tests - with: - test-types: '["vision"]' - stack-config: 'server:ci-tests' # re-recording must be done in server mode - provider: 'ollama' - inference-mode: 'record' - run-vision-tests: 'true' - - run-replay-mode-tests: - # Skip this job if we're in recording mode (handled by record-tests job) - if: ${{ needs.discover-tests.outputs.rerecord-tests != 'true' }} - needs: discover-tests - runs-on: ubuntu-latest - - strategy: - fail-fast: false - matrix: - client-type: [library] - provider: [ollama] - python-version: ["3.12"] - client-version: ["latest"] - - steps: - - name: Checkout repository - uses: actions/checkout@11bd71901bbe5b1630ceea73d27597364c9af683 # v4.2.2 - - - name: Setup test environment - uses: ./.github/actions/setup-test-environment - with: - python-version: ${{ matrix.python-version }} - client-version: ${{ matrix.client-version }} - provider: ${{ matrix.provider }} - run-vision-tests: 'true' - - - name: Run and record tests - uses: ./.github/actions/run-and-record-tests - with: - test-types: '["vision"]' - stack-config: ${{ matrix.client-type == 'library' && 'ci-tests' || 'server:ci-tests' }} - provider: ${{ matrix.provider }} - inference-mode: 'replay' - run-vision-tests: 'true' From 89ff93182c3dd9ae9cfb03b3ad31112a01b50606 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 15:32:05 -0700 Subject: [PATCH 18/19] feat(ci): only run on 3.12, run on both 3.12 and 3.13 nightly (#3000) We don't need to run on all python versions all the time --- .github/workflows/integration-tests.yml | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/.github/workflows/integration-tests.yml b/.github/workflows/integration-tests.yml index 513b7c861..487980de9 100644 --- a/.github/workflows/integration-tests.yml +++ b/.github/workflows/integration-tests.yml @@ -67,8 +67,9 @@ jobs: client-type: [library, server] # Use vllm on weekly schedule, otherwise use test-provider input (defaults to ollama) provider: ${{ (github.event.schedule == '1 0 * * 0') && fromJSON('["vllm"]') || fromJSON(format('["{0}"]', github.event.inputs.test-provider || 'ollama')) }} - python-version: ["3.12", "3.13"] - client-version: ${{ (github.event.schedule == '0 0 * * 0' || github.event.inputs.test-all-client-versions == 'true') && fromJSON('["published", "latest"]') || fromJSON('["latest"]') }} + # Use Python 3.13 only on nightly schedule (daily latest client test), otherwise use 3.12 + python-version: ${{ github.event.schedule == '0 0 * * *' && fromJSON('["3.12", "3.13"]') || fromJSON('["3.12"]') }} + client-version: ${{ (github.event.schedule == '0 0 * * *' || github.event.inputs.test-all-client-versions == 'true') && fromJSON('["published", "latest"]') || fromJSON('["latest"]') }} run-vision-tests: ['true', 'false'] steps: From 369286f95b527e6f2709f3b5b2346a08a7bff9f4 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 31 Jul 2025 15:35:42 -0700 Subject: [PATCH 19/19] fix(ci): syntax error in the disabled workflow --- .github/workflows/test-external-provider-module.yml | 6 ++---- 1 file changed, 2 insertions(+), 4 deletions(-) diff --git a/.github/workflows/test-external-provider-module.yml b/.github/workflows/test-external-provider-module.yml index c3d11f28a..d61b0dfe9 100644 --- a/.github/workflows/test-external-provider-module.yml +++ b/.github/workflows/test-external-provider-module.yml @@ -12,15 +12,13 @@ on: - 'tests/integration/**' - 'uv.lock' - 'pyproject.toml' - - 'requirements.txt' - 'tests/external/*' - '.github/workflows/test-external-provider-module.yml' # This workflow -# This workflow is disabled. See https://github.com/meta-llama/llama-stack/pull/2975#issuecomment-3138702984 for details -if: false - jobs: test-external-providers-from-module: + # This workflow is disabled. See https://github.com/meta-llama/llama-stack/pull/2975#issuecomment-3138702984 for details + if: false runs-on: ubuntu-latest strategy: matrix: