mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-08 13:00:52 +00:00
Merge f93461de5e
into sapling-pr-archive-ehhuang
This commit is contained in:
commit
d103bf6c03
104 changed files with 15263 additions and 5592 deletions
2
.github/workflows/integration-auth-tests.yml
vendored
2
.github/workflows/integration-auth-tests.yml
vendored
|
@ -86,7 +86,7 @@ jobs:
|
|||
|
||||
# avoid line breaks in the server log, especially because we grep it below.
|
||||
export COLUMNS=1984
|
||||
nohup uv run llama stack run $run_dir/run.yaml --image-type venv > server.log 2>&1 &
|
||||
nohup uv run llama stack run $run_dir/run.yaml > server.log 2>&1 &
|
||||
|
||||
- name: Wait for Llama Stack server to be ready
|
||||
run: |
|
||||
|
|
|
@ -59,7 +59,7 @@ jobs:
|
|||
# Use the virtual environment created by the build step (name comes from build config)
|
||||
source ramalama-stack-test/bin/activate
|
||||
uv pip list
|
||||
nohup llama stack run tests/external/ramalama-stack/run.yaml --image-type ${{ matrix.image-type }} > server.log 2>&1 &
|
||||
nohup llama stack run tests/external/ramalama-stack/run.yaml > server.log 2>&1 &
|
||||
|
||||
- name: Wait for Llama Stack server to be ready
|
||||
run: |
|
||||
|
|
2
.github/workflows/test-external.yml
vendored
2
.github/workflows/test-external.yml
vendored
|
@ -59,7 +59,7 @@ jobs:
|
|||
# Use the virtual environment created by the build step (name comes from build config)
|
||||
source ci-test/bin/activate
|
||||
uv pip list
|
||||
nohup llama stack run tests/external/run-byoa.yaml --image-type ${{ matrix.image-type }} > server.log 2>&1 &
|
||||
nohup llama stack run tests/external/run-byoa.yaml > server.log 2>&1 &
|
||||
|
||||
- name: Wait for Llama Stack server to be ready
|
||||
run: |
|
||||
|
|
|
@ -52,7 +52,7 @@ You can access the HuggingFace trainer via the `starter` distribution:
|
|||
|
||||
```bash
|
||||
llama stack build --distro starter --image-type venv
|
||||
llama stack run --image-type venv ~/.llama/distributions/starter/starter-run.yaml
|
||||
llama stack run ~/.llama/distributions/starter/starter-run.yaml
|
||||
```
|
||||
|
||||
### Usage Example
|
||||
|
|
|
@ -219,13 +219,10 @@ group_tools = client.tools.list_tools(toolgroup_id="search_tools")
|
|||
<TabItem value="setup" label="Setup & Configuration">
|
||||
|
||||
1. Start by registering a Tavily API key at [Tavily](https://tavily.com/).
|
||||
2. [Optional] Provide the API key directly to the Llama Stack server
|
||||
2. [Optional] Set the API key in your environment before starting the Llama Stack server
|
||||
```bash
|
||||
export TAVILY_SEARCH_API_KEY="your key"
|
||||
```
|
||||
```bash
|
||||
--env TAVILY_SEARCH_API_KEY=${TAVILY_SEARCH_API_KEY}
|
||||
```
|
||||
|
||||
</TabItem>
|
||||
<TabItem value="implementation" label="Implementation">
|
||||
|
@ -273,9 +270,9 @@ for log in EventLogger().log(response):
|
|||
<TabItem value="setup" label="Setup & Configuration">
|
||||
|
||||
1. Start by registering for a WolframAlpha API key at [WolframAlpha Developer Portal](https://developer.wolframalpha.com/access).
|
||||
2. Provide the API key either when starting the Llama Stack server:
|
||||
2. Provide the API key either by setting it in your environment before starting the Llama Stack server:
|
||||
```bash
|
||||
--env WOLFRAM_ALPHA_API_KEY=${WOLFRAM_ALPHA_API_KEY}
|
||||
export WOLFRAM_ALPHA_API_KEY="your key"
|
||||
```
|
||||
or from the client side:
|
||||
```python
|
||||
|
|
|
@ -76,7 +76,7 @@ Integration tests are located in [tests/integration](https://github.com/meta-lla
|
|||
Consult [tests/integration/README.md](https://github.com/meta-llama/llama-stack/blob/main/tests/integration/README.md) for more details on how to run the tests.
|
||||
|
||||
Note that each provider's `sample_run_config()` method (in the configuration class for that provider)
|
||||
typically references some environment variables for specifying API keys and the like. You can set these in the environment or pass these via the `--env` flag to the test command.
|
||||
typically references some environment variables for specifying API keys and the like. You can set these in the environment before running the test command.
|
||||
|
||||
|
||||
### 2. Unit Testing
|
||||
|
|
|
@ -289,10 +289,10 @@ After this step is successful, you should be able to find the built container im
|
|||
docker run -d \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
-e INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
-e OLLAMA_URL=http://host.docker.internal:11434 \
|
||||
localhost/distribution-ollama:dev \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env OLLAMA_URL=http://host.docker.internal:11434
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
Here are the docker flags and their uses:
|
||||
|
@ -305,12 +305,12 @@ Here are the docker flags and their uses:
|
|||
|
||||
* `localhost/distribution-ollama:dev`: The name and tag of the container image to run
|
||||
|
||||
* `-e INFERENCE_MODEL=$INFERENCE_MODEL`: Sets the INFERENCE_MODEL environment variable in the container
|
||||
|
||||
* `-e OLLAMA_URL=http://host.docker.internal:11434`: Sets the OLLAMA_URL environment variable in the container
|
||||
|
||||
* `--port $LLAMA_STACK_PORT`: Port number for the server to listen on
|
||||
|
||||
* `--env INFERENCE_MODEL=$INFERENCE_MODEL`: Sets the model to use for inference
|
||||
|
||||
* `--env OLLAMA_URL=http://host.docker.internal:11434`: Configures the URL for the Ollama service
|
||||
|
||||
</TabItem>
|
||||
</Tabs>
|
||||
|
||||
|
@ -320,23 +320,22 @@ Now, let's start the Llama Stack Distribution Server. You will need the YAML con
|
|||
|
||||
```
|
||||
llama stack run -h
|
||||
usage: llama stack run [-h] [--port PORT] [--image-name IMAGE_NAME] [--env KEY=VALUE]
|
||||
usage: llama stack run [-h] [--port PORT] [--image-name IMAGE_NAME]
|
||||
[--image-type {venv}] [--enable-ui]
|
||||
[config | template]
|
||||
[config | distro]
|
||||
|
||||
Start the server for a Llama Stack Distribution. You should have already built (or downloaded) and configured the distribution.
|
||||
|
||||
positional arguments:
|
||||
config | template Path to config file to use for the run or name of known template (`llama stack list` for a list). (default: None)
|
||||
config | distro Path to config file to use for the run or name of known distro (`llama stack list` for a list). (default: None)
|
||||
|
||||
options:
|
||||
-h, --help show this help message and exit
|
||||
--port PORT Port to run the server on. It can also be passed via the env var LLAMA_STACK_PORT. (default: 8321)
|
||||
--image-name IMAGE_NAME
|
||||
Name of the image to run. Defaults to the current environment (default: None)
|
||||
--env KEY=VALUE Environment variables to pass to the server in KEY=VALUE format. Can be specified multiple times. (default: None)
|
||||
[DEPRECATED] This flag is no longer supported. Please activate your virtual environment before running. (default: None)
|
||||
--image-type {venv}
|
||||
Image Type used during the build. This should be venv. (default: None)
|
||||
[DEPRECATED] This flag is no longer supported. Please activate your virtual environment before running. (default: None)
|
||||
--enable-ui Start the UI server (default: False)
|
||||
```
|
||||
|
||||
|
@ -348,9 +347,6 @@ llama stack run tgi
|
|||
|
||||
# Start using config file
|
||||
llama stack run ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml
|
||||
|
||||
# Start using a venv
|
||||
llama stack run --image-type venv ~/.llama/distributions/llamastack-my-local-stack/my-local-stack-run.yaml
|
||||
```
|
||||
|
||||
```
|
||||
|
|
|
@ -101,7 +101,7 @@ A few things to note:
|
|||
- The id is a string you can choose freely.
|
||||
- You can instantiate any number of provider instances of the same type.
|
||||
- The configuration dictionary is provider-specific.
|
||||
- Notice that configuration can reference environment variables (with default values), which are expanded at runtime. When you run a stack server (via docker or via `llama stack run`), you can specify `--env OLLAMA_URL=http://my-server:11434` to override the default value.
|
||||
- Notice that configuration can reference environment variables (with default values), which are expanded at runtime. When you run a stack server, you can set environment variables in your shell before running `llama stack run` to override the default values.
|
||||
|
||||
### Environment Variable Substitution
|
||||
|
||||
|
@ -173,13 +173,10 @@ optional_token: ${env.OPTIONAL_TOKEN:+}
|
|||
|
||||
#### Runtime Override
|
||||
|
||||
You can override environment variables at runtime when starting the server:
|
||||
You can override environment variables at runtime by setting them in your shell before starting the server:
|
||||
|
||||
```bash
|
||||
# Override specific environment variables
|
||||
llama stack run --config run.yaml --env API_KEY=sk-123 --env BASE_URL=https://custom-api.com
|
||||
|
||||
# Or set them in your shell
|
||||
# Set environment variables in your shell
|
||||
export API_KEY=sk-123
|
||||
export BASE_URL=https://custom-api.com
|
||||
llama stack run --config run.yaml
|
||||
|
|
|
@ -69,10 +69,10 @@ docker run \
|
|||
-it \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ./run.yaml:/root/my-run.yaml \
|
||||
-e WATSONX_API_KEY=$WATSONX_API_KEY \
|
||||
-e WATSONX_PROJECT_ID=$WATSONX_PROJECT_ID \
|
||||
-e WATSONX_BASE_URL=$WATSONX_BASE_URL \
|
||||
llamastack/distribution-watsonx \
|
||||
--config /root/my-run.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env WATSONX_API_KEY=$WATSONX_API_KEY \
|
||||
--env WATSONX_PROJECT_ID=$WATSONX_PROJECT_ID \
|
||||
--env WATSONX_BASE_URL=$WATSONX_BASE_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
|
|
@ -129,11 +129,11 @@ docker run -it \
|
|||
# NOTE: mount the llama-stack / llama-model directories if testing local changes else not needed
|
||||
-v $HOME/git/llama-stack:/app/llama-stack-source -v $HOME/git/llama-models:/app/llama-models-source \
|
||||
# localhost/distribution-dell:dev if building / testing locally
|
||||
llamastack/distribution-dell\
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
-e INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
-e DEH_URL=$DEH_URL \
|
||||
-e CHROMA_URL=$CHROMA_URL \
|
||||
llamastack/distribution-dell \
|
||||
--port $LLAMA_STACK_PORT
|
||||
|
||||
```
|
||||
|
||||
|
@ -154,14 +154,14 @@ docker run \
|
|||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v $HOME/.llama:/root/.llama \
|
||||
-v ./llama_stack/distributions/tgi/run-with-safety.yaml:/root/my-run.yaml \
|
||||
-e INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
-e DEH_URL=$DEH_URL \
|
||||
-e SAFETY_MODEL=$SAFETY_MODEL \
|
||||
-e DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
-e CHROMA_URL=$CHROMA_URL \
|
||||
llamastack/distribution-dell \
|
||||
--config /root/my-run.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env SAFETY_MODEL=$SAFETY_MODEL \
|
||||
--env DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
### Via venv
|
||||
|
@ -170,21 +170,21 @@ Make sure you have done `pip install llama-stack` and have the Llama Stack CLI a
|
|||
|
||||
```bash
|
||||
llama stack build --distro dell --image-type venv
|
||||
llama stack run dell
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
DEH_URL=$DEH_URL \
|
||||
CHROMA_URL=$CHROMA_URL \
|
||||
llama stack run dell \
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
||||
```bash
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
DEH_URL=$DEH_URL \
|
||||
SAFETY_MODEL=$SAFETY_MODEL \
|
||||
DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
CHROMA_URL=$CHROMA_URL \
|
||||
llama stack run ./run-with-safety.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env SAFETY_MODEL=$SAFETY_MODEL \
|
||||
--env DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
|
|
@ -84,9 +84,9 @@ docker run \
|
|||
--gpu all \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
-e INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
llamastack/distribution-meta-reference-gpu \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
@ -98,10 +98,10 @@ docker run \
|
|||
--gpu all \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
-e INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
-e SAFETY_MODEL=meta-llama/Llama-Guard-3-1B \
|
||||
llamastack/distribution-meta-reference-gpu \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
### Via venv
|
||||
|
@ -110,16 +110,16 @@ Make sure you have done `uv pip install llama-stack` and have the Llama Stack CL
|
|||
|
||||
```bash
|
||||
llama stack build --distro meta-reference-gpu --image-type venv
|
||||
INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
llama stack run distributions/meta-reference-gpu/run.yaml \
|
||||
--port 8321 \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--port 8321
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
||||
```bash
|
||||
INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
SAFETY_MODEL=meta-llama/Llama-Guard-3-1B \
|
||||
llama stack run distributions/meta-reference-gpu/run-with-safety.yaml \
|
||||
--port 8321 \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--port 8321
|
||||
```
|
||||
|
|
|
@ -129,10 +129,10 @@ docker run \
|
|||
--pull always \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ./run.yaml:/root/my-run.yaml \
|
||||
-e NVIDIA_API_KEY=$NVIDIA_API_KEY \
|
||||
llamastack/distribution-nvidia \
|
||||
--config /root/my-run.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env NVIDIA_API_KEY=$NVIDIA_API_KEY
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
### Via venv
|
||||
|
@ -142,10 +142,10 @@ If you've set up your local development environment, you can also build the imag
|
|||
```bash
|
||||
INFERENCE_MODEL=meta-llama/Llama-3.1-8B-Instruct
|
||||
llama stack build --distro nvidia --image-type venv
|
||||
NVIDIA_API_KEY=$NVIDIA_API_KEY \
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
llama stack run ./run.yaml \
|
||||
--port 8321 \
|
||||
--env NVIDIA_API_KEY=$NVIDIA_API_KEY \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL
|
||||
--port 8321
|
||||
```
|
||||
|
||||
## Example Notebooks
|
||||
|
|
|
@ -86,9 +86,9 @@ docker run -it \
|
|||
--pull always \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
-e OLLAMA_URL=http://host.docker.internal:11434 \
|
||||
llamastack/distribution-starter \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env OLLAMA_URL=http://host.docker.internal:11434
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
Note to start the container with Podman, you can do the same but replace `docker` at the start of the command with
|
||||
`podman`. If you are using `podman` older than `4.7.0`, please also replace `host.docker.internal` in the `OLLAMA_URL`
|
||||
|
@ -106,9 +106,9 @@ docker run -it \
|
|||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
--network=host \
|
||||
-e OLLAMA_URL=http://localhost:11434 \
|
||||
llamastack/distribution-starter \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env OLLAMA_URL=http://localhost:11434
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
:::
|
||||
You will see output like below:
|
||||
|
|
|
@ -15,6 +15,7 @@ Anthropic inference provider for accessing Claude models and Anthropic's AI serv
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `api_key` | `str \| None` | No | | API key for Anthropic models |
|
||||
|
||||
## Sample Configuration
|
||||
|
|
|
@ -22,6 +22,7 @@ https://learn.microsoft.com/en-us/azure/ai-foundry/openai/overview
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `api_key` | `<class 'pydantic.types.SecretStr'>` | No | | Azure API key for Azure |
|
||||
| `api_base` | `<class 'pydantic.networks.HttpUrl'>` | No | | Azure API base for Azure (e.g., https://your-resource-name.openai.azure.com) |
|
||||
| `api_version` | `str \| None` | No | | Azure API version for Azure (e.g., 2024-12-01-preview) |
|
||||
|
|
|
@ -15,6 +15,7 @@ AWS Bedrock inference provider for accessing various AI models through AWS's man
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `aws_access_key_id` | `str \| None` | No | | The AWS access key to use. Default use environment variable: AWS_ACCESS_KEY_ID |
|
||||
| `aws_secret_access_key` | `str \| None` | No | | The AWS secret access key to use. Default use environment variable: AWS_SECRET_ACCESS_KEY |
|
||||
| `aws_session_token` | `str \| None` | No | | The AWS session token to use. Default use environment variable: AWS_SESSION_TOKEN |
|
||||
|
|
|
@ -15,6 +15,7 @@ Cerebras inference provider for running models on Cerebras Cloud platform.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `base_url` | `<class 'str'>` | No | https://api.cerebras.ai | Base URL for the Cerebras API |
|
||||
| `api_key` | `<class 'pydantic.types.SecretStr'>` | No | | Cerebras API Key |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ Databricks inference provider for running models on Databricks' unified analytic
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `str \| None` | No | | The URL for the Databricks model serving endpoint |
|
||||
| `api_token` | `<class 'pydantic.types.SecretStr'>` | No | | The Databricks API token |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ Fireworks AI inference provider for Llama models and other AI models on the Fire
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | https://api.fireworks.ai/inference/v1 | The URL for the Fireworks server |
|
||||
| `api_key` | `pydantic.types.SecretStr \| None` | No | | The Fireworks.ai API Key |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ Google Gemini inference provider for accessing Gemini models and Google's AI ser
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `api_key` | `str \| None` | No | | API key for Gemini models |
|
||||
|
||||
## Sample Configuration
|
||||
|
|
|
@ -15,6 +15,7 @@ Groq inference provider for ultra-fast inference using Groq's LPU technology.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `api_key` | `str \| None` | No | | The Groq API key |
|
||||
| `url` | `<class 'str'>` | No | https://api.groq.com | The URL for the Groq AI server |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ Llama OpenAI-compatible provider for using Llama models with OpenAI API format.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `api_key` | `str \| None` | No | | The Llama API key |
|
||||
| `openai_compat_api_base` | `<class 'str'>` | No | https://api.llama.com/compat/v1/ | The URL for the Llama API server |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ NVIDIA inference provider for accessing NVIDIA NIM models and AI services.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | https://integrate.api.nvidia.com | A base url for accessing the NVIDIA NIM |
|
||||
| `api_key` | `pydantic.types.SecretStr \| None` | No | | The NVIDIA API key, only needed of using the hosted service |
|
||||
| `timeout` | `<class 'int'>` | No | 60 | Timeout for the HTTP requests |
|
||||
|
|
|
@ -15,8 +15,8 @@ Ollama inference provider for running local models through the Ollama runtime.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | http://localhost:11434 | |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically |
|
||||
|
||||
## Sample Configuration
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ OpenAI inference provider for accessing GPT models and other OpenAI services.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `api_key` | `str \| None` | No | | API key for OpenAI models |
|
||||
| `base_url` | `<class 'str'>` | No | https://api.openai.com/v1 | Base URL for OpenAI API |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ Passthrough inference provider for connecting to any external inference service
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | | The URL for the passthrough endpoint |
|
||||
| `api_key` | `pydantic.types.SecretStr \| None` | No | | API Key for the passthrouth endpoint |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ RunPod inference provider for running models on RunPod's cloud GPU platform.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `str \| None` | No | | The URL for the Runpod model serving endpoint |
|
||||
| `api_token` | `str \| None` | No | | The API token |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ SambaNova inference provider for running models on SambaNova's dataflow architec
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | https://api.sambanova.ai/v1 | The URL for the SambaNova AI server |
|
||||
| `api_key` | `pydantic.types.SecretStr \| None` | No | | The SambaNova cloud API Key |
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ Text Generation Inference (TGI) provider for HuggingFace model serving.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | | The URL for the TGI serving endpoint |
|
||||
|
||||
## Sample Configuration
|
||||
|
|
|
@ -15,6 +15,7 @@ Together AI inference provider for open-source models and collaborative AI devel
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | https://api.together.xyz/v1 | The URL for the Together AI server |
|
||||
| `api_key` | `pydantic.types.SecretStr \| None` | No | | The Together AI API Key |
|
||||
|
||||
|
|
|
@ -54,6 +54,7 @@ Available Models:
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `project` | `<class 'str'>` | No | | Google Cloud project ID for Vertex AI |
|
||||
| `location` | `<class 'str'>` | No | us-central1 | Google Cloud location for Vertex AI |
|
||||
|
||||
|
|
|
@ -15,11 +15,11 @@ Remote vLLM inference provider for connecting to vLLM servers.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `str \| None` | No | | The URL for the vLLM model serving endpoint |
|
||||
| `max_tokens` | `<class 'int'>` | No | 4096 | Maximum number of tokens to generate. |
|
||||
| `api_token` | `str \| None` | No | fake | The API token |
|
||||
| `tls_verify` | `bool \| str` | No | True | Whether to verify TLS certificates. Can be a boolean or a path to a CA certificate file. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically |
|
||||
|
||||
## Sample Configuration
|
||||
|
||||
|
|
|
@ -15,6 +15,7 @@ IBM WatsonX inference provider for accessing AI models on IBM's WatsonX platform
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `url` | `<class 'str'>` | No | https://us-south.ml.cloud.ibm.com | A base url for accessing the watsonx.ai |
|
||||
| `api_key` | `pydantic.types.SecretStr \| None` | No | | The watsonx API key |
|
||||
| `project_id` | `str \| None` | No | | The Project ID key |
|
||||
|
|
|
@ -15,6 +15,7 @@ AWS Bedrock safety provider for content moderation using AWS's safety services.
|
|||
| Field | Type | Required | Default | Description |
|
||||
|-------|------|----------|---------|-------------|
|
||||
| `allowed_models` | `list[str \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
||||
| `refresh_models` | `<class 'bool'>` | No | False | Whether to refresh models periodically from the provider |
|
||||
| `aws_access_key_id` | `str \| None` | No | | The AWS access key to use. Default use environment variable: AWS_ACCESS_KEY_ID |
|
||||
| `aws_secret_access_key` | `str \| None` | No | | The AWS secret access key to use. Default use environment variable: AWS_SECRET_ACCESS_KEY |
|
||||
| `aws_session_token` | `str \| None` | No | | The AWS session token to use. Default use environment variable: AWS_SESSION_TOKEN |
|
||||
|
|
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
File diff suppressed because one or more lines are too long
|
@ -1,366 +1,315 @@
|
|||
{
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "c1e7571c",
|
||||
"metadata": {
|
||||
"id": "c1e7571c"
|
||||
},
|
||||
"source": [
|
||||
"[](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb)\n",
|
||||
"\n",
|
||||
"# Llama Stack - Building AI Applications\n",
|
||||
"\n",
|
||||
"<img src=\"https://llamastack.github.io/latest/_images/llama-stack.png\" alt=\"drawing\" width=\"500\"/>\n",
|
||||
"\n",
|
||||
"Get started with Llama Stack in minutes!\n",
|
||||
"\n",
|
||||
"[Llama Stack](https://github.com/meta-llama/llama-stack) is a stateful service with REST APIs to support the seamless transition of AI applications across different environments. You can build and test using a local server first and deploy to a hosted endpoint for production.\n",
|
||||
"\n",
|
||||
"In this guide, we'll walk through how to build a RAG application locally using Llama Stack with [Ollama](https://ollama.com/)\n",
|
||||
"as the inference [provider](docs/source/providers/index.md#inference) for a Llama Model.\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "4CV1Q19BDMVw",
|
||||
"metadata": {
|
||||
"id": "4CV1Q19BDMVw"
|
||||
},
|
||||
"source": [
|
||||
"## Step 1: Install and setup"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "K4AvfUAJZOeS",
|
||||
"metadata": {
|
||||
"id": "K4AvfUAJZOeS"
|
||||
},
|
||||
"source": [
|
||||
"### 1.1. Install uv and test inference with Ollama\n",
|
||||
"\n",
|
||||
"We'll install [uv](https://docs.astral.sh/uv/) to setup the Python virtual environment, along with [colab-xterm](https://github.com/InfuseAI/colab-xterm) for running command-line tools, and [Ollama](https://ollama.com/download) as the inference provider."
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "7a2d7b85",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"%pip install uv llama_stack llama-stack-client\n",
|
||||
"\n",
|
||||
"## If running on Collab:\n",
|
||||
"# !pip install colab-xterm\n",
|
||||
"# %load_ext colabxterm\n",
|
||||
"\n",
|
||||
"!curl https://ollama.ai/install.sh | sh"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "39fa584b",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"### 1.2. Test inference with Ollama"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "3bf81522",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"We’ll now launch a terminal and run inference on a Llama model with Ollama to verify that the model is working correctly."
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "a7e8e0f1",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"## If running on Colab:\n",
|
||||
"# %xterm\n",
|
||||
"\n",
|
||||
"## To be ran in the terminal:\n",
|
||||
"# ollama serve &\n",
|
||||
"# ollama run llama3.2:3b --keepalive 60m"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "f3c5f243",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"If successful, you should see the model respond to a prompt.\n",
|
||||
"\n",
|
||||
"...\n",
|
||||
"```\n",
|
||||
">>> hi\n",
|
||||
"Hello! How can I assist you today?\n",
|
||||
"```"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "oDUB7M_qe-Gs",
|
||||
"metadata": {
|
||||
"id": "oDUB7M_qe-Gs"
|
||||
},
|
||||
"source": [
|
||||
"## Step 2: Run the Llama Stack server\n",
|
||||
"\n",
|
||||
"In this showcase, we will start a Llama Stack server that is running locally."
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "732eadc6",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"### 2.1. Setup the Llama Stack Server"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "J2kGed0R5PSf",
|
||||
"metadata": {
|
||||
"colab": {
|
||||
"base_uri": "https://localhost:8080/"
|
||||
},
|
||||
"collapsed": true,
|
||||
"id": "J2kGed0R5PSf",
|
||||
"outputId": "2478ea60-8d35-48a1-b011-f233831740c5"
|
||||
},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"import os\n",
|
||||
"import subprocess\n",
|
||||
"\n",
|
||||
"if \"UV_SYSTEM_PYTHON\" in os.environ:\n",
|
||||
" del os.environ[\"UV_SYSTEM_PYTHON\"]\n",
|
||||
"\n",
|
||||
"# this command installs all the dependencies needed for the llama stack server with the ollama inference provider\n",
|
||||
"!uv run --with llama-stack llama stack build --distro starter --image-type venv\n",
|
||||
"\n",
|
||||
"def run_llama_stack_server_background():\n",
|
||||
" log_file = open(\"llama_stack_server.log\", \"w\")\n",
|
||||
" process = subprocess.Popen(\n",
|
||||
" f\"OLLAMA_URL=http://localhost:11434 uv run --with llama-stack llama stack run starter --image-type venv\n",
|
||||
" shell=True,\n",
|
||||
" stdout=log_file,\n",
|
||||
" stderr=log_file,\n",
|
||||
" text=True\n",
|
||||
" )\n",
|
||||
"\n",
|
||||
" print(f\"Starting Llama Stack server with PID: {process.pid}\")\n",
|
||||
" return process\n",
|
||||
"\n",
|
||||
"def wait_for_server_to_start():\n",
|
||||
" import requests\n",
|
||||
" from requests.exceptions import ConnectionError\n",
|
||||
" import time\n",
|
||||
"\n",
|
||||
" url = \"http://0.0.0.0:8321/v1/health\"\n",
|
||||
" max_retries = 30\n",
|
||||
" retry_interval = 1\n",
|
||||
"\n",
|
||||
" print(\"Waiting for server to start\", end=\"\")\n",
|
||||
" for _ in range(max_retries):\n",
|
||||
" try:\n",
|
||||
" response = requests.get(url)\n",
|
||||
" if response.status_code == 200:\n",
|
||||
" print(\"\\nServer is ready!\")\n",
|
||||
" return True\n",
|
||||
" except ConnectionError:\n",
|
||||
" print(\".\", end=\"\", flush=True)\n",
|
||||
" time.sleep(retry_interval)\n",
|
||||
"\n",
|
||||
" print(\"\\nServer failed to start after\", max_retries * retry_interval, \"seconds\")\n",
|
||||
" return False\n",
|
||||
"\n",
|
||||
"\n",
|
||||
"# use this helper if needed to kill the server\n",
|
||||
"def kill_llama_stack_server():\n",
|
||||
" # Kill any existing llama stack server processes\n",
|
||||
" os.system(\"ps aux | grep -v grep | grep llama_stack.core.server.server | awk '{print $2}' | xargs kill -9\")\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "c40e9efd",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"### 2.2. Start the Llama Stack Server"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": 7,
|
||||
"id": "f779283d",
|
||||
"metadata": {},
|
||||
"outputs": [
|
||||
{
|
||||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"Starting Llama Stack server with PID: 787100\n",
|
||||
"Waiting for server to start\n",
|
||||
"Server is ready!\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"server_process = run_llama_stack_server_background()\n",
|
||||
"assert wait_for_server_to_start()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "28477c03",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"## Step 3: Run the demo"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": 8,
|
||||
"id": "7da71011",
|
||||
"metadata": {},
|
||||
"outputs": [
|
||||
{
|
||||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"rag_tool> Ingesting document: https://www.paulgraham.com/greatwork.html\n",
|
||||
"prompt> How do you do great work?\n",
|
||||
"\u001b[33minference> \u001b[0m\u001b[33m[k\u001b[0m\u001b[33mnowledge\u001b[0m\u001b[33m_search\u001b[0m\u001b[33m(query\u001b[0m\u001b[33m=\"\u001b[0m\u001b[33mWhat\u001b[0m\u001b[33m is\u001b[0m\u001b[33m the\u001b[0m\u001b[33m key\u001b[0m\u001b[33m to\u001b[0m\u001b[33m doing\u001b[0m\u001b[33m great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m\")]\u001b[0m\u001b[97m\u001b[0m\n",
|
||||
"\u001b[32mtool_execution> Tool:knowledge_search Args:{'query': 'What is the key to doing great work'}\u001b[0m\n",
|
||||
"\u001b[32mtool_execution> Tool:knowledge_search Response:[TextContentItem(text='knowledge_search tool found 5 chunks:\\nBEGIN of knowledge_search tool results.\\n', type='text'), TextContentItem(text=\"Result 1:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 2:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 3:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 4:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 5:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text='END of knowledge_search tool results.\\n', type='text'), TextContentItem(text='The above results were retrieved to help answer the user\\'s query: \"What is the key to doing great work\". Use them as supporting information only in answering this query.\\n', type='text')]\u001b[0m\n",
|
||||
"\u001b[33minference> \u001b[0m\u001b[33mDoing\u001b[0m\u001b[33m great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m means\u001b[0m\u001b[33m doing\u001b[0m\u001b[33m something\u001b[0m\u001b[33m important\u001b[0m\u001b[33m so\u001b[0m\u001b[33m well\u001b[0m\u001b[33m that\u001b[0m\u001b[33m you\u001b[0m\u001b[33m expand\u001b[0m\u001b[33m people\u001b[0m\u001b[33m's\u001b[0m\u001b[33m ideas\u001b[0m\u001b[33m of\u001b[0m\u001b[33m what\u001b[0m\u001b[33m's\u001b[0m\u001b[33m possible\u001b[0m\u001b[33m.\u001b[0m\u001b[33m However\u001b[0m\u001b[33m,\u001b[0m\u001b[33m there\u001b[0m\u001b[33m's\u001b[0m\u001b[33m no\u001b[0m\u001b[33m threshold\u001b[0m\u001b[33m for\u001b[0m\u001b[33m importance\u001b[0m\u001b[33m,\u001b[0m\u001b[33m and\u001b[0m\u001b[33m it\u001b[0m\u001b[33m's\u001b[0m\u001b[33m often\u001b[0m\u001b[33m hard\u001b[0m\u001b[33m to\u001b[0m\u001b[33m judge\u001b[0m\u001b[33m at\u001b[0m\u001b[33m the\u001b[0m\u001b[33m time\u001b[0m\u001b[33m anyway\u001b[0m\u001b[33m.\u001b[0m\u001b[33m Great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m is\u001b[0m\u001b[33m a\u001b[0m\u001b[33m matter\u001b[0m\u001b[33m of\u001b[0m\u001b[33m degree\u001b[0m\u001b[33m,\u001b[0m\u001b[33m and\u001b[0m\u001b[33m it\u001b[0m\u001b[33m can\u001b[0m\u001b[33m be\u001b[0m\u001b[33m difficult\u001b[0m\u001b[33m to\u001b[0m\u001b[33m determine\u001b[0m\u001b[33m whether\u001b[0m\u001b[33m someone\u001b[0m\u001b[33m has\u001b[0m\u001b[33m done\u001b[0m\u001b[33m great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m until\u001b[0m\u001b[33m after\u001b[0m\u001b[33m the\u001b[0m\u001b[33m fact\u001b[0m\u001b[33m.\u001b[0m\u001b[97m\u001b[0m\n",
|
||||
"\u001b[30m\u001b[0m"
|
||||
]
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"from llama_stack_client import Agent, AgentEventLogger, RAGDocument, LlamaStackClient\n",
|
||||
"\n",
|
||||
"vector_db_id = \"my_demo_vector_db\"\n",
|
||||
"client = LlamaStackClient(base_url=\"http://0.0.0.0:8321\")\n",
|
||||
"\n",
|
||||
"models = client.models.list()\n",
|
||||
"\n",
|
||||
"# Select the first ollama and first ollama's embedding model\n",
|
||||
"model_id = next(m for m in models if m.model_type == \"llm\" and m.provider_id == \"ollama\").identifier\n",
|
||||
"embedding_model = next(m for m in models if m.model_type == \"embedding\" and m.provider_id == \"ollama\")\n",
|
||||
"embedding_model_id = embedding_model.identifier\n",
|
||||
"embedding_dimension = embedding_model.metadata[\"embedding_dimension\"]\n",
|
||||
"\n",
|
||||
"_ = client.vector_dbs.register(\n",
|
||||
" vector_db_id=vector_db_id,\n",
|
||||
" embedding_model=embedding_model_id,\n",
|
||||
" embedding_dimension=embedding_dimension,\n",
|
||||
" provider_id=\"faiss\",\n",
|
||||
")\n",
|
||||
"source = \"https://www.paulgraham.com/greatwork.html\"\n",
|
||||
"print(\"rag_tool> Ingesting document:\", source)\n",
|
||||
"document = RAGDocument(\n",
|
||||
" document_id=\"document_1\",\n",
|
||||
" content=source,\n",
|
||||
" mime_type=\"text/html\",\n",
|
||||
" metadata={},\n",
|
||||
")\n",
|
||||
"client.tool_runtime.rag_tool.insert(\n",
|
||||
" documents=[document],\n",
|
||||
" vector_db_id=vector_db_id,\n",
|
||||
" chunk_size_in_tokens=50,\n",
|
||||
")\n",
|
||||
"agent = Agent(\n",
|
||||
" client,\n",
|
||||
" model=model_id,\n",
|
||||
" instructions=\"You are a helpful assistant\",\n",
|
||||
" tools=[\n",
|
||||
" {\n",
|
||||
" \"name\": \"builtin::rag/knowledge_search\",\n",
|
||||
" \"args\": {\"vector_db_ids\": [vector_db_id]},\n",
|
||||
" }\n",
|
||||
" ],\n",
|
||||
")\n",
|
||||
"\n",
|
||||
"prompt = \"How do you do great work?\"\n",
|
||||
"print(\"prompt>\", prompt)\n",
|
||||
"\n",
|
||||
"response = agent.create_turn(\n",
|
||||
" messages=[{\"role\": \"user\", \"content\": prompt}],\n",
|
||||
" session_id=agent.create_session(\"rag_session\"),\n",
|
||||
" stream=True,\n",
|
||||
")\n",
|
||||
"\n",
|
||||
"for log in AgentEventLogger().log(response):\n",
|
||||
" log.print()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "341aaadf",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"Congratulations! You've successfully built your first RAG application using Llama Stack! 🎉🥳"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "e88e1185",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"## Next Steps"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "bcb73600",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"Now you're ready to dive deeper into Llama Stack!\n",
|
||||
"- Explore the [Detailed Tutorial](./detailed_tutorial.md).\n",
|
||||
"- Try the [Getting Started Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb).\n",
|
||||
"- Browse more [Notebooks on GitHub](https://github.com/meta-llama/llama-stack/tree/main/docs/notebooks).\n",
|
||||
"- Learn about Llama Stack [Concepts](../concepts/index.md).\n",
|
||||
"- Discover how to [Build Llama Stacks](../distributions/index.md).\n",
|
||||
"- Refer to our [References](../references/index.md) for details on the Llama CLI and Python SDK.\n",
|
||||
"- Check out the [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repository for example applications and tutorials."
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"accelerator": "GPU",
|
||||
"colab": {
|
||||
"gpuType": "T4",
|
||||
"provenance": []
|
||||
},
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.10.6"
|
||||
}
|
||||
"cells": [
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "c1e7571c",
|
||||
"metadata": {
|
||||
"id": "c1e7571c"
|
||||
},
|
||||
"source": [
|
||||
"[](https://colab.research.google.com/github/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb)\n",
|
||||
"\n",
|
||||
"# Llama Stack - Building AI Applications\n",
|
||||
"\n",
|
||||
"<img src=\"https://llamastack.github.io/latest/_images/llama-stack.png\" alt=\"drawing\" width=\"500\"/>\n",
|
||||
"\n",
|
||||
"Get started with Llama Stack in minutes!\n",
|
||||
"\n",
|
||||
"[Llama Stack](https://github.com/meta-llama/llama-stack) is a stateful service with REST APIs to support the seamless transition of AI applications across different environments. You can build and test using a local server first and deploy to a hosted endpoint for production.\n",
|
||||
"\n",
|
||||
"In this guide, we'll walk through how to build a RAG application locally using Llama Stack with [Ollama](https://ollama.com/)\n",
|
||||
"as the inference [provider](docs/source/providers/index.md#inference) for a Llama Model.\n"
|
||||
]
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "4CV1Q19BDMVw",
|
||||
"metadata": {
|
||||
"id": "4CV1Q19BDMVw"
|
||||
},
|
||||
"source": [
|
||||
"## Step 1: Install and setup"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "K4AvfUAJZOeS",
|
||||
"metadata": {
|
||||
"id": "K4AvfUAJZOeS"
|
||||
},
|
||||
"source": [
|
||||
"### 1.1. Install uv and test inference with Ollama\n",
|
||||
"\n",
|
||||
"We'll install [uv](https://docs.astral.sh/uv/) to setup the Python virtual environment, along with [colab-xterm](https://github.com/InfuseAI/colab-xterm) for running command-line tools, and [Ollama](https://ollama.com/download) as the inference provider."
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "7a2d7b85",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"%pip install uv llama_stack llama-stack-client\n",
|
||||
"\n",
|
||||
"## If running on Collab:\n",
|
||||
"# !pip install colab-xterm\n",
|
||||
"# %load_ext colabxterm\n",
|
||||
"\n",
|
||||
"!curl https://ollama.ai/install.sh | sh"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "39fa584b",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"### 1.2. Test inference with Ollama"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "3bf81522",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"We’ll now launch a terminal and run inference on a Llama model with Ollama to verify that the model is working correctly."
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "a7e8e0f1",
|
||||
"metadata": {},
|
||||
"outputs": [],
|
||||
"source": [
|
||||
"## If running on Colab:\n",
|
||||
"# %xterm\n",
|
||||
"\n",
|
||||
"## To be ran in the terminal:\n",
|
||||
"# ollama serve &\n",
|
||||
"# ollama run llama3.2:3b --keepalive 60m"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "f3c5f243",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"If successful, you should see the model respond to a prompt.\n",
|
||||
"\n",
|
||||
"...\n",
|
||||
"```\n",
|
||||
">>> hi\n",
|
||||
"Hello! How can I assist you today?\n",
|
||||
"```"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "oDUB7M_qe-Gs",
|
||||
"metadata": {
|
||||
"id": "oDUB7M_qe-Gs"
|
||||
},
|
||||
"source": [
|
||||
"## Step 2: Run the Llama Stack server\n",
|
||||
"\n",
|
||||
"In this showcase, we will start a Llama Stack server that is running locally."
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "732eadc6",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"### 2.1. Setup the Llama Stack Server"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": null,
|
||||
"id": "J2kGed0R5PSf",
|
||||
"metadata": {
|
||||
"colab": {
|
||||
"base_uri": "https://localhost:8080/"
|
||||
},
|
||||
"collapsed": true,
|
||||
"id": "J2kGed0R5PSf",
|
||||
"outputId": "2478ea60-8d35-48a1-b011-f233831740c5"
|
||||
},
|
||||
"outputs": [],
|
||||
"source": "import os\nimport subprocess\n\nif \"UV_SYSTEM_PYTHON\" in os.environ:\n del os.environ[\"UV_SYSTEM_PYTHON\"]\n\n# this command installs all the dependencies needed for the llama stack server with the ollama inference provider\n!uv run --with llama-stack llama stack build --distro starter --image-type venv\n\ndef run_llama_stack_server_background():\n log_file = open(\"llama_stack_server.log\", \"w\")\n process = subprocess.Popen(\n f\"OLLAMA_URL=http://localhost:11434 uv run --with llama-stack llama stack run starter\",\n shell=True,\n stdout=log_file,\n stderr=log_file,\n text=True\n )\n\n print(f\"Starting Llama Stack server with PID: {process.pid}\")\n return process\n\ndef wait_for_server_to_start():\n import requests\n from requests.exceptions import ConnectionError\n import time\n\n url = \"http://0.0.0.0:8321/v1/health\"\n max_retries = 30\n retry_interval = 1\n\n print(\"Waiting for server to start\", end=\"\")\n for _ in range(max_retries):\n try:\n response = requests.get(url)\n if response.status_code == 200:\n print(\"\\nServer is ready!\")\n return True\n except ConnectionError:\n print(\".\", end=\"\", flush=True)\n time.sleep(retry_interval)\n\n print(\"\\nServer failed to start after\", max_retries * retry_interval, \"seconds\")\n return False\n\n\n# use this helper if needed to kill the server\ndef kill_llama_stack_server():\n # Kill any existing llama stack server processes\n os.system(\"ps aux | grep -v grep | grep llama_stack.core.server.server | awk '{print $2}' | xargs kill -9\")\n"
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "c40e9efd",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"### 2.2. Start the Llama Stack Server"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": 7,
|
||||
"id": "f779283d",
|
||||
"metadata": {},
|
||||
"outputs": [
|
||||
{
|
||||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"Starting Llama Stack server with PID: 787100\n",
|
||||
"Waiting for server to start\n",
|
||||
"Server is ready!\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"server_process = run_llama_stack_server_background()\n",
|
||||
"assert wait_for_server_to_start()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "28477c03",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"## Step 3: Run the demo"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "code",
|
||||
"execution_count": 8,
|
||||
"id": "7da71011",
|
||||
"metadata": {},
|
||||
"outputs": [
|
||||
{
|
||||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"rag_tool> Ingesting document: https://www.paulgraham.com/greatwork.html\n",
|
||||
"prompt> How do you do great work?\n",
|
||||
"\u001b[33minference> \u001b[0m\u001b[33m[k\u001b[0m\u001b[33mnowledge\u001b[0m\u001b[33m_search\u001b[0m\u001b[33m(query\u001b[0m\u001b[33m=\"\u001b[0m\u001b[33mWhat\u001b[0m\u001b[33m is\u001b[0m\u001b[33m the\u001b[0m\u001b[33m key\u001b[0m\u001b[33m to\u001b[0m\u001b[33m doing\u001b[0m\u001b[33m great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m\")]\u001b[0m\u001b[97m\u001b[0m\n",
|
||||
"\u001b[32mtool_execution> Tool:knowledge_search Args:{'query': 'What is the key to doing great work'}\u001b[0m\n",
|
||||
"\u001b[32mtool_execution> Tool:knowledge_search Response:[TextContentItem(text='knowledge_search tool found 5 chunks:\\nBEGIN of knowledge_search tool results.\\n', type='text'), TextContentItem(text=\"Result 1:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 2:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 3:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 4:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text=\"Result 5:\\nDocument_id:docum\\nContent: work. Doing great work means doing something important\\nso well that you expand people's ideas of what's possible. But\\nthere's no threshold for importance. It's a matter of degree, and\\noften hard to judge at the time anyway.\\n\", type='text'), TextContentItem(text='END of knowledge_search tool results.\\n', type='text'), TextContentItem(text='The above results were retrieved to help answer the user\\'s query: \"What is the key to doing great work\". Use them as supporting information only in answering this query.\\n', type='text')]\u001b[0m\n",
|
||||
"\u001b[33minference> \u001b[0m\u001b[33mDoing\u001b[0m\u001b[33m great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m means\u001b[0m\u001b[33m doing\u001b[0m\u001b[33m something\u001b[0m\u001b[33m important\u001b[0m\u001b[33m so\u001b[0m\u001b[33m well\u001b[0m\u001b[33m that\u001b[0m\u001b[33m you\u001b[0m\u001b[33m expand\u001b[0m\u001b[33m people\u001b[0m\u001b[33m's\u001b[0m\u001b[33m ideas\u001b[0m\u001b[33m of\u001b[0m\u001b[33m what\u001b[0m\u001b[33m's\u001b[0m\u001b[33m possible\u001b[0m\u001b[33m.\u001b[0m\u001b[33m However\u001b[0m\u001b[33m,\u001b[0m\u001b[33m there\u001b[0m\u001b[33m's\u001b[0m\u001b[33m no\u001b[0m\u001b[33m threshold\u001b[0m\u001b[33m for\u001b[0m\u001b[33m importance\u001b[0m\u001b[33m,\u001b[0m\u001b[33m and\u001b[0m\u001b[33m it\u001b[0m\u001b[33m's\u001b[0m\u001b[33m often\u001b[0m\u001b[33m hard\u001b[0m\u001b[33m to\u001b[0m\u001b[33m judge\u001b[0m\u001b[33m at\u001b[0m\u001b[33m the\u001b[0m\u001b[33m time\u001b[0m\u001b[33m anyway\u001b[0m\u001b[33m.\u001b[0m\u001b[33m Great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m is\u001b[0m\u001b[33m a\u001b[0m\u001b[33m matter\u001b[0m\u001b[33m of\u001b[0m\u001b[33m degree\u001b[0m\u001b[33m,\u001b[0m\u001b[33m and\u001b[0m\u001b[33m it\u001b[0m\u001b[33m can\u001b[0m\u001b[33m be\u001b[0m\u001b[33m difficult\u001b[0m\u001b[33m to\u001b[0m\u001b[33m determine\u001b[0m\u001b[33m whether\u001b[0m\u001b[33m someone\u001b[0m\u001b[33m has\u001b[0m\u001b[33m done\u001b[0m\u001b[33m great\u001b[0m\u001b[33m work\u001b[0m\u001b[33m until\u001b[0m\u001b[33m after\u001b[0m\u001b[33m the\u001b[0m\u001b[33m fact\u001b[0m\u001b[33m.\u001b[0m\u001b[97m\u001b[0m\n",
|
||||
"\u001b[30m\u001b[0m"
|
||||
]
|
||||
}
|
||||
],
|
||||
"source": [
|
||||
"from llama_stack_client import Agent, AgentEventLogger, RAGDocument, LlamaStackClient\n",
|
||||
"\n",
|
||||
"vector_db_id = \"my_demo_vector_db\"\n",
|
||||
"client = LlamaStackClient(base_url=\"http://0.0.0.0:8321\")\n",
|
||||
"\n",
|
||||
"models = client.models.list()\n",
|
||||
"\n",
|
||||
"# Select the first ollama and first ollama's embedding model\n",
|
||||
"model_id = next(m for m in models if m.model_type == \"llm\" and m.provider_id == \"ollama\").identifier\n",
|
||||
"embedding_model = next(m for m in models if m.model_type == \"embedding\" and m.provider_id == \"ollama\")\n",
|
||||
"embedding_model_id = embedding_model.identifier\n",
|
||||
"embedding_dimension = embedding_model.metadata[\"embedding_dimension\"]\n",
|
||||
"\n",
|
||||
"_ = client.vector_dbs.register(\n",
|
||||
" vector_db_id=vector_db_id,\n",
|
||||
" embedding_model=embedding_model_id,\n",
|
||||
" embedding_dimension=embedding_dimension,\n",
|
||||
" provider_id=\"faiss\",\n",
|
||||
")\n",
|
||||
"source = \"https://www.paulgraham.com/greatwork.html\"\n",
|
||||
"print(\"rag_tool> Ingesting document:\", source)\n",
|
||||
"document = RAGDocument(\n",
|
||||
" document_id=\"document_1\",\n",
|
||||
" content=source,\n",
|
||||
" mime_type=\"text/html\",\n",
|
||||
" metadata={},\n",
|
||||
")\n",
|
||||
"client.tool_runtime.rag_tool.insert(\n",
|
||||
" documents=[document],\n",
|
||||
" vector_db_id=vector_db_id,\n",
|
||||
" chunk_size_in_tokens=50,\n",
|
||||
")\n",
|
||||
"agent = Agent(\n",
|
||||
" client,\n",
|
||||
" model=model_id,\n",
|
||||
" instructions=\"You are a helpful assistant\",\n",
|
||||
" tools=[\n",
|
||||
" {\n",
|
||||
" \"name\": \"builtin::rag/knowledge_search\",\n",
|
||||
" \"args\": {\"vector_db_ids\": [vector_db_id]},\n",
|
||||
" }\n",
|
||||
" ],\n",
|
||||
")\n",
|
||||
"\n",
|
||||
"prompt = \"How do you do great work?\"\n",
|
||||
"print(\"prompt>\", prompt)\n",
|
||||
"\n",
|
||||
"response = agent.create_turn(\n",
|
||||
" messages=[{\"role\": \"user\", \"content\": prompt}],\n",
|
||||
" session_id=agent.create_session(\"rag_session\"),\n",
|
||||
" stream=True,\n",
|
||||
")\n",
|
||||
"\n",
|
||||
"for log in AgentEventLogger().log(response):\n",
|
||||
" log.print()"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "341aaadf",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"Congratulations! You've successfully built your first RAG application using Llama Stack! 🎉🥳"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "e88e1185",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"## Next Steps"
|
||||
]
|
||||
},
|
||||
{
|
||||
"cell_type": "markdown",
|
||||
"id": "bcb73600",
|
||||
"metadata": {},
|
||||
"source": [
|
||||
"Now you're ready to dive deeper into Llama Stack!\n",
|
||||
"- Explore the [Detailed Tutorial](./detailed_tutorial.md).\n",
|
||||
"- Try the [Getting Started Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/getting_started.ipynb).\n",
|
||||
"- Browse more [Notebooks on GitHub](https://github.com/meta-llama/llama-stack/tree/main/docs/notebooks).\n",
|
||||
"- Learn about Llama Stack [Concepts](../concepts/index.md).\n",
|
||||
"- Discover how to [Build Llama Stacks](../distributions/index.md).\n",
|
||||
"- Refer to our [References](../references/index.md) for details on the Llama CLI and Python SDK.\n",
|
||||
"- Check out the [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repository for example applications and tutorials."
|
||||
]
|
||||
}
|
||||
],
|
||||
"metadata": {
|
||||
"accelerator": "GPU",
|
||||
"colab": {
|
||||
"gpuType": "T4",
|
||||
"provenance": []
|
||||
},
|
||||
"kernelspec": {
|
||||
"display_name": "Python 3",
|
||||
"language": "python",
|
||||
"name": "python3"
|
||||
},
|
||||
"language_info": {
|
||||
"codemirror_mode": {
|
||||
"name": "ipython",
|
||||
"version": 3
|
||||
},
|
||||
"file_extension": ".py",
|
||||
"mimetype": "text/x-python",
|
||||
"name": "python",
|
||||
"nbconvert_exporter": "python",
|
||||
"pygments_lexer": "ipython3",
|
||||
"version": "3.10.6"
|
||||
}
|
||||
},
|
||||
"nbformat": 4,
|
||||
"nbformat_minor": 5
|
||||
}
|
||||
|
|
|
@ -88,7 +88,7 @@ If you're looking for more specific topics, we have a [Zero to Hero Guide](#next
|
|||
...
|
||||
Build Successful!
|
||||
You can find the newly-built template here: ~/.llama/distributions/starter/starter-run.yaml
|
||||
You can run the new Llama Stack Distro via: uv run --with llama-stack llama stack run starter --image-type venv
|
||||
You can run the new Llama Stack Distro via: uv run --with llama-stack llama stack run starter
|
||||
```
|
||||
|
||||
3. **Set the ENV variables by exporting them to the terminal**:
|
||||
|
@ -102,12 +102,11 @@ If you're looking for more specific topics, we have a [Zero to Hero Guide](#next
|
|||
3. **Run the Llama Stack**:
|
||||
Run the stack using uv:
|
||||
```bash
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
SAFETY_MODEL=$SAFETY_MODEL \
|
||||
OLLAMA_URL=$OLLAMA_URL \
|
||||
uv run --with llama-stack llama stack run starter \
|
||||
--image-type venv \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env SAFETY_MODEL=$SAFETY_MODEL \
|
||||
--env OLLAMA_URL=$OLLAMA_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
Note: Every time you run a new model with `ollama run`, you will need to restart the llama stack. Otherwise it won't see the new model.
|
||||
|
||||
|
|
|
@ -444,12 +444,24 @@ def _run_stack_build_command_from_build_config(
|
|||
|
||||
cprint("Build Successful!", color="green", file=sys.stderr)
|
||||
cprint(f"You can find the newly-built distribution here: {run_config_file}", color="blue", file=sys.stderr)
|
||||
cprint(
|
||||
"You can run the new Llama Stack distro via: "
|
||||
+ colored(f"llama stack run {run_config_file} --image-type {build_config.image_type}", "blue"),
|
||||
color="green",
|
||||
file=sys.stderr,
|
||||
)
|
||||
if build_config.image_type == LlamaStackImageType.VENV:
|
||||
cprint(
|
||||
"You can run the new Llama Stack distro (after activating "
|
||||
+ colored(image_name, "cyan")
|
||||
+ ") via: "
|
||||
+ colored(f"llama stack run {run_config_file}", "blue"),
|
||||
color="green",
|
||||
file=sys.stderr,
|
||||
)
|
||||
elif build_config.image_type == LlamaStackImageType.CONTAINER:
|
||||
cprint(
|
||||
"You can run the container with: "
|
||||
+ colored(
|
||||
f"docker run -p 8321:8321 -v ~/.llama:/root/.llama localhost/{image_name} --port 8321", "blue"
|
||||
),
|
||||
color="green",
|
||||
file=sys.stderr,
|
||||
)
|
||||
return distro_path
|
||||
else:
|
||||
return _generate_run_config(build_config, build_dir, image_name)
|
||||
|
|
|
@ -16,7 +16,7 @@ import yaml
|
|||
from llama_stack.cli.stack.utils import ImageType
|
||||
from llama_stack.cli.subcommand import Subcommand
|
||||
from llama_stack.core.datatypes import LoggingConfig, StackRunConfig
|
||||
from llama_stack.core.stack import cast_image_name_to_string, replace_env_vars, validate_env_pair
|
||||
from llama_stack.core.stack import cast_image_name_to_string, replace_env_vars
|
||||
from llama_stack.core.utils.config_resolution import Mode, resolve_config_or_distro
|
||||
from llama_stack.log import get_logger
|
||||
|
||||
|
@ -55,18 +55,12 @@ class StackRun(Subcommand):
|
|||
"--image-name",
|
||||
type=str,
|
||||
default=None,
|
||||
help="Name of the image to run. Defaults to the current environment",
|
||||
)
|
||||
self.parser.add_argument(
|
||||
"--env",
|
||||
action="append",
|
||||
help="Environment variables to pass to the server in KEY=VALUE format. Can be specified multiple times.",
|
||||
metavar="KEY=VALUE",
|
||||
help="[DEPRECATED] This flag is no longer supported. Please activate your virtual environment before running.",
|
||||
)
|
||||
self.parser.add_argument(
|
||||
"--image-type",
|
||||
type=str,
|
||||
help="Image Type used during the build. This can be only venv.",
|
||||
help="[DEPRECATED] This flag is no longer supported. Please activate your virtual environment before running.",
|
||||
choices=[e.value for e in ImageType if e.value != ImageType.CONTAINER.value],
|
||||
)
|
||||
self.parser.add_argument(
|
||||
|
@ -75,48 +69,22 @@ class StackRun(Subcommand):
|
|||
help="Start the UI server",
|
||||
)
|
||||
|
||||
def _resolve_config_and_distro(self, args: argparse.Namespace) -> tuple[Path | None, str | None]:
|
||||
"""Resolve config file path and distribution name from args.config"""
|
||||
from llama_stack.core.utils.config_dirs import DISTRIBS_BASE_DIR
|
||||
|
||||
if not args.config:
|
||||
return None, None
|
||||
|
||||
config_file = Path(args.config)
|
||||
has_yaml_suffix = args.config.endswith(".yaml")
|
||||
distro_name = None
|
||||
|
||||
if not config_file.exists() and not has_yaml_suffix:
|
||||
# check if this is a distribution
|
||||
config_file = Path(REPO_ROOT) / "llama_stack" / "distributions" / args.config / "run.yaml"
|
||||
if config_file.exists():
|
||||
distro_name = args.config
|
||||
|
||||
if not config_file.exists() and not has_yaml_suffix:
|
||||
# check if it's a build config saved to ~/.llama dir
|
||||
config_file = Path(DISTRIBS_BASE_DIR / f"llamastack-{args.config}" / f"{args.config}-run.yaml")
|
||||
|
||||
if not config_file.exists():
|
||||
self.parser.error(
|
||||
f"File {str(config_file)} does not exist.\n\nPlease run `llama stack build` to generate (and optionally edit) a run.yaml file"
|
||||
)
|
||||
|
||||
if not config_file.is_file():
|
||||
self.parser.error(
|
||||
f"Config file must be a valid file path, '{config_file}' is not a file: type={type(config_file)}"
|
||||
)
|
||||
|
||||
return config_file, distro_name
|
||||
|
||||
def _run_stack_run_cmd(self, args: argparse.Namespace) -> None:
|
||||
import yaml
|
||||
|
||||
from llama_stack.core.configure import parse_and_maybe_upgrade_config
|
||||
from llama_stack.core.utils.exec import formulate_run_args, run_command
|
||||
|
||||
if args.image_type or args.image_name:
|
||||
self.parser.error(
|
||||
"The --image-type and --image-name flags are no longer supported.\n\n"
|
||||
"Please activate your virtual environment manually before running `llama stack run`.\n\n"
|
||||
"For example:\n"
|
||||
" source /path/to/venv/bin/activate\n"
|
||||
" llama stack run <config>\n"
|
||||
)
|
||||
|
||||
if args.enable_ui:
|
||||
self._start_ui_development_server(args.port)
|
||||
image_type, image_name = args.image_type, args.image_name
|
||||
|
||||
if args.config:
|
||||
try:
|
||||
|
@ -128,10 +96,6 @@ class StackRun(Subcommand):
|
|||
else:
|
||||
config_file = None
|
||||
|
||||
# Check if config is required based on image type
|
||||
if image_type == ImageType.VENV.value and not config_file:
|
||||
self.parser.error("Config file is required for venv environment")
|
||||
|
||||
if config_file:
|
||||
logger.info(f"Using run configuration: {config_file}")
|
||||
|
||||
|
@ -146,50 +110,13 @@ class StackRun(Subcommand):
|
|||
os.makedirs(str(config.external_providers_dir), exist_ok=True)
|
||||
except AttributeError as e:
|
||||
self.parser.error(f"failed to parse config file '{config_file}':\n {e}")
|
||||
else:
|
||||
config = None
|
||||
|
||||
# If neither image type nor image name is provided, assume the server should be run directly
|
||||
# using the current environment packages.
|
||||
if not image_type and not image_name:
|
||||
logger.info("No image type or image name provided. Assuming environment packages.")
|
||||
self._uvicorn_run(config_file, args)
|
||||
else:
|
||||
run_args = formulate_run_args(image_type, image_name)
|
||||
|
||||
run_args.extend([str(args.port)])
|
||||
|
||||
if config_file:
|
||||
run_args.extend(["--config", str(config_file)])
|
||||
|
||||
if args.env:
|
||||
for env_var in args.env:
|
||||
if "=" not in env_var:
|
||||
self.parser.error(f"Environment variable '{env_var}' must be in KEY=VALUE format")
|
||||
return
|
||||
key, value = env_var.split("=", 1) # split on first = only
|
||||
if not key:
|
||||
self.parser.error(f"Environment variable '{env_var}' has empty key")
|
||||
return
|
||||
run_args.extend(["--env", f"{key}={value}"])
|
||||
|
||||
run_command(run_args)
|
||||
self._uvicorn_run(config_file, args)
|
||||
|
||||
def _uvicorn_run(self, config_file: Path | None, args: argparse.Namespace) -> None:
|
||||
if not config_file:
|
||||
self.parser.error("Config file is required")
|
||||
|
||||
# Set environment variables if provided
|
||||
if args.env:
|
||||
for env_pair in args.env:
|
||||
try:
|
||||
key, value = validate_env_pair(env_pair)
|
||||
logger.info(f"Setting environment variable {key} => {value}")
|
||||
os.environ[key] = value
|
||||
except ValueError as e:
|
||||
logger.error(f"Error: {str(e)}")
|
||||
self.parser.error(f"Invalid environment variable format: {env_pair}")
|
||||
|
||||
config_file = resolve_config_or_distro(str(config_file), Mode.RUN)
|
||||
with open(config_file) as fp:
|
||||
config_contents = yaml.safe_load(fp)
|
||||
|
|
|
@ -33,7 +33,7 @@ class ModelsRoutingTable(CommonRoutingTableImpl, Models):
|
|||
try:
|
||||
models = await provider.list_models()
|
||||
except Exception as e:
|
||||
logger.warning(f"Model refresh failed for provider {provider_id}: {e}")
|
||||
logger.debug(f"Model refresh failed for provider {provider_id}: {e}")
|
||||
continue
|
||||
|
||||
self.listed_providers.add(provider_id)
|
||||
|
|
|
@ -245,3 +245,65 @@ class VectorDBsRoutingTable(CommonRoutingTableImpl, VectorDBs):
|
|||
vector_store_id=vector_store_id,
|
||||
file_id=file_id,
|
||||
)
|
||||
|
||||
async def openai_create_vector_store_file_batch(
|
||||
self,
|
||||
vector_store_id: str,
|
||||
file_ids: list[str],
|
||||
attributes: dict[str, Any] | None = None,
|
||||
chunking_strategy: Any | None = None,
|
||||
):
|
||||
await self.assert_action_allowed("update", "vector_db", vector_store_id)
|
||||
provider = await self.get_provider_impl(vector_store_id)
|
||||
return await provider.openai_create_vector_store_file_batch(
|
||||
vector_store_id=vector_store_id,
|
||||
file_ids=file_ids,
|
||||
attributes=attributes,
|
||||
chunking_strategy=chunking_strategy,
|
||||
)
|
||||
|
||||
async def openai_retrieve_vector_store_file_batch(
|
||||
self,
|
||||
batch_id: str,
|
||||
vector_store_id: str,
|
||||
):
|
||||
await self.assert_action_allowed("read", "vector_db", vector_store_id)
|
||||
provider = await self.get_provider_impl(vector_store_id)
|
||||
return await provider.openai_retrieve_vector_store_file_batch(
|
||||
batch_id=batch_id,
|
||||
vector_store_id=vector_store_id,
|
||||
)
|
||||
|
||||
async def openai_list_files_in_vector_store_file_batch(
|
||||
self,
|
||||
batch_id: str,
|
||||
vector_store_id: str,
|
||||
after: str | None = None,
|
||||
before: str | None = None,
|
||||
filter: str | None = None,
|
||||
limit: int | None = 20,
|
||||
order: str | None = "desc",
|
||||
):
|
||||
await self.assert_action_allowed("read", "vector_db", vector_store_id)
|
||||
provider = await self.get_provider_impl(vector_store_id)
|
||||
return await provider.openai_list_files_in_vector_store_file_batch(
|
||||
batch_id=batch_id,
|
||||
vector_store_id=vector_store_id,
|
||||
after=after,
|
||||
before=before,
|
||||
filter=filter,
|
||||
limit=limit,
|
||||
order=order,
|
||||
)
|
||||
|
||||
async def openai_cancel_vector_store_file_batch(
|
||||
self,
|
||||
batch_id: str,
|
||||
vector_store_id: str,
|
||||
):
|
||||
await self.assert_action_allowed("update", "vector_db", vector_store_id)
|
||||
provider = await self.get_provider_impl(vector_store_id)
|
||||
return await provider.openai_cancel_vector_store_file_batch(
|
||||
batch_id=batch_id,
|
||||
vector_store_id=vector_store_id,
|
||||
)
|
||||
|
|
|
@ -274,22 +274,6 @@ def cast_image_name_to_string(config_dict: dict[str, Any]) -> dict[str, Any]:
|
|||
return config_dict
|
||||
|
||||
|
||||
def validate_env_pair(env_pair: str) -> tuple[str, str]:
|
||||
"""Validate and split an environment variable key-value pair."""
|
||||
try:
|
||||
key, value = env_pair.split("=", 1)
|
||||
key = key.strip()
|
||||
if not key:
|
||||
raise ValueError(f"Empty key in environment variable pair: {env_pair}")
|
||||
if not all(c.isalnum() or c == "_" for c in key):
|
||||
raise ValueError(f"Key must contain only alphanumeric characters and underscores: {key}")
|
||||
return key, value
|
||||
except ValueError as e:
|
||||
raise ValueError(
|
||||
f"Invalid environment variable format '{env_pair}': {str(e)}. Expected format: KEY=value"
|
||||
) from e
|
||||
|
||||
|
||||
def add_internal_implementations(impls: dict[Api, Any], run_config: StackRunConfig) -> None:
|
||||
"""Add internal implementations (inspect and providers) to the implementations dictionary.
|
||||
|
||||
|
|
|
@ -25,7 +25,7 @@ error_handler() {
|
|||
trap 'error_handler ${LINENO}' ERR
|
||||
|
||||
if [ $# -lt 3 ]; then
|
||||
echo "Usage: $0 <env_type> <env_path_or_name> <port> [--config <yaml_config>] [--env KEY=VALUE]..."
|
||||
echo "Usage: $0 <env_type> <env_path_or_name> <port> [--config <yaml_config>]"
|
||||
exit 1
|
||||
fi
|
||||
|
||||
|
@ -43,7 +43,6 @@ SCRIPT_DIR=$(dirname "$(readlink -f "$0")")
|
|||
|
||||
# Initialize variables
|
||||
yaml_config=""
|
||||
env_vars=""
|
||||
other_args=""
|
||||
|
||||
# Process remaining arguments
|
||||
|
@ -58,15 +57,6 @@ while [[ $# -gt 0 ]]; do
|
|||
exit 1
|
||||
fi
|
||||
;;
|
||||
--env)
|
||||
if [[ -n "$2" ]]; then
|
||||
env_vars="$env_vars --env $2"
|
||||
shift 2
|
||||
else
|
||||
echo -e "${RED}Error: --env requires a KEY=VALUE argument${NC}" >&2
|
||||
exit 1
|
||||
fi
|
||||
;;
|
||||
*)
|
||||
other_args="$other_args $1"
|
||||
shift
|
||||
|
@ -119,7 +109,6 @@ if [[ "$env_type" == "venv" ]]; then
|
|||
llama stack run \
|
||||
$yaml_config_arg \
|
||||
--port "$port" \
|
||||
$env_vars \
|
||||
$other_args
|
||||
elif [[ "$env_type" == "container" ]]; then
|
||||
echo -e "${RED}Warning: Llama Stack no longer supports running Containers via the 'llama stack run' command.${NC}"
|
||||
|
|
|
@ -117,11 +117,11 @@ docker run -it \
|
|||
# NOTE: mount the llama-stack directory if testing local changes else not needed
|
||||
-v $HOME/git/llama-stack:/app/llama-stack-source \
|
||||
# localhost/distribution-dell:dev if building / testing locally
|
||||
-e INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
-e DEH_URL=$DEH_URL \
|
||||
-e CHROMA_URL=$CHROMA_URL \
|
||||
llamastack/distribution-{{ name }}\
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
|
||||
```
|
||||
|
||||
|
@ -142,14 +142,14 @@ docker run \
|
|||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v $HOME/.llama:/root/.llama \
|
||||
-v ./llama_stack/distributions/tgi/run-with-safety.yaml:/root/my-run.yaml \
|
||||
-e INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
-e DEH_URL=$DEH_URL \
|
||||
-e SAFETY_MODEL=$SAFETY_MODEL \
|
||||
-e DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
-e CHROMA_URL=$CHROMA_URL \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--config /root/my-run.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env SAFETY_MODEL=$SAFETY_MODEL \
|
||||
--env DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
### Via Conda
|
||||
|
@ -158,21 +158,21 @@ Make sure you have done `pip install llama-stack` and have the Llama Stack CLI a
|
|||
|
||||
```bash
|
||||
llama stack build --distro {{ name }} --image-type conda
|
||||
llama stack run {{ name }}
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
DEH_URL=$DEH_URL \
|
||||
CHROMA_URL=$CHROMA_URL \
|
||||
llama stack run {{ name }} \
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
||||
```bash
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
DEH_URL=$DEH_URL \
|
||||
SAFETY_MODEL=$SAFETY_MODEL \
|
||||
DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
CHROMA_URL=$CHROMA_URL \
|
||||
llama stack run ./run-with-safety.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
--env DEH_URL=$DEH_URL \
|
||||
--env SAFETY_MODEL=$SAFETY_MODEL \
|
||||
--env DEH_SAFETY_URL=$DEH_SAFETY_URL \
|
||||
--env CHROMA_URL=$CHROMA_URL
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
|
|
@ -72,9 +72,9 @@ docker run \
|
|||
--gpu all \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
-e INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
@ -86,10 +86,10 @@ docker run \
|
|||
--gpu all \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ~/.llama:/root/.llama \
|
||||
-e INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
-e SAFETY_MODEL=meta-llama/Llama-Guard-3-1B \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
### Via venv
|
||||
|
@ -98,16 +98,16 @@ Make sure you have done `uv pip install llama-stack` and have the Llama Stack CL
|
|||
|
||||
```bash
|
||||
llama stack build --distro {{ name }} --image-type venv
|
||||
INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
llama stack run distributions/{{ name }}/run.yaml \
|
||||
--port 8321 \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct
|
||||
--port 8321
|
||||
```
|
||||
|
||||
If you are using Llama Stack Safety / Shield APIs, use:
|
||||
|
||||
```bash
|
||||
INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
SAFETY_MODEL=meta-llama/Llama-Guard-3-1B \
|
||||
llama stack run distributions/{{ name }}/run-with-safety.yaml \
|
||||
--port 8321 \
|
||||
--env INFERENCE_MODEL=meta-llama/Llama-3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-llama/Llama-Guard-3-1B
|
||||
--port 8321
|
||||
```
|
||||
|
|
|
@ -118,10 +118,10 @@ docker run \
|
|||
--pull always \
|
||||
-p $LLAMA_STACK_PORT:$LLAMA_STACK_PORT \
|
||||
-v ./run.yaml:/root/my-run.yaml \
|
||||
-e NVIDIA_API_KEY=$NVIDIA_API_KEY \
|
||||
llamastack/distribution-{{ name }} \
|
||||
--config /root/my-run.yaml \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env NVIDIA_API_KEY=$NVIDIA_API_KEY
|
||||
--port $LLAMA_STACK_PORT
|
||||
```
|
||||
|
||||
### Via venv
|
||||
|
@ -131,10 +131,10 @@ If you've set up your local development environment, you can also build the imag
|
|||
```bash
|
||||
INFERENCE_MODEL=meta-llama/Llama-3.1-8B-Instruct
|
||||
llama stack build --distro nvidia --image-type venv
|
||||
NVIDIA_API_KEY=$NVIDIA_API_KEY \
|
||||
INFERENCE_MODEL=$INFERENCE_MODEL \
|
||||
llama stack run ./run.yaml \
|
||||
--port 8321 \
|
||||
--env NVIDIA_API_KEY=$NVIDIA_API_KEY \
|
||||
--env INFERENCE_MODEL=$INFERENCE_MODEL
|
||||
--port 8321
|
||||
```
|
||||
|
||||
## Example Notebooks
|
||||
|
|
|
@ -22,6 +22,7 @@ async def get_provider_impl(config: MetaReferenceAgentsImplConfig, deps: dict[Ap
|
|||
deps[Api.tool_runtime],
|
||||
deps[Api.tool_groups],
|
||||
policy,
|
||||
Api.telemetry in deps,
|
||||
)
|
||||
await impl.initialize()
|
||||
return impl
|
||||
|
|
|
@ -110,6 +110,7 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
persistence_store: KVStore,
|
||||
created_at: str,
|
||||
policy: list[AccessRule],
|
||||
telemetry_enabled: bool = False,
|
||||
):
|
||||
self.agent_id = agent_id
|
||||
self.agent_config = agent_config
|
||||
|
@ -120,6 +121,7 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
self.tool_runtime_api = tool_runtime_api
|
||||
self.tool_groups_api = tool_groups_api
|
||||
self.created_at = created_at
|
||||
self.telemetry_enabled = telemetry_enabled
|
||||
|
||||
ShieldRunnerMixin.__init__(
|
||||
self,
|
||||
|
@ -188,28 +190,30 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
|
||||
async def create_and_execute_turn(self, request: AgentTurnCreateRequest) -> AsyncGenerator:
|
||||
turn_id = str(uuid.uuid4())
|
||||
span = tracing.get_current_span()
|
||||
if span:
|
||||
span.set_attribute("session_id", request.session_id)
|
||||
span.set_attribute("agent_id", self.agent_id)
|
||||
span.set_attribute("request", request.model_dump_json())
|
||||
span.set_attribute("turn_id", turn_id)
|
||||
if self.agent_config.name:
|
||||
span.set_attribute("agent_name", self.agent_config.name)
|
||||
if self.telemetry_enabled:
|
||||
span = tracing.get_current_span()
|
||||
if span is not None:
|
||||
span.set_attribute("session_id", request.session_id)
|
||||
span.set_attribute("agent_id", self.agent_id)
|
||||
span.set_attribute("request", request.model_dump_json())
|
||||
span.set_attribute("turn_id", turn_id)
|
||||
if self.agent_config.name:
|
||||
span.set_attribute("agent_name", self.agent_config.name)
|
||||
|
||||
await self._initialize_tools(request.toolgroups)
|
||||
async for chunk in self._run_turn(request, turn_id):
|
||||
yield chunk
|
||||
|
||||
async def resume_turn(self, request: AgentTurnResumeRequest) -> AsyncGenerator:
|
||||
span = tracing.get_current_span()
|
||||
if span:
|
||||
span.set_attribute("agent_id", self.agent_id)
|
||||
span.set_attribute("session_id", request.session_id)
|
||||
span.set_attribute("request", request.model_dump_json())
|
||||
span.set_attribute("turn_id", request.turn_id)
|
||||
if self.agent_config.name:
|
||||
span.set_attribute("agent_name", self.agent_config.name)
|
||||
if self.telemetry_enabled:
|
||||
span = tracing.get_current_span()
|
||||
if span is not None:
|
||||
span.set_attribute("agent_id", self.agent_id)
|
||||
span.set_attribute("session_id", request.session_id)
|
||||
span.set_attribute("request", request.model_dump_json())
|
||||
span.set_attribute("turn_id", request.turn_id)
|
||||
if self.agent_config.name:
|
||||
span.set_attribute("agent_name", self.agent_config.name)
|
||||
|
||||
await self._initialize_tools()
|
||||
async for chunk in self._run_turn(request):
|
||||
|
@ -395,9 +399,12 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
touchpoint: str,
|
||||
) -> AsyncGenerator:
|
||||
async with tracing.span("run_shields") as span:
|
||||
span.set_attribute("input", [m.model_dump_json() for m in messages])
|
||||
if self.telemetry_enabled and span is not None:
|
||||
span.set_attribute("input", [m.model_dump_json() for m in messages])
|
||||
if len(shields) == 0:
|
||||
span.set_attribute("output", "no shields")
|
||||
|
||||
if len(shields) == 0:
|
||||
span.set_attribute("output", "no shields")
|
||||
return
|
||||
|
||||
step_id = str(uuid.uuid4())
|
||||
|
@ -430,7 +437,8 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
)
|
||||
)
|
||||
)
|
||||
span.set_attribute("output", e.violation.model_dump_json())
|
||||
if self.telemetry_enabled and span is not None:
|
||||
span.set_attribute("output", e.violation.model_dump_json())
|
||||
|
||||
yield CompletionMessage(
|
||||
content=str(e),
|
||||
|
@ -453,7 +461,8 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
)
|
||||
)
|
||||
)
|
||||
span.set_attribute("output", "no violations")
|
||||
if self.telemetry_enabled and span is not None:
|
||||
span.set_attribute("output", "no violations")
|
||||
|
||||
async def _run(
|
||||
self,
|
||||
|
@ -518,8 +527,9 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
stop_reason: StopReason | None = None
|
||||
|
||||
async with tracing.span("inference") as span:
|
||||
if self.agent_config.name:
|
||||
span.set_attribute("agent_name", self.agent_config.name)
|
||||
if self.telemetry_enabled and span is not None:
|
||||
if self.agent_config.name:
|
||||
span.set_attribute("agent_name", self.agent_config.name)
|
||||
|
||||
def _serialize_nested(value):
|
||||
"""Recursively serialize nested Pydantic models to dicts."""
|
||||
|
@ -637,18 +647,19 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
else:
|
||||
raise ValueError(f"Unexpected delta type {type(delta)}")
|
||||
|
||||
span.set_attribute("stop_reason", stop_reason or StopReason.end_of_turn)
|
||||
span.set_attribute(
|
||||
"input",
|
||||
json.dumps([json.loads(m.model_dump_json()) for m in input_messages]),
|
||||
)
|
||||
output_attr = json.dumps(
|
||||
{
|
||||
"content": content,
|
||||
"tool_calls": [json.loads(t.model_dump_json()) for t in tool_calls],
|
||||
}
|
||||
)
|
||||
span.set_attribute("output", output_attr)
|
||||
if self.telemetry_enabled and span is not None:
|
||||
span.set_attribute("stop_reason", stop_reason or StopReason.end_of_turn)
|
||||
span.set_attribute(
|
||||
"input",
|
||||
json.dumps([json.loads(m.model_dump_json()) for m in input_messages]),
|
||||
)
|
||||
output_attr = json.dumps(
|
||||
{
|
||||
"content": content,
|
||||
"tool_calls": [json.loads(t.model_dump_json()) for t in tool_calls],
|
||||
}
|
||||
)
|
||||
span.set_attribute("output", output_attr)
|
||||
|
||||
n_iter += 1
|
||||
await self.storage.set_num_infer_iters_in_turn(session_id, turn_id, n_iter)
|
||||
|
@ -756,7 +767,9 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
{
|
||||
"tool_name": tool_call.tool_name,
|
||||
"input": message.model_dump_json(),
|
||||
},
|
||||
}
|
||||
if self.telemetry_enabled
|
||||
else {},
|
||||
) as span:
|
||||
tool_execution_start_time = datetime.now(UTC).isoformat()
|
||||
tool_result = await self.execute_tool_call_maybe(
|
||||
|
@ -771,7 +784,8 @@ class ChatAgent(ShieldRunnerMixin):
|
|||
call_id=tool_call.call_id,
|
||||
content=tool_result.content,
|
||||
)
|
||||
span.set_attribute("output", result_message.model_dump_json())
|
||||
if self.telemetry_enabled and span is not None:
|
||||
span.set_attribute("output", result_message.model_dump_json())
|
||||
|
||||
# Store tool execution step
|
||||
tool_execution_step = ToolExecutionStep(
|
||||
|
|
|
@ -64,6 +64,7 @@ class MetaReferenceAgentsImpl(Agents):
|
|||
tool_runtime_api: ToolRuntime,
|
||||
tool_groups_api: ToolGroups,
|
||||
policy: list[AccessRule],
|
||||
telemetry_enabled: bool = False,
|
||||
):
|
||||
self.config = config
|
||||
self.inference_api = inference_api
|
||||
|
@ -71,6 +72,7 @@ class MetaReferenceAgentsImpl(Agents):
|
|||
self.safety_api = safety_api
|
||||
self.tool_runtime_api = tool_runtime_api
|
||||
self.tool_groups_api = tool_groups_api
|
||||
self.telemetry_enabled = telemetry_enabled
|
||||
|
||||
self.in_memory_store = InmemoryKVStoreImpl()
|
||||
self.openai_responses_impl: OpenAIResponsesImpl | None = None
|
||||
|
@ -135,6 +137,7 @@ class MetaReferenceAgentsImpl(Agents):
|
|||
),
|
||||
created_at=agent_info.created_at,
|
||||
policy=self.policy,
|
||||
telemetry_enabled=self.telemetry_enabled,
|
||||
)
|
||||
|
||||
async def create_agent_session(
|
||||
|
|
|
@ -200,12 +200,10 @@ class FaissIndex(EmbeddingIndex):
|
|||
|
||||
class FaissVectorIOAdapter(OpenAIVectorStoreMixin, VectorIO, VectorDBsProtocolPrivate):
|
||||
def __init__(self, config: FaissVectorIOConfig, inference_api: Inference, files_api: Files | None) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
self.config = config
|
||||
self.inference_api = inference_api
|
||||
self.files_api = files_api
|
||||
self.cache: dict[str, VectorDBWithIndex] = {}
|
||||
self.kvstore: KVStore | None = None
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
|
||||
async def initialize(self) -> None:
|
||||
self.kvstore = await kvstore_impl(self.config.kvstore)
|
||||
|
|
|
@ -410,12 +410,10 @@ class SQLiteVecVectorIOAdapter(OpenAIVectorStoreMixin, VectorIO, VectorDBsProtoc
|
|||
"""
|
||||
|
||||
def __init__(self, config, inference_api: Inference, files_api: Files | None) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
self.config = config
|
||||
self.inference_api = inference_api
|
||||
self.files_api = files_api
|
||||
self.cache: dict[str, VectorDBWithIndex] = {}
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
self.kvstore: KVStore | None = None
|
||||
|
||||
async def initialize(self) -> None:
|
||||
self.kvstore = await kvstore_impl(self.config.kvstore)
|
||||
|
|
|
@ -36,6 +36,9 @@ def available_providers() -> list[ProviderSpec]:
|
|||
Api.tool_runtime,
|
||||
Api.tool_groups,
|
||||
],
|
||||
optional_api_dependencies=[
|
||||
Api.telemetry,
|
||||
],
|
||||
description="Meta's reference implementation of an agent system that can use tools, access vector databases, and perform complex reasoning tasks.",
|
||||
),
|
||||
]
|
||||
|
|
|
@ -11,6 +11,7 @@ from llama_stack.providers.datatypes import (
|
|||
ProviderSpec,
|
||||
RemoteProviderSpec,
|
||||
)
|
||||
from llama_stack.providers.registry.vector_io import DEFAULT_VECTOR_IO_DEPS
|
||||
|
||||
|
||||
def available_providers() -> list[ProviderSpec]:
|
||||
|
@ -18,9 +19,8 @@ def available_providers() -> list[ProviderSpec]:
|
|||
InlineProviderSpec(
|
||||
api=Api.tool_runtime,
|
||||
provider_type="inline::rag-runtime",
|
||||
pip_packages=[
|
||||
"chardet",
|
||||
"pypdf",
|
||||
pip_packages=DEFAULT_VECTOR_IO_DEPS
|
||||
+ [
|
||||
"tqdm",
|
||||
"numpy",
|
||||
"scikit-learn",
|
||||
|
|
|
@ -12,13 +12,16 @@ from llama_stack.providers.datatypes import (
|
|||
RemoteProviderSpec,
|
||||
)
|
||||
|
||||
# Common dependencies for all vector IO providers that support document processing
|
||||
DEFAULT_VECTOR_IO_DEPS = ["chardet", "pypdf"]
|
||||
|
||||
|
||||
def available_providers() -> list[ProviderSpec]:
|
||||
return [
|
||||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::meta-reference",
|
||||
pip_packages=["faiss-cpu"],
|
||||
pip_packages=["faiss-cpu"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.faiss",
|
||||
config_class="llama_stack.providers.inline.vector_io.faiss.FaissVectorIOConfig",
|
||||
deprecation_warning="Please use the `inline::faiss` provider instead.",
|
||||
|
@ -29,7 +32,7 @@ def available_providers() -> list[ProviderSpec]:
|
|||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::faiss",
|
||||
pip_packages=["faiss-cpu"],
|
||||
pip_packages=["faiss-cpu"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.faiss",
|
||||
config_class="llama_stack.providers.inline.vector_io.faiss.FaissVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -82,7 +85,7 @@ more details about Faiss in general.
|
|||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::sqlite-vec",
|
||||
pip_packages=["sqlite-vec"],
|
||||
pip_packages=["sqlite-vec"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.sqlite_vec",
|
||||
config_class="llama_stack.providers.inline.vector_io.sqlite_vec.SQLiteVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -289,7 +292,7 @@ See [sqlite-vec's GitHub repo](https://github.com/asg017/sqlite-vec/tree/main) f
|
|||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::sqlite_vec",
|
||||
pip_packages=["sqlite-vec"],
|
||||
pip_packages=["sqlite-vec"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.sqlite_vec",
|
||||
config_class="llama_stack.providers.inline.vector_io.sqlite_vec.SQLiteVectorIOConfig",
|
||||
deprecation_warning="Please use the `inline::sqlite-vec` provider (notice the hyphen instead of underscore) instead.",
|
||||
|
@ -303,7 +306,7 @@ Please refer to the sqlite-vec provider documentation.
|
|||
api=Api.vector_io,
|
||||
adapter_type="chromadb",
|
||||
provider_type="remote::chromadb",
|
||||
pip_packages=["chromadb-client"],
|
||||
pip_packages=["chromadb-client"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.remote.vector_io.chroma",
|
||||
config_class="llama_stack.providers.remote.vector_io.chroma.ChromaVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -345,7 +348,7 @@ See [Chroma's documentation](https://docs.trychroma.com/docs/overview/introducti
|
|||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::chromadb",
|
||||
pip_packages=["chromadb"],
|
||||
pip_packages=["chromadb"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.chroma",
|
||||
config_class="llama_stack.providers.inline.vector_io.chroma.ChromaVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -389,7 +392,7 @@ See [Chroma's documentation](https://docs.trychroma.com/docs/overview/introducti
|
|||
api=Api.vector_io,
|
||||
adapter_type="pgvector",
|
||||
provider_type="remote::pgvector",
|
||||
pip_packages=["psycopg2-binary"],
|
||||
pip_packages=["psycopg2-binary"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.remote.vector_io.pgvector",
|
||||
config_class="llama_stack.providers.remote.vector_io.pgvector.PGVectorVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -500,7 +503,7 @@ See [PGVector's documentation](https://github.com/pgvector/pgvector) for more de
|
|||
api=Api.vector_io,
|
||||
adapter_type="weaviate",
|
||||
provider_type="remote::weaviate",
|
||||
pip_packages=["weaviate-client>=4.16.5"],
|
||||
pip_packages=["weaviate-client>=4.16.5"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.remote.vector_io.weaviate",
|
||||
config_class="llama_stack.providers.remote.vector_io.weaviate.WeaviateVectorIOConfig",
|
||||
provider_data_validator="llama_stack.providers.remote.vector_io.weaviate.WeaviateRequestProviderData",
|
||||
|
@ -541,7 +544,7 @@ See [Weaviate's documentation](https://weaviate.io/developers/weaviate) for more
|
|||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::qdrant",
|
||||
pip_packages=["qdrant-client"],
|
||||
pip_packages=["qdrant-client"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.qdrant",
|
||||
config_class="llama_stack.providers.inline.vector_io.qdrant.QdrantVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -594,7 +597,7 @@ See the [Qdrant documentation](https://qdrant.tech/documentation/) for more deta
|
|||
api=Api.vector_io,
|
||||
adapter_type="qdrant",
|
||||
provider_type="remote::qdrant",
|
||||
pip_packages=["qdrant-client"],
|
||||
pip_packages=["qdrant-client"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.remote.vector_io.qdrant",
|
||||
config_class="llama_stack.providers.remote.vector_io.qdrant.QdrantVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -607,7 +610,7 @@ Please refer to the inline provider documentation.
|
|||
api=Api.vector_io,
|
||||
adapter_type="milvus",
|
||||
provider_type="remote::milvus",
|
||||
pip_packages=["pymilvus>=2.4.10"],
|
||||
pip_packages=["pymilvus>=2.4.10"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.remote.vector_io.milvus",
|
||||
config_class="llama_stack.providers.remote.vector_io.milvus.MilvusVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
@ -813,7 +816,7 @@ For more details on TLS configuration, refer to the [TLS setup guide](https://mi
|
|||
InlineProviderSpec(
|
||||
api=Api.vector_io,
|
||||
provider_type="inline::milvus",
|
||||
pip_packages=["pymilvus[milvus-lite]>=2.4.10"],
|
||||
pip_packages=["pymilvus[milvus-lite]>=2.4.10"] + DEFAULT_VECTOR_IO_DEPS,
|
||||
module="llama_stack.providers.inline.vector_io.milvus",
|
||||
config_class="llama_stack.providers.inline.vector_io.milvus.MilvusVectorIOConfig",
|
||||
api_dependencies=[Api.inference],
|
||||
|
|
|
@ -41,9 +41,6 @@ class DatabricksInferenceAdapter(OpenAIMixin):
|
|||
).serving_endpoints.list() # TODO: this is not async
|
||||
]
|
||||
|
||||
async def should_refresh_models(self) -> bool:
|
||||
return False
|
||||
|
||||
async def openai_completion(
|
||||
self,
|
||||
model: str,
|
||||
|
|
|
@ -6,8 +6,6 @@
|
|||
|
||||
from typing import Any
|
||||
|
||||
from pydantic import Field
|
||||
|
||||
from llama_stack.providers.utils.inference.model_registry import RemoteInferenceProviderConfig
|
||||
|
||||
DEFAULT_OLLAMA_URL = "http://localhost:11434"
|
||||
|
@ -15,10 +13,6 @@ DEFAULT_OLLAMA_URL = "http://localhost:11434"
|
|||
|
||||
class OllamaImplConfig(RemoteInferenceProviderConfig):
|
||||
url: str = DEFAULT_OLLAMA_URL
|
||||
refresh_models: bool = Field(
|
||||
default=False,
|
||||
description="Whether to refresh models periodically",
|
||||
)
|
||||
|
||||
@classmethod
|
||||
def sample_run_config(cls, url: str = "${env.OLLAMA_URL:=http://localhost:11434}", **kwargs) -> dict[str, Any]:
|
||||
|
|
|
@ -72,9 +72,6 @@ class OllamaInferenceAdapter(OpenAIMixin):
|
|||
f"Ollama Server is not running (message: {r['message']}). Make sure to start it using `ollama serve` in a separate terminal"
|
||||
)
|
||||
|
||||
async def should_refresh_models(self) -> bool:
|
||||
return self.config.refresh_models
|
||||
|
||||
async def health(self) -> HealthResponse:
|
||||
"""
|
||||
Performs a health check by verifying connectivity to the Ollama server.
|
||||
|
|
|
@ -11,6 +11,6 @@ async def get_adapter_impl(config: RunpodImplConfig, _deps):
|
|||
from .runpod import RunpodInferenceAdapter
|
||||
|
||||
assert isinstance(config, RunpodImplConfig), f"Unexpected config type: {type(config)}"
|
||||
impl = RunpodInferenceAdapter(config)
|
||||
impl = RunpodInferenceAdapter(config=config)
|
||||
await impl.initialize()
|
||||
return impl
|
||||
|
|
|
@ -4,69 +4,86 @@
|
|||
# This source code is licensed under the terms described in the LICENSE file in
|
||||
# the root directory of this source tree.
|
||||
|
||||
from typing import Any
|
||||
|
||||
from llama_stack.apis.inference import * # noqa: F403
|
||||
from llama_stack.apis.inference import OpenAIEmbeddingsResponse
|
||||
|
||||
# from llama_stack.providers.datatypes import ModelsProtocolPrivate
|
||||
from llama_stack.providers.utils.inference.model_registry import ModelRegistryHelper, build_hf_repo_model_entry
|
||||
from llama_stack.providers.utils.inference.openai_compat import (
|
||||
get_sampling_options,
|
||||
)
|
||||
from llama_stack.providers.utils.inference.prompt_adapter import (
|
||||
chat_completion_request_to_prompt,
|
||||
from llama_stack.apis.inference import (
|
||||
OpenAIMessageParam,
|
||||
OpenAIResponseFormatParam,
|
||||
)
|
||||
from llama_stack.providers.utils.inference.openai_mixin import OpenAIMixin
|
||||
|
||||
from .config import RunpodImplConfig
|
||||
|
||||
# https://docs.runpod.io/serverless/vllm/overview#compatible-models
|
||||
# https://github.com/runpod-workers/worker-vllm/blob/main/README.md#compatible-model-architectures
|
||||
RUNPOD_SUPPORTED_MODELS = {
|
||||
"Llama3.1-8B": "meta-llama/Llama-3.1-8B",
|
||||
"Llama3.1-70B": "meta-llama/Llama-3.1-70B",
|
||||
"Llama3.1-405B:bf16-mp8": "meta-llama/Llama-3.1-405B",
|
||||
"Llama3.1-405B": "meta-llama/Llama-3.1-405B-FP8",
|
||||
"Llama3.1-405B:bf16-mp16": "meta-llama/Llama-3.1-405B",
|
||||
"Llama3.1-8B-Instruct": "meta-llama/Llama-3.1-8B-Instruct",
|
||||
"Llama3.1-70B-Instruct": "meta-llama/Llama-3.1-70B-Instruct",
|
||||
"Llama3.1-405B-Instruct:bf16-mp8": "meta-llama/Llama-3.1-405B-Instruct",
|
||||
"Llama3.1-405B-Instruct": "meta-llama/Llama-3.1-405B-Instruct-FP8",
|
||||
"Llama3.1-405B-Instruct:bf16-mp16": "meta-llama/Llama-3.1-405B-Instruct",
|
||||
"Llama3.2-1B": "meta-llama/Llama-3.2-1B",
|
||||
"Llama3.2-3B": "meta-llama/Llama-3.2-3B",
|
||||
}
|
||||
|
||||
SAFETY_MODELS_ENTRIES = []
|
||||
class RunpodInferenceAdapter(OpenAIMixin):
|
||||
"""
|
||||
Adapter for RunPod's OpenAI-compatible API endpoints.
|
||||
Supports VLLM for serverless endpoint self-hosted or public endpoints.
|
||||
Can work with any runpod endpoints that support OpenAI-compatible API
|
||||
"""
|
||||
|
||||
# Create MODEL_ENTRIES from RUNPOD_SUPPORTED_MODELS for compatibility with starter template
|
||||
MODEL_ENTRIES = [
|
||||
build_hf_repo_model_entry(provider_model_id, model_descriptor)
|
||||
for provider_model_id, model_descriptor in RUNPOD_SUPPORTED_MODELS.items()
|
||||
] + SAFETY_MODELS_ENTRIES
|
||||
config: RunpodImplConfig
|
||||
|
||||
def get_api_key(self) -> str:
|
||||
"""Get API key for OpenAI client."""
|
||||
return self.config.api_token
|
||||
|
||||
class RunpodInferenceAdapter(
|
||||
ModelRegistryHelper,
|
||||
Inference,
|
||||
):
|
||||
def __init__(self, config: RunpodImplConfig) -> None:
|
||||
ModelRegistryHelper.__init__(self, stack_to_provider_models_map=RUNPOD_SUPPORTED_MODELS)
|
||||
self.config = config
|
||||
def get_base_url(self) -> str:
|
||||
"""Get base URL for OpenAI client."""
|
||||
return self.config.url
|
||||
|
||||
def _get_params(self, request: ChatCompletionRequest) -> dict:
|
||||
return {
|
||||
"model": self.map_to_provider_model(request.model),
|
||||
"prompt": chat_completion_request_to_prompt(request),
|
||||
"stream": request.stream,
|
||||
**get_sampling_options(request.sampling_params),
|
||||
}
|
||||
|
||||
async def openai_embeddings(
|
||||
async def openai_chat_completion(
|
||||
self,
|
||||
model: str,
|
||||
input: str | list[str],
|
||||
encoding_format: str | None = "float",
|
||||
dimensions: int | None = None,
|
||||
messages: list[OpenAIMessageParam],
|
||||
frequency_penalty: float | None = None,
|
||||
function_call: str | dict[str, Any] | None = None,
|
||||
functions: list[dict[str, Any]] | None = None,
|
||||
logit_bias: dict[str, float] | None = None,
|
||||
logprobs: bool | None = None,
|
||||
max_completion_tokens: int | None = None,
|
||||
max_tokens: int | None = None,
|
||||
n: int | None = None,
|
||||
parallel_tool_calls: bool | None = None,
|
||||
presence_penalty: float | None = None,
|
||||
response_format: OpenAIResponseFormatParam | None = None,
|
||||
seed: int | None = None,
|
||||
stop: str | list[str] | None = None,
|
||||
stream: bool | None = None,
|
||||
stream_options: dict[str, Any] | None = None,
|
||||
temperature: float | None = None,
|
||||
tool_choice: str | dict[str, Any] | None = None,
|
||||
tools: list[dict[str, Any]] | None = None,
|
||||
top_logprobs: int | None = None,
|
||||
top_p: float | None = None,
|
||||
user: str | None = None,
|
||||
) -> OpenAIEmbeddingsResponse:
|
||||
raise NotImplementedError()
|
||||
):
|
||||
"""Override to add RunPod-specific stream_options requirement."""
|
||||
if stream and not stream_options:
|
||||
stream_options = {"include_usage": True}
|
||||
|
||||
return await super().openai_chat_completion(
|
||||
model=model,
|
||||
messages=messages,
|
||||
frequency_penalty=frequency_penalty,
|
||||
function_call=function_call,
|
||||
functions=functions,
|
||||
logit_bias=logit_bias,
|
||||
logprobs=logprobs,
|
||||
max_completion_tokens=max_completion_tokens,
|
||||
max_tokens=max_tokens,
|
||||
n=n,
|
||||
parallel_tool_calls=parallel_tool_calls,
|
||||
presence_penalty=presence_penalty,
|
||||
response_format=response_format,
|
||||
seed=seed,
|
||||
stop=stop,
|
||||
stream=stream,
|
||||
stream_options=stream_options,
|
||||
temperature=temperature,
|
||||
tool_choice=tool_choice,
|
||||
tools=tools,
|
||||
top_logprobs=top_logprobs,
|
||||
top_p=top_p,
|
||||
user=user,
|
||||
)
|
||||
|
|
|
@ -63,9 +63,6 @@ class TogetherInferenceAdapter(OpenAIMixin, NeedsRequestProviderData):
|
|||
# Together's /v1/models is not compatible with OpenAI's /v1/models. Together support ticket #13355 -> will not fix, use Together's own client
|
||||
return [m.id for m in await self._get_client().models.list()]
|
||||
|
||||
async def should_refresh_models(self) -> bool:
|
||||
return True
|
||||
|
||||
async def openai_embeddings(
|
||||
self,
|
||||
model: str,
|
||||
|
|
|
@ -30,10 +30,6 @@ class VLLMInferenceAdapterConfig(RemoteInferenceProviderConfig):
|
|||
default=True,
|
||||
description="Whether to verify TLS certificates. Can be a boolean or a path to a CA certificate file.",
|
||||
)
|
||||
refresh_models: bool = Field(
|
||||
default=False,
|
||||
description="Whether to refresh models periodically",
|
||||
)
|
||||
|
||||
@field_validator("tls_verify")
|
||||
@classmethod
|
||||
|
|
|
@ -53,10 +53,6 @@ class VLLMInferenceAdapter(OpenAIMixin):
|
|||
"You must provide a URL in run.yaml (or via the VLLM_URL environment variable) to use vLLM."
|
||||
)
|
||||
|
||||
async def should_refresh_models(self) -> bool:
|
||||
# Strictly respecting the refresh_models directive
|
||||
return self.config.refresh_models
|
||||
|
||||
async def health(self) -> HealthResponse:
|
||||
"""
|
||||
Performs a health check by verifying connectivity to the remote vLLM server.
|
||||
|
|
|
@ -140,14 +140,13 @@ class ChromaVectorIOAdapter(OpenAIVectorStoreMixin, VectorIO, VectorDBsProtocolP
|
|||
inference_api: Api.inference,
|
||||
files_api: Files | None,
|
||||
) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
log.info(f"Initializing ChromaVectorIOAdapter with url: {config}")
|
||||
self.config = config
|
||||
self.inference_api = inference_api
|
||||
self.client = None
|
||||
self.cache = {}
|
||||
self.kvstore: KVStore | None = None
|
||||
self.vector_db_store = None
|
||||
self.files_api = files_api
|
||||
|
||||
async def initialize(self) -> None:
|
||||
self.kvstore = await kvstore_impl(self.config.kvstore)
|
||||
|
|
|
@ -309,14 +309,12 @@ class MilvusVectorIOAdapter(OpenAIVectorStoreMixin, VectorIO, VectorDBsProtocolP
|
|||
inference_api: Inference,
|
||||
files_api: Files | None,
|
||||
) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
self.config = config
|
||||
self.cache = {}
|
||||
self.client = None
|
||||
self.inference_api = inference_api
|
||||
self.files_api = files_api
|
||||
self.kvstore: KVStore | None = None
|
||||
self.vector_db_store = None
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
self.metadata_collection_name = "openai_vector_stores_metadata"
|
||||
|
||||
async def initialize(self) -> None:
|
||||
|
|
|
@ -345,14 +345,12 @@ class PGVectorVectorIOAdapter(OpenAIVectorStoreMixin, VectorIO, VectorDBsProtoco
|
|||
inference_api: Api.inference,
|
||||
files_api: Files | None = None,
|
||||
) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
self.config = config
|
||||
self.inference_api = inference_api
|
||||
self.conn = None
|
||||
self.cache = {}
|
||||
self.files_api = files_api
|
||||
self.kvstore: KVStore | None = None
|
||||
self.vector_db_store = None
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
self.metadata_collection_name = "openai_vector_stores_metadata"
|
||||
|
||||
async def initialize(self) -> None:
|
||||
|
|
|
@ -27,7 +27,7 @@ from llama_stack.apis.vector_io import (
|
|||
from llama_stack.log import get_logger
|
||||
from llama_stack.providers.datatypes import Api, VectorDBsProtocolPrivate
|
||||
from llama_stack.providers.inline.vector_io.qdrant import QdrantVectorIOConfig as InlineQdrantVectorIOConfig
|
||||
from llama_stack.providers.utils.kvstore import KVStore, kvstore_impl
|
||||
from llama_stack.providers.utils.kvstore import kvstore_impl
|
||||
from llama_stack.providers.utils.memory.openai_vector_store_mixin import OpenAIVectorStoreMixin
|
||||
from llama_stack.providers.utils.memory.vector_store import (
|
||||
ChunkForDeletion,
|
||||
|
@ -162,14 +162,12 @@ class QdrantVectorIOAdapter(OpenAIVectorStoreMixin, VectorIO, VectorDBsProtocolP
|
|||
inference_api: Api.inference,
|
||||
files_api: Files | None = None,
|
||||
) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
self.config = config
|
||||
self.client: AsyncQdrantClient = None
|
||||
self.cache = {}
|
||||
self.inference_api = inference_api
|
||||
self.files_api = files_api
|
||||
self.vector_db_store = None
|
||||
self.kvstore: KVStore | None = None
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
self._qdrant_lock = asyncio.Lock()
|
||||
|
||||
async def initialize(self) -> None:
|
||||
|
|
|
@ -284,14 +284,12 @@ class WeaviateVectorIOAdapter(
|
|||
inference_api: Api.inference,
|
||||
files_api: Files | None,
|
||||
) -> None:
|
||||
super().__init__(files_api=files_api, kvstore=None)
|
||||
self.config = config
|
||||
self.inference_api = inference_api
|
||||
self.client_cache = {}
|
||||
self.cache = {}
|
||||
self.files_api = files_api
|
||||
self.kvstore: KVStore | None = None
|
||||
self.vector_db_store = None
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
self.metadata_collection_name = "openai_vector_stores_metadata"
|
||||
|
||||
def _get_client(self) -> weaviate.WeaviateClient:
|
||||
|
|
|
@ -24,6 +24,10 @@ class RemoteInferenceProviderConfig(BaseModel):
|
|||
default=None,
|
||||
description="List of models that should be registered with the model registry. If None, all models are allowed.",
|
||||
)
|
||||
refresh_models: bool = Field(
|
||||
default=False,
|
||||
description="Whether to refresh models periodically from the provider",
|
||||
)
|
||||
|
||||
|
||||
# TODO: this class is more confusing than useful right now. We need to make it
|
||||
|
|
|
@ -484,7 +484,7 @@ class OpenAIMixin(NeedsRequestProviderData, ABC, BaseModel):
|
|||
return model in self._model_cache
|
||||
|
||||
async def should_refresh_models(self) -> bool:
|
||||
return False
|
||||
return self.config.refresh_models
|
||||
|
||||
#
|
||||
# The model_dump implementations are to avoid serializing the extra fields,
|
||||
|
|
|
@ -12,6 +12,8 @@ import uuid
|
|||
from abc import ABC, abstractmethod
|
||||
from typing import Any
|
||||
|
||||
from pydantic import TypeAdapter
|
||||
|
||||
from llama_stack.apis.common.errors import VectorStoreNotFoundError
|
||||
from llama_stack.apis.files import Files, OpenAIFileObject
|
||||
from llama_stack.apis.vector_dbs import VectorDB
|
||||
|
@ -50,12 +52,16 @@ logger = get_logger(name=__name__, category="providers::utils")
|
|||
|
||||
# Constants for OpenAI vector stores
|
||||
CHUNK_MULTIPLIER = 5
|
||||
FILE_BATCH_CLEANUP_INTERVAL_SECONDS = 24 * 60 * 60 # 1 day in seconds
|
||||
MAX_CONCURRENT_FILES_PER_BATCH = 3 # Maximum concurrent file processing within a batch
|
||||
FILE_BATCH_CHUNK_SIZE = 10 # Process files in chunks of this size
|
||||
|
||||
VERSION = "v3"
|
||||
VECTOR_DBS_PREFIX = f"vector_dbs:{VERSION}::"
|
||||
OPENAI_VECTOR_STORES_PREFIX = f"openai_vector_stores:{VERSION}::"
|
||||
OPENAI_VECTOR_STORES_FILES_PREFIX = f"openai_vector_stores_files:{VERSION}::"
|
||||
OPENAI_VECTOR_STORES_FILES_CONTENTS_PREFIX = f"openai_vector_stores_files_contents:{VERSION}::"
|
||||
OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX = f"openai_vector_stores_file_batches:{VERSION}::"
|
||||
|
||||
|
||||
class OpenAIVectorStoreMixin(ABC):
|
||||
|
@ -65,11 +71,15 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
an openai_vector_stores in-memory cache.
|
||||
"""
|
||||
|
||||
# These should be provided by the implementing class
|
||||
openai_vector_stores: dict[str, dict[str, Any]]
|
||||
files_api: Files | None
|
||||
# KV store for persisting OpenAI vector store metadata
|
||||
kvstore: KVStore | None
|
||||
# Implementing classes should call super().__init__() in their __init__ method
|
||||
# to properly initialize the mixin attributes.
|
||||
def __init__(self, files_api: Files | None = None, kvstore: KVStore | None = None):
|
||||
self.openai_vector_stores: dict[str, dict[str, Any]] = {}
|
||||
self.openai_file_batches: dict[str, dict[str, Any]] = {}
|
||||
self.files_api = files_api
|
||||
self.kvstore = kvstore
|
||||
self._last_file_batch_cleanup_time = 0
|
||||
self._file_batch_tasks: dict[str, asyncio.Task[None]] = {}
|
||||
|
||||
async def _save_openai_vector_store(self, store_id: str, store_info: dict[str, Any]) -> None:
|
||||
"""Save vector store metadata to persistent storage."""
|
||||
|
@ -159,9 +169,129 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
for idx in range(len(raw_items)):
|
||||
await self.kvstore.delete(f"{contents_prefix}{idx}")
|
||||
|
||||
async def _save_openai_vector_store_file_batch(self, batch_id: str, batch_info: dict[str, Any]) -> None:
|
||||
"""Save file batch metadata to persistent storage."""
|
||||
assert self.kvstore
|
||||
key = f"{OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX}{batch_id}"
|
||||
await self.kvstore.set(key=key, value=json.dumps(batch_info))
|
||||
# update in-memory cache
|
||||
self.openai_file_batches[batch_id] = batch_info
|
||||
|
||||
async def _load_openai_vector_store_file_batches(self) -> dict[str, dict[str, Any]]:
|
||||
"""Load all file batch metadata from persistent storage."""
|
||||
assert self.kvstore
|
||||
start_key = OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX
|
||||
end_key = f"{OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX}\xff"
|
||||
stored_data = await self.kvstore.values_in_range(start_key, end_key)
|
||||
|
||||
batches: dict[str, dict[str, Any]] = {}
|
||||
for item in stored_data:
|
||||
info = json.loads(item)
|
||||
batches[info["id"]] = info
|
||||
return batches
|
||||
|
||||
async def _delete_openai_vector_store_file_batch(self, batch_id: str) -> None:
|
||||
"""Delete file batch metadata from persistent storage and in-memory cache."""
|
||||
assert self.kvstore
|
||||
key = f"{OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX}{batch_id}"
|
||||
await self.kvstore.delete(key)
|
||||
# remove from in-memory cache
|
||||
self.openai_file_batches.pop(batch_id, None)
|
||||
|
||||
async def _cleanup_expired_file_batches(self) -> None:
|
||||
"""Clean up expired file batches from persistent storage."""
|
||||
assert self.kvstore
|
||||
start_key = OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX
|
||||
end_key = f"{OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX}\xff"
|
||||
stored_data = await self.kvstore.values_in_range(start_key, end_key)
|
||||
|
||||
current_time = int(time.time())
|
||||
expired_count = 0
|
||||
|
||||
for item in stored_data:
|
||||
info = json.loads(item)
|
||||
expires_at = info.get("expires_at")
|
||||
if expires_at and current_time > expires_at:
|
||||
logger.info(f"Cleaning up expired file batch: {info['id']}")
|
||||
await self.kvstore.delete(f"{OPENAI_VECTOR_STORES_FILE_BATCHES_PREFIX}{info['id']}")
|
||||
# Remove from in-memory cache if present
|
||||
self.openai_file_batches.pop(info["id"], None)
|
||||
expired_count += 1
|
||||
|
||||
if expired_count > 0:
|
||||
logger.info(f"Cleaned up {expired_count} expired file batches")
|
||||
|
||||
async def _get_completed_files_in_batch(self, vector_store_id: str, file_ids: list[str]) -> set[str]:
|
||||
"""Determine which files in a batch are actually completed by checking vector store file_ids."""
|
||||
if vector_store_id not in self.openai_vector_stores:
|
||||
return set()
|
||||
|
||||
store_info = self.openai_vector_stores[vector_store_id]
|
||||
completed_files = set(file_ids) & set(store_info["file_ids"])
|
||||
return completed_files
|
||||
|
||||
async def _analyze_batch_completion_on_resume(self, batch_id: str, batch_info: dict[str, Any]) -> list[str]:
|
||||
"""Analyze batch completion status and return remaining files to process.
|
||||
|
||||
Returns:
|
||||
List of file IDs that still need processing. Empty list if batch is complete.
|
||||
"""
|
||||
vector_store_id = batch_info["vector_store_id"]
|
||||
all_file_ids = batch_info["file_ids"]
|
||||
|
||||
# Find files that are actually completed
|
||||
completed_files = await self._get_completed_files_in_batch(vector_store_id, all_file_ids)
|
||||
remaining_files = [file_id for file_id in all_file_ids if file_id not in completed_files]
|
||||
|
||||
completed_count = len(completed_files)
|
||||
total_count = len(all_file_ids)
|
||||
remaining_count = len(remaining_files)
|
||||
|
||||
# Update file counts to reflect actual state
|
||||
batch_info["file_counts"] = {
|
||||
"completed": completed_count,
|
||||
"failed": 0, # We don't track failed files during resume - they'll be retried
|
||||
"in_progress": remaining_count,
|
||||
"cancelled": 0,
|
||||
"total": total_count,
|
||||
}
|
||||
|
||||
# If all files are already completed, mark batch as completed
|
||||
if remaining_count == 0:
|
||||
batch_info["status"] = "completed"
|
||||
logger.info(f"Batch {batch_id} is already fully completed, updating status")
|
||||
|
||||
# Save updated batch info
|
||||
await self._save_openai_vector_store_file_batch(batch_id, batch_info)
|
||||
|
||||
return remaining_files
|
||||
|
||||
async def _resume_incomplete_batches(self) -> None:
|
||||
"""Resume processing of incomplete file batches after server restart."""
|
||||
for batch_id, batch_info in self.openai_file_batches.items():
|
||||
if batch_info["status"] == "in_progress":
|
||||
logger.info(f"Analyzing incomplete file batch: {batch_id}")
|
||||
|
||||
remaining_files = await self._analyze_batch_completion_on_resume(batch_id, batch_info)
|
||||
|
||||
# Check if batch is now completed after analysis
|
||||
if batch_info["status"] == "completed":
|
||||
continue
|
||||
|
||||
if remaining_files:
|
||||
logger.info(f"Resuming batch {batch_id} with {len(remaining_files)} remaining files")
|
||||
# Restart the background processing task with only remaining files
|
||||
task = asyncio.create_task(self._process_file_batch_async(batch_id, batch_info, remaining_files))
|
||||
self._file_batch_tasks[batch_id] = task
|
||||
|
||||
async def initialize_openai_vector_stores(self) -> None:
|
||||
"""Load existing OpenAI vector stores into the in-memory cache."""
|
||||
"""Load existing OpenAI vector stores and file batches into the in-memory cache."""
|
||||
self.openai_vector_stores = await self._load_openai_vector_stores()
|
||||
self.openai_file_batches = await self._load_openai_vector_store_file_batches()
|
||||
self._file_batch_tasks = {}
|
||||
# TODO: Resume only works for single worker deployment. Jobs with multiple workers will need to be handled differently.
|
||||
await self._resume_incomplete_batches()
|
||||
self._last_file_batch_cleanup_time = 0
|
||||
|
||||
@abstractmethod
|
||||
async def delete_chunks(self, store_id: str, chunks_for_deletion: list[ChunkForDeletion]) -> None:
|
||||
|
@ -570,6 +700,14 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
if vector_store_id not in self.openai_vector_stores:
|
||||
raise VectorStoreNotFoundError(vector_store_id)
|
||||
|
||||
# Check if file is already attached to this vector store
|
||||
store_info = self.openai_vector_stores[vector_store_id]
|
||||
if file_id in store_info["file_ids"]:
|
||||
logger.warning(f"File {file_id} is already attached to vector store {vector_store_id}, skipping")
|
||||
# Return existing file object
|
||||
file_info = await self._load_openai_vector_store_file(vector_store_id, file_id)
|
||||
return VectorStoreFileObject(**file_info)
|
||||
|
||||
attributes = attributes or {}
|
||||
chunking_strategy = chunking_strategy or VectorStoreChunkingStrategyAuto()
|
||||
created_at = int(time.time())
|
||||
|
@ -615,7 +753,6 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
chunk_overlap_tokens,
|
||||
attributes,
|
||||
)
|
||||
|
||||
if not chunks:
|
||||
vector_store_file_object.status = "failed"
|
||||
vector_store_file_object.last_error = VectorStoreFileLastError(
|
||||
|
@ -828,7 +965,230 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
chunking_strategy: VectorStoreChunkingStrategy | None = None,
|
||||
) -> VectorStoreFileBatchObject:
|
||||
"""Create a vector store file batch."""
|
||||
raise NotImplementedError("openai_create_vector_store_file_batch is not implemented yet")
|
||||
if vector_store_id not in self.openai_vector_stores:
|
||||
raise VectorStoreNotFoundError(vector_store_id)
|
||||
|
||||
chunking_strategy = chunking_strategy or VectorStoreChunkingStrategyAuto()
|
||||
|
||||
created_at = int(time.time())
|
||||
batch_id = f"batch_{uuid.uuid4()}"
|
||||
# File batches expire after 7 days
|
||||
expires_at = created_at + (7 * 24 * 60 * 60)
|
||||
|
||||
# Initialize batch file counts - all files start as in_progress
|
||||
file_counts = VectorStoreFileCounts(
|
||||
completed=0,
|
||||
cancelled=0,
|
||||
failed=0,
|
||||
in_progress=len(file_ids),
|
||||
total=len(file_ids),
|
||||
)
|
||||
|
||||
# Create batch object immediately with in_progress status
|
||||
batch_object = VectorStoreFileBatchObject(
|
||||
id=batch_id,
|
||||
created_at=created_at,
|
||||
vector_store_id=vector_store_id,
|
||||
status="in_progress",
|
||||
file_counts=file_counts,
|
||||
)
|
||||
|
||||
batch_info = {
|
||||
**batch_object.model_dump(),
|
||||
"file_ids": file_ids,
|
||||
"attributes": attributes,
|
||||
"chunking_strategy": chunking_strategy.model_dump(),
|
||||
"expires_at": expires_at,
|
||||
}
|
||||
await self._save_openai_vector_store_file_batch(batch_id, batch_info)
|
||||
|
||||
# Start background processing of files
|
||||
task = asyncio.create_task(self._process_file_batch_async(batch_id, batch_info))
|
||||
self._file_batch_tasks[batch_id] = task
|
||||
|
||||
# Run cleanup if needed (throttled to once every 1 day)
|
||||
current_time = int(time.time())
|
||||
if current_time - self._last_file_batch_cleanup_time >= FILE_BATCH_CLEANUP_INTERVAL_SECONDS:
|
||||
logger.info("Running throttled cleanup of expired file batches")
|
||||
asyncio.create_task(self._cleanup_expired_file_batches())
|
||||
self._last_file_batch_cleanup_time = current_time
|
||||
|
||||
return batch_object
|
||||
|
||||
async def _process_files_with_concurrency(
|
||||
self,
|
||||
file_ids: list[str],
|
||||
vector_store_id: str,
|
||||
attributes: dict[str, Any],
|
||||
chunking_strategy_obj: Any,
|
||||
batch_id: str,
|
||||
batch_info: dict[str, Any],
|
||||
) -> None:
|
||||
"""Process files with controlled concurrency and chunking."""
|
||||
semaphore = asyncio.Semaphore(MAX_CONCURRENT_FILES_PER_BATCH)
|
||||
|
||||
async def process_single_file(file_id: str) -> tuple[str, bool]:
|
||||
"""Process a single file with concurrency control."""
|
||||
async with semaphore:
|
||||
try:
|
||||
vector_store_file_object = await self.openai_attach_file_to_vector_store(
|
||||
vector_store_id=vector_store_id,
|
||||
file_id=file_id,
|
||||
attributes=attributes,
|
||||
chunking_strategy=chunking_strategy_obj,
|
||||
)
|
||||
return file_id, vector_store_file_object.status == "completed"
|
||||
except Exception as e:
|
||||
logger.error(f"Failed to process file {file_id} in batch {batch_id}: {e}")
|
||||
return file_id, False
|
||||
|
||||
# Process files in chunks to avoid creating too many tasks at once
|
||||
total_files = len(file_ids)
|
||||
for chunk_start in range(0, total_files, FILE_BATCH_CHUNK_SIZE):
|
||||
chunk_end = min(chunk_start + FILE_BATCH_CHUNK_SIZE, total_files)
|
||||
chunk = file_ids[chunk_start:chunk_end]
|
||||
|
||||
chunk_num = chunk_start // FILE_BATCH_CHUNK_SIZE + 1
|
||||
total_chunks = (total_files + FILE_BATCH_CHUNK_SIZE - 1) // FILE_BATCH_CHUNK_SIZE
|
||||
logger.info(
|
||||
f"Processing chunk {chunk_num} of {total_chunks} ({len(chunk)} files, {chunk_start + 1}-{chunk_end} of {total_files} total files)"
|
||||
)
|
||||
|
||||
async with asyncio.TaskGroup() as tg:
|
||||
chunk_tasks = [tg.create_task(process_single_file(file_id)) for file_id in chunk]
|
||||
|
||||
chunk_results = [task.result() for task in chunk_tasks]
|
||||
|
||||
# Update counts after each chunk for progressive feedback
|
||||
for _, success in chunk_results:
|
||||
self._update_file_counts(batch_info, success=success)
|
||||
|
||||
# Save progress after each chunk
|
||||
await self._save_openai_vector_store_file_batch(batch_id, batch_info)
|
||||
|
||||
def _update_file_counts(self, batch_info: dict[str, Any], success: bool) -> None:
|
||||
"""Update file counts based on processing result."""
|
||||
if success:
|
||||
batch_info["file_counts"]["completed"] += 1
|
||||
else:
|
||||
batch_info["file_counts"]["failed"] += 1
|
||||
batch_info["file_counts"]["in_progress"] -= 1
|
||||
|
||||
def _update_batch_status(self, batch_info: dict[str, Any]) -> None:
|
||||
"""Update final batch status based on file processing results."""
|
||||
if batch_info["file_counts"]["failed"] == 0:
|
||||
batch_info["status"] = "completed"
|
||||
elif batch_info["file_counts"]["completed"] == 0:
|
||||
batch_info["status"] = "failed"
|
||||
else:
|
||||
batch_info["status"] = "completed" # Partial success counts as completed
|
||||
|
||||
async def _process_file_batch_async(
|
||||
self,
|
||||
batch_id: str,
|
||||
batch_info: dict[str, Any],
|
||||
override_file_ids: list[str] | None = None,
|
||||
) -> None:
|
||||
"""Process files in a batch asynchronously in the background."""
|
||||
file_ids = override_file_ids if override_file_ids is not None else batch_info["file_ids"]
|
||||
attributes = batch_info["attributes"]
|
||||
chunking_strategy = batch_info["chunking_strategy"]
|
||||
vector_store_id = batch_info["vector_store_id"]
|
||||
chunking_strategy_adapter: TypeAdapter[VectorStoreChunkingStrategy] = TypeAdapter(VectorStoreChunkingStrategy)
|
||||
chunking_strategy_obj = chunking_strategy_adapter.validate_python(chunking_strategy)
|
||||
|
||||
try:
|
||||
# Process all files with controlled concurrency
|
||||
await self._process_files_with_concurrency(
|
||||
file_ids=file_ids,
|
||||
vector_store_id=vector_store_id,
|
||||
attributes=attributes,
|
||||
chunking_strategy_obj=chunking_strategy_obj,
|
||||
batch_id=batch_id,
|
||||
batch_info=batch_info,
|
||||
)
|
||||
|
||||
# Update final batch status
|
||||
self._update_batch_status(batch_info)
|
||||
await self._save_openai_vector_store_file_batch(batch_id, batch_info)
|
||||
|
||||
logger.info(f"File batch {batch_id} processing completed with status: {batch_info['status']}")
|
||||
|
||||
except asyncio.CancelledError:
|
||||
logger.info(f"File batch {batch_id} processing was cancelled")
|
||||
# Clean up task reference if it still exists
|
||||
self._file_batch_tasks.pop(batch_id, None)
|
||||
raise # Re-raise to ensure proper cancellation propagation
|
||||
finally:
|
||||
# Always clean up task reference when processing ends
|
||||
self._file_batch_tasks.pop(batch_id, None)
|
||||
|
||||
def _get_and_validate_batch(self, batch_id: str, vector_store_id: str) -> dict[str, Any]:
|
||||
"""Get and validate batch exists and belongs to vector store."""
|
||||
if vector_store_id not in self.openai_vector_stores:
|
||||
raise VectorStoreNotFoundError(vector_store_id)
|
||||
|
||||
if batch_id not in self.openai_file_batches:
|
||||
raise ValueError(f"File batch {batch_id} not found")
|
||||
|
||||
batch_info = self.openai_file_batches[batch_id]
|
||||
|
||||
# Check if batch has expired (read-only check)
|
||||
expires_at = batch_info.get("expires_at")
|
||||
if expires_at:
|
||||
current_time = int(time.time())
|
||||
if current_time > expires_at:
|
||||
raise ValueError(f"File batch {batch_id} has expired after 7 days from creation")
|
||||
|
||||
if batch_info["vector_store_id"] != vector_store_id:
|
||||
raise ValueError(f"File batch {batch_id} does not belong to vector store {vector_store_id}")
|
||||
|
||||
return batch_info
|
||||
|
||||
def _paginate_objects(
|
||||
self,
|
||||
objects: list[Any],
|
||||
limit: int | None = 20,
|
||||
after: str | None = None,
|
||||
before: str | None = None,
|
||||
) -> tuple[list[Any], bool, str | None, str | None]:
|
||||
"""Apply pagination to a list of objects with id fields."""
|
||||
limit = min(limit or 20, 100) # Cap at 100 as per OpenAI
|
||||
|
||||
# Find start index
|
||||
start_idx = 0
|
||||
if after:
|
||||
for i, obj in enumerate(objects):
|
||||
if obj.id == after:
|
||||
start_idx = i + 1
|
||||
break
|
||||
|
||||
# Find end index
|
||||
end_idx = start_idx + limit
|
||||
if before:
|
||||
for i, obj in enumerate(objects[start_idx:], start_idx):
|
||||
if obj.id == before:
|
||||
end_idx = i
|
||||
break
|
||||
|
||||
# Apply pagination
|
||||
paginated_objects = objects[start_idx:end_idx]
|
||||
|
||||
# Determine pagination info
|
||||
has_more = end_idx < len(objects)
|
||||
first_id = paginated_objects[0].id if paginated_objects else None
|
||||
last_id = paginated_objects[-1].id if paginated_objects else None
|
||||
|
||||
return paginated_objects, has_more, first_id, last_id
|
||||
|
||||
async def openai_retrieve_vector_store_file_batch(
|
||||
self,
|
||||
batch_id: str,
|
||||
vector_store_id: str,
|
||||
) -> VectorStoreFileBatchObject:
|
||||
"""Retrieve a vector store file batch."""
|
||||
batch_info = self._get_and_validate_batch(batch_id, vector_store_id)
|
||||
return VectorStoreFileBatchObject(**batch_info)
|
||||
|
||||
async def openai_list_files_in_vector_store_file_batch(
|
||||
self,
|
||||
|
@ -841,15 +1201,39 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
order: str | None = "desc",
|
||||
) -> VectorStoreFilesListInBatchResponse:
|
||||
"""Returns a list of vector store files in a batch."""
|
||||
raise NotImplementedError("openai_list_files_in_vector_store_file_batch is not implemented yet")
|
||||
batch_info = self._get_and_validate_batch(batch_id, vector_store_id)
|
||||
batch_file_ids = batch_info["file_ids"]
|
||||
|
||||
async def openai_retrieve_vector_store_file_batch(
|
||||
self,
|
||||
batch_id: str,
|
||||
vector_store_id: str,
|
||||
) -> VectorStoreFileBatchObject:
|
||||
"""Retrieve a vector store file batch."""
|
||||
raise NotImplementedError("openai_retrieve_vector_store_file_batch is not implemented yet")
|
||||
# Load file objects for files in this batch
|
||||
batch_file_objects = []
|
||||
|
||||
for file_id in batch_file_ids:
|
||||
try:
|
||||
file_info = await self._load_openai_vector_store_file(vector_store_id, file_id)
|
||||
file_object = VectorStoreFileObject(**file_info)
|
||||
|
||||
# Apply status filter if provided
|
||||
if filter and file_object.status != filter:
|
||||
continue
|
||||
|
||||
batch_file_objects.append(file_object)
|
||||
except Exception as e:
|
||||
logger.warning(f"Could not load file {file_id} from batch {batch_id}: {e}")
|
||||
continue
|
||||
|
||||
# Sort by created_at
|
||||
reverse_order = order == "desc"
|
||||
batch_file_objects.sort(key=lambda x: x.created_at, reverse=reverse_order)
|
||||
|
||||
# Apply pagination using helper
|
||||
paginated_files, has_more, first_id, last_id = self._paginate_objects(batch_file_objects, limit, after, before)
|
||||
|
||||
return VectorStoreFilesListInBatchResponse(
|
||||
data=paginated_files,
|
||||
first_id=first_id,
|
||||
last_id=last_id,
|
||||
has_more=has_more,
|
||||
)
|
||||
|
||||
async def openai_cancel_vector_store_file_batch(
|
||||
self,
|
||||
|
@ -857,4 +1241,24 @@ class OpenAIVectorStoreMixin(ABC):
|
|||
vector_store_id: str,
|
||||
) -> VectorStoreFileBatchObject:
|
||||
"""Cancel a vector store file batch."""
|
||||
raise NotImplementedError("openai_cancel_vector_store_file_batch is not implemented yet")
|
||||
batch_info = self._get_and_validate_batch(batch_id, vector_store_id)
|
||||
|
||||
if batch_info["status"] not in ["in_progress"]:
|
||||
raise ValueError(f"Cannot cancel batch {batch_id} with status {batch_info['status']}")
|
||||
|
||||
# Cancel the actual processing task if it exists
|
||||
if batch_id in self._file_batch_tasks:
|
||||
task = self._file_batch_tasks[batch_id]
|
||||
if not task.done():
|
||||
task.cancel()
|
||||
logger.info(f"Cancelled processing task for file batch: {batch_id}")
|
||||
# Remove from task tracking
|
||||
del self._file_batch_tasks[batch_id]
|
||||
|
||||
batch_info["status"] = "cancelled"
|
||||
|
||||
await self._save_openai_vector_store_file_batch(batch_id, batch_info)
|
||||
|
||||
updated_batch = VectorStoreFileBatchObject(**batch_info)
|
||||
|
||||
return updated_batch
|
||||
|
|
|
@ -221,8 +221,8 @@ fi
|
|||
cmd=( run -d "${PLATFORM_OPTS[@]}" --name llama-stack \
|
||||
--network llama-net \
|
||||
-p "${PORT}:${PORT}" \
|
||||
"${SERVER_IMAGE}" --port "${PORT}" \
|
||||
--env OLLAMA_URL="http://ollama-server:${OLLAMA_PORT}")
|
||||
-e OLLAMA_URL="http://ollama-server:${OLLAMA_PORT}" \
|
||||
"${SERVER_IMAGE}" --port "${PORT}")
|
||||
|
||||
log "🦙 Starting Llama Stack..."
|
||||
if ! execute_with_log $ENGINE "${cmd[@]}"; then
|
||||
|
|
|
@ -186,7 +186,7 @@ if [[ "$STACK_CONFIG" == *"server:"* ]]; then
|
|||
echo "Llama Stack Server is already running, skipping start"
|
||||
else
|
||||
echo "=== Starting Llama Stack Server ==="
|
||||
nohup llama stack run ci-tests --image-type venv > server.log 2>&1 &
|
||||
nohup llama stack run ci-tests > server.log 2>&1 &
|
||||
|
||||
echo "Waiting for Llama Stack Server to start..."
|
||||
for i in {1..30}; do
|
||||
|
|
|
@ -14,23 +14,22 @@
|
|||
"__data__": {
|
||||
"models": [
|
||||
{
|
||||
"model": "llama3.2:3b",
|
||||
"name": "llama3.2:3b",
|
||||
"digest": "a80c4f17acd55265feec403c7aef86be0c25983ab279d83f3bcd3abbcb5b8b72",
|
||||
"expires_at": "2025-10-04T12:20:09.202126-07:00",
|
||||
"size": 3367856128,
|
||||
"size_vram": 3367856128,
|
||||
"model": "all-minilm:l6-v2",
|
||||
"name": "all-minilm:l6-v2",
|
||||
"digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef",
|
||||
"expires_at": "2025-10-06T16:41:45.231544-07:00",
|
||||
"size": 590204928,
|
||||
"size_vram": 590204928,
|
||||
"details": {
|
||||
"parent_model": "",
|
||||
"format": "gguf",
|
||||
"family": "llama",
|
||||
"family": "bert",
|
||||
"families": [
|
||||
"llama"
|
||||
"bert"
|
||||
],
|
||||
"parameter_size": "3.2B",
|
||||
"quantization_level": "Q4_K_M"
|
||||
},
|
||||
"context_length": 4096
|
||||
"parameter_size": "23M",
|
||||
"quantization_level": "F16"
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
|
|
|
@ -0,0 +1,807 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "nomic-embed-text:latest",
|
||||
"input": [
|
||||
"This is the content of test file 2"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "nomic-embed-text:latest"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
0.036525182,
|
||||
-0.0072787926,
|
||||
-0.15320797,
|
||||
-0.028591355,
|
||||
0.028115708,
|
||||
-0.0033384967,
|
||||
0.021494914,
|
||||
-0.023697548,
|
||||
-0.059428893,
|
||||
-0.04412936,
|
||||
-0.014445912,
|
||||
0.06520278,
|
||||
0.013848802,
|
||||
-0.029918822,
|
||||
-0.022794332,
|
||||
-0.012578859,
|
||||
0.060358867,
|
||||
-0.031223036,
|
||||
-0.012306958,
|
||||
-0.028883344,
|
||||
-0.014677056,
|
||||
-0.024171423,
|
||||
-0.047258105,
|
||||
-0.019668069,
|
||||
0.10096786,
|
||||
0.042677063,
|
||||
-0.012945782,
|
||||
0.05772575,
|
||||
-0.09481949,
|
||||
-0.013522372,
|
||||
0.058091108,
|
||||
-0.035321448,
|
||||
0.02014728,
|
||||
-0.06760144,
|
||||
-0.012323442,
|
||||
-0.045392025,
|
||||
0.04685687,
|
||||
0.024927035,
|
||||
-0.0017673819,
|
||||
0.036423087,
|
||||
-0.020881223,
|
||||
-0.010788712,
|
||||
-0.01838111,
|
||||
-0.007801951,
|
||||
-0.011164214,
|
||||
-0.022797823,
|
||||
-0.01222212,
|
||||
0.05638416,
|
||||
-0.01662934,
|
||||
-0.04117776,
|
||||
0.004534807,
|
||||
0.019233605,
|
||||
-0.008680182,
|
||||
0.03177389,
|
||||
0.06082287,
|
||||
-0.010224672,
|
||||
-0.018689552,
|
||||
-0.019074611,
|
||||
0.029412521,
|
||||
-0.06990004,
|
||||
0.054043073,
|
||||
0.027053045,
|
||||
-0.049923293,
|
||||
0.058975294,
|
||||
0.0018301148,
|
||||
-0.06718531,
|
||||
-0.044889167,
|
||||
0.032761537,
|
||||
-0.022604113,
|
||||
0.043496683,
|
||||
0.08500273,
|
||||
0.008184364,
|
||||
0.0011824819,
|
||||
-0.0417156,
|
||||
-0.015855035,
|
||||
-0.06935983,
|
||||
0.01533393,
|
||||
-0.03297617,
|
||||
-0.043794934,
|
||||
0.008973833,
|
||||
0.0415081,
|
||||
0.037018586,
|
||||
0.004035694,
|
||||
0.0067175985,
|
||||
0.058073524,
|
||||
-0.033033613,
|
||||
-0.049569633,
|
||||
-0.011724681,
|
||||
-0.0049699075,
|
||||
0.04405061,
|
||||
0.02349984,
|
||||
0.049434863,
|
||||
0.05952279,
|
||||
0.007926657,
|
||||
-0.019564645,
|
||||
0.028824113,
|
||||
0.030559592,
|
||||
0.044332445,
|
||||
-0.03705847,
|
||||
0.013914022,
|
||||
-0.01584405,
|
||||
0.012503536,
|
||||
0.039434463,
|
||||
0.020927113,
|
||||
0.02458832,
|
||||
0.033364173,
|
||||
-0.0013068066,
|
||||
0.025197528,
|
||||
-0.05292493,
|
||||
0.010358094,
|
||||
-0.018871995,
|
||||
0.039280638,
|
||||
-0.048534855,
|
||||
0.004642058,
|
||||
0.011491514,
|
||||
-0.036328327,
|
||||
0.0637683,
|
||||
-0.0360448,
|
||||
-0.04317744,
|
||||
0.03721341,
|
||||
0.009880729,
|
||||
-0.032810695,
|
||||
0.012197031,
|
||||
0.06644975,
|
||||
0.04497407,
|
||||
0.0018043267,
|
||||
-0.076117076,
|
||||
0.0028520897,
|
||||
0.025521474,
|
||||
-0.04780887,
|
||||
-0.015784036,
|
||||
-0.004914463,
|
||||
-0.0003810333,
|
||||
-0.008213055,
|
||||
-0.0040868036,
|
||||
0.0026211734,
|
||||
0.005037653,
|
||||
-0.0054035867,
|
||||
-0.054472372,
|
||||
-0.04214955,
|
||||
-0.036636207,
|
||||
0.005277914,
|
||||
0.025802922,
|
||||
0.054448027,
|
||||
0.010910778,
|
||||
-0.019098252,
|
||||
0.06248315,
|
||||
0.019785397,
|
||||
-0.02148464,
|
||||
-0.023303429,
|
||||
0.0124828555,
|
||||
-0.02455258,
|
||||
0.0053893207,
|
||||
0.006655952,
|
||||
0.020618292,
|
||||
-0.05195774,
|
||||
0.001724354,
|
||||
-0.049451906,
|
||||
0.031900283,
|
||||
0.08541784,
|
||||
0.02900063,
|
||||
0.006208959,
|
||||
-0.009606019,
|
||||
0.0030572556,
|
||||
-0.018463623,
|
||||
0.014401457,
|
||||
0.0007510511,
|
||||
0.08289015,
|
||||
0.062720075,
|
||||
-0.010840198,
|
||||
-0.04971401,
|
||||
-0.038808372,
|
||||
0.0044536674,
|
||||
0.011472072,
|
||||
-0.031167375,
|
||||
-0.031224154,
|
||||
0.011706744,
|
||||
-0.022990009,
|
||||
0.04747808,
|
||||
-0.0016337503,
|
||||
0.015181135,
|
||||
0.005154193,
|
||||
0.00949444,
|
||||
0.042812645,
|
||||
0.001253686,
|
||||
-0.050080713,
|
||||
0.038098942,
|
||||
-0.014367589,
|
||||
-0.043111958,
|
||||
-0.0059632747,
|
||||
-0.022602718,
|
||||
-0.0042201183,
|
||||
-0.09451348,
|
||||
-0.042164654,
|
||||
-0.010821582,
|
||||
-0.04681359,
|
||||
0.016275495,
|
||||
-0.0033313567,
|
||||
0.027538816,
|
||||
-0.019907625,
|
||||
0.00040033093,
|
||||
-0.030824887,
|
||||
-0.058938056,
|
||||
0.0014922265,
|
||||
-0.027667042,
|
||||
0.015573365,
|
||||
-0.04173136,
|
||||
-0.015453809,
|
||||
-0.01595607,
|
||||
0.03898053,
|
||||
0.043484144,
|
||||
0.0075124763,
|
||||
-0.0025220348,
|
||||
0.038111743,
|
||||
0.041447856,
|
||||
-0.011153068,
|
||||
-0.01717726,
|
||||
-0.045249123,
|
||||
-0.010734678,
|
||||
-0.03552057,
|
||||
0.033035237,
|
||||
-0.0077519426,
|
||||
0.048082184,
|
||||
-0.06981011,
|
||||
0.034551185,
|
||||
0.011257734,
|
||||
-0.043801457,
|
||||
-0.018373946,
|
||||
-0.04797999,
|
||||
-0.017102923,
|
||||
0.0029698398,
|
||||
-0.09975526,
|
||||
0.00053959514,
|
||||
0.0074329274,
|
||||
-0.018584883,
|
||||
-0.0094977375,
|
||||
-0.05056549,
|
||||
0.08929669,
|
||||
0.011828429,
|
||||
0.040005405,
|
||||
-0.03369542,
|
||||
0.07867971,
|
||||
0.025032107,
|
||||
0.016890414,
|
||||
0.014425124,
|
||||
0.00064274436,
|
||||
0.009868133,
|
||||
-0.034772366,
|
||||
0.05254746,
|
||||
0.071544185,
|
||||
-0.01852601,
|
||||
-0.0013607002,
|
||||
0.010325862,
|
||||
0.0647086,
|
||||
0.013452749,
|
||||
-0.009807788,
|
||||
-0.01738053,
|
||||
-0.012833702,
|
||||
-0.0037767375,
|
||||
-0.046967912,
|
||||
0.017845146,
|
||||
-0.0682881,
|
||||
0.011557345,
|
||||
0.01458601,
|
||||
-0.048856564,
|
||||
-0.01423403,
|
||||
-0.03424404,
|
||||
0.021640293,
|
||||
-0.025939032,
|
||||
-0.001273354,
|
||||
0.0033471219,
|
||||
0.02255794,
|
||||
-0.05386608,
|
||||
0.02134696,
|
||||
0.012213072,
|
||||
-0.027799206,
|
||||
0.041816894,
|
||||
0.013318655,
|
||||
-0.027756989,
|
||||
0.03054267,
|
||||
-0.025455547,
|
||||
0.014977695,
|
||||
0.03629763,
|
||||
0.05029929,
|
||||
0.017317088,
|
||||
0.0008021539,
|
||||
-0.027486524,
|
||||
0.0011794426,
|
||||
0.021061994,
|
||||
0.038059466,
|
||||
0.014114616,
|
||||
0.014319938,
|
||||
0.012650396,
|
||||
0.04102732,
|
||||
0.018222608,
|
||||
0.0115328785,
|
||||
0.043359082,
|
||||
-0.0028082337,
|
||||
-0.016893078,
|
||||
-0.03791571,
|
||||
0.023969462,
|
||||
0.0077467947,
|
||||
0.033167463,
|
||||
0.018768141,
|
||||
0.00804635,
|
||||
-0.05316497,
|
||||
0.021600094,
|
||||
-0.032088757,
|
||||
0.056640208,
|
||||
0.010592809,
|
||||
-0.06282453,
|
||||
-0.003963599,
|
||||
-0.0054780785,
|
||||
0.0057015507,
|
||||
-0.026736109,
|
||||
0.03140229,
|
||||
0.021742998,
|
||||
0.037487593,
|
||||
0.04916904,
|
||||
-0.015454876,
|
||||
0.0036427178,
|
||||
-0.06809397,
|
||||
-0.005600329,
|
||||
0.006426826,
|
||||
0.029163402,
|
||||
0.008698685,
|
||||
0.013447198,
|
||||
0.028116653,
|
||||
-0.032959465,
|
||||
-0.046715226,
|
||||
0.062885955,
|
||||
0.07805104,
|
||||
-0.075704284,
|
||||
-0.026722923,
|
||||
0.031568483,
|
||||
0.029869428,
|
||||
0.014207811,
|
||||
0.058283728,
|
||||
-0.0009454238,
|
||||
0.049990628,
|
||||
0.09433687,
|
||||
0.011483032,
|
||||
0.0073822956,
|
||||
0.001765557,
|
||||
0.014384013,
|
||||
-0.0805711,
|
||||
-0.057262138,
|
||||
0.0033087756,
|
||||
0.017576102,
|
||||
0.050261848,
|
||||
-0.0058530914,
|
||||
-0.00258757,
|
||||
0.009722071,
|
||||
0.0044941446,
|
||||
0.009631424,
|
||||
0.027689122,
|
||||
0.012394503,
|
||||
-0.04055002,
|
||||
0.055514883,
|
||||
-0.028808117,
|
||||
0.0297643,
|
||||
-0.034311485,
|
||||
0.021378465,
|
||||
-0.033280674,
|
||||
0.019441161,
|
||||
-0.009369208,
|
||||
0.0030489776,
|
||||
-0.016572703,
|
||||
0.042294934,
|
||||
0.015723946,
|
||||
0.0022674324,
|
||||
-0.0014906601,
|
||||
0.01840701,
|
||||
0.059862193,
|
||||
0.053135127,
|
||||
0.020754104,
|
||||
-0.06374346,
|
||||
0.001787633,
|
||||
-0.036681958,
|
||||
0.03553359,
|
||||
0.06609074,
|
||||
-0.0107706385,
|
||||
0.045129295,
|
||||
0.06838197,
|
||||
0.025984539,
|
||||
-0.06558362,
|
||||
0.027897354,
|
||||
-0.00621841,
|
||||
0.03920637,
|
||||
0.009362378,
|
||||
-0.062093496,
|
||||
0.021269219,
|
||||
-0.06091154,
|
||||
-0.027098468,
|
||||
0.008638457,
|
||||
-0.050488967,
|
||||
0.04693317,
|
||||
0.043328438,
|
||||
-0.025587596,
|
||||
0.03407469,
|
||||
-0.048816204,
|
||||
-0.004734613,
|
||||
-0.0008902356,
|
||||
0.024133636,
|
||||
-0.022534605,
|
||||
0.035635088,
|
||||
-0.053277653,
|
||||
-0.055609506,
|
||||
0.0523981,
|
||||
0.0014473854,
|
||||
0.032570753,
|
||||
-0.005762427,
|
||||
-0.016173452,
|
||||
-0.06672014,
|
||||
0.0013724786,
|
||||
0.007844828,
|
||||
0.02429992,
|
||||
0.0032019925,
|
||||
0.0016553001,
|
||||
-0.022802994,
|
||||
0.001800882,
|
||||
0.032480165,
|
||||
-0.002195562,
|
||||
-0.03154405,
|
||||
-0.013679192,
|
||||
-0.011184489,
|
||||
0.033688888,
|
||||
0.04774288,
|
||||
0.0018061483,
|
||||
-0.09035719,
|
||||
-0.0047670994,
|
||||
-0.02052915,
|
||||
0.036272082,
|
||||
0.020193182,
|
||||
-0.036813166,
|
||||
0.039460275,
|
||||
-0.015967365,
|
||||
-0.0033895948,
|
||||
-0.031828586,
|
||||
0.053221144,
|
||||
0.021549668,
|
||||
-0.07595095,
|
||||
-0.044737455,
|
||||
-0.010761814,
|
||||
0.0025511624,
|
||||
0.14498504,
|
||||
0.08222001,
|
||||
-0.037528154,
|
||||
-0.032176156,
|
||||
0.013683398,
|
||||
0.01410672,
|
||||
0.019557275,
|
||||
0.062485218,
|
||||
0.027925756,
|
||||
0.079192385,
|
||||
-0.026622739,
|
||||
0.02323037,
|
||||
-0.016175434,
|
||||
-0.032527965,
|
||||
-0.008870566,
|
||||
-0.009013046,
|
||||
-0.009945577,
|
||||
0.025208296,
|
||||
0.0073141777,
|
||||
0.044331536,
|
||||
-0.020921648,
|
||||
-0.016868133,
|
||||
-0.026842397,
|
||||
0.03165012,
|
||||
0.043120645,
|
||||
-0.048179835,
|
||||
-0.05591947,
|
||||
0.029399967,
|
||||
-0.069223806,
|
||||
0.03508237,
|
||||
0.00804212,
|
||||
-0.041150257,
|
||||
0.008898182,
|
||||
0.0006015489,
|
||||
0.023109462,
|
||||
0.027766718,
|
||||
0.012039964,
|
||||
-0.030886615,
|
||||
-0.030401329,
|
||||
0.038484607,
|
||||
-0.0247026,
|
||||
0.0018090954,
|
||||
0.028525416,
|
||||
0.054761168,
|
||||
-0.0062592058,
|
||||
0.029739086,
|
||||
0.033199638,
|
||||
0.0488184,
|
||||
0.028191078,
|
||||
-0.020734766,
|
||||
0.00060847827,
|
||||
0.029920708,
|
||||
-0.0490555,
|
||||
0.007290553,
|
||||
0.0026984178,
|
||||
0.063341014,
|
||||
0.018249765,
|
||||
0.019682994,
|
||||
0.0063302247,
|
||||
-0.029094942,
|
||||
-0.030193835,
|
||||
0.042414594,
|
||||
-0.05859321,
|
||||
-0.09094711,
|
||||
-0.025345713,
|
||||
-0.034041878,
|
||||
-0.014829038,
|
||||
0.0030920506,
|
||||
0.015670862,
|
||||
0.073476,
|
||||
0.017715238,
|
||||
0.052982714,
|
||||
0.012198469,
|
||||
-0.021962965,
|
||||
0.017349334,
|
||||
0.025136312,
|
||||
0.006353252,
|
||||
0.03436416,
|
||||
-0.01633907,
|
||||
-0.08311436,
|
||||
0.04788054,
|
||||
0.0032672018,
|
||||
-0.0318856,
|
||||
0.06784985,
|
||||
0.072452076,
|
||||
0.009116457,
|
||||
0.017004106,
|
||||
-0.040795434,
|
||||
-0.023130566,
|
||||
-0.0017866351,
|
||||
-0.020753238,
|
||||
-0.028738804,
|
||||
0.0031001552,
|
||||
-0.012533389,
|
||||
0.047431413,
|
||||
-0.059432007,
|
||||
-0.019904893,
|
||||
0.009464013,
|
||||
-0.016388606,
|
||||
0.028543858,
|
||||
-0.026128467,
|
||||
-0.03368374,
|
||||
-0.0040021804,
|
||||
3.1505784e-05,
|
||||
-0.10005339,
|
||||
0.020524276,
|
||||
-0.06320255,
|
||||
-0.026909621,
|
||||
-0.009929203,
|
||||
0.03084924,
|
||||
-0.041759893,
|
||||
0.02034976,
|
||||
-0.008311877,
|
||||
-0.0042031757,
|
||||
-0.04709363,
|
||||
0.030620687,
|
||||
-0.028947143,
|
||||
-0.007556809,
|
||||
0.01617724,
|
||||
0.037857477,
|
||||
-0.039480377,
|
||||
-0.008805032,
|
||||
0.051410846,
|
||||
0.017079966,
|
||||
0.0032464731,
|
||||
0.023022559,
|
||||
-0.017350538,
|
||||
0.03471975,
|
||||
-0.02863222,
|
||||
-0.024592673,
|
||||
-0.0077179587,
|
||||
0.03141146,
|
||||
0.03583118,
|
||||
-0.0130302245,
|
||||
-0.057425633,
|
||||
0.040003538,
|
||||
-0.0046423534,
|
||||
0.019725544,
|
||||
0.0397109,
|
||||
-0.0025461344,
|
||||
0.046675395,
|
||||
0.011516851,
|
||||
-0.029444098,
|
||||
0.03419632,
|
||||
-0.043872464,
|
||||
-0.021072017,
|
||||
-0.010389852,
|
||||
0.01248914,
|
||||
-0.03476949,
|
||||
0.02083105,
|
||||
-0.021170666,
|
||||
-0.010824939,
|
||||
-0.034223318,
|
||||
0.0008804664,
|
||||
-0.00975538,
|
||||
-0.004145119,
|
||||
0.0062736045,
|
||||
0.017810361,
|
||||
-0.05057402,
|
||||
0.0028876425,
|
||||
-0.012459405,
|
||||
0.024415256,
|
||||
-0.009684934,
|
||||
-0.032268245,
|
||||
-1.0135974e-05,
|
||||
0.015377202,
|
||||
-0.008089165,
|
||||
-0.08534785,
|
||||
0.011209079,
|
||||
-0.006432232,
|
||||
-0.05970185,
|
||||
0.03646468,
|
||||
-0.024002092,
|
||||
-0.022855703,
|
||||
-0.051673587,
|
||||
0.038473092,
|
||||
-0.028756764,
|
||||
0.041329525,
|
||||
-0.06377881,
|
||||
-0.014500157,
|
||||
-0.018372798,
|
||||
-0.008677442,
|
||||
0.036858637,
|
||||
0.038448237,
|
||||
0.044321943,
|
||||
-0.046770208,
|
||||
0.026638264,
|
||||
-0.04069364,
|
||||
-0.051563717,
|
||||
-0.054425545,
|
||||
-0.007966239,
|
||||
-0.045169767,
|
||||
-0.0006122694,
|
||||
0.013411372,
|
||||
0.04263278,
|
||||
0.03749003,
|
||||
0.010722818,
|
||||
-0.041889716,
|
||||
-0.036726084,
|
||||
0.014166507,
|
||||
0.038341004,
|
||||
0.004509263,
|
||||
0.035988707,
|
||||
0.02634235,
|
||||
-0.02256134,
|
||||
0.08171513,
|
||||
0.09104147,
|
||||
0.06757358,
|
||||
-0.0016213343,
|
||||
-0.018941583,
|
||||
-0.0014519675,
|
||||
0.03409365,
|
||||
-0.060576558,
|
||||
-0.028001321,
|
||||
-0.08352477,
|
||||
0.011629786,
|
||||
0.014637305,
|
||||
-0.021191692,
|
||||
0.009192876,
|
||||
0.0025693115,
|
||||
0.03831378,
|
||||
-0.00035758872,
|
||||
-0.032391928,
|
||||
0.006118005,
|
||||
-0.05773841,
|
||||
0.033030152,
|
||||
0.03268179,
|
||||
0.031052263,
|
||||
-0.0018795256,
|
||||
-0.0463158,
|
||||
0.017675944,
|
||||
0.039604764,
|
||||
0.056545958,
|
||||
-0.002072885,
|
||||
-0.0374349,
|
||||
-0.014934615,
|
||||
-0.046360567,
|
||||
0.060439337,
|
||||
-5.3795357e-05,
|
||||
0.027416907,
|
||||
-0.08041611,
|
||||
0.00016825287,
|
||||
-0.08668716,
|
||||
-0.03210328,
|
||||
0.016515074,
|
||||
-0.0062253834,
|
||||
-0.00093463395,
|
||||
-0.027180947,
|
||||
-0.049670145,
|
||||
-0.033094753,
|
||||
-0.0051170597,
|
||||
0.031569846,
|
||||
-0.014995825,
|
||||
-0.016850019,
|
||||
0.04239559,
|
||||
0.020676404,
|
||||
0.0319851,
|
||||
-0.008854607,
|
||||
0.04452473,
|
||||
-0.023021534,
|
||||
0.007295005,
|
||||
0.05227394,
|
||||
0.0040576537,
|
||||
-0.0655794,
|
||||
-0.067981854,
|
||||
0.03440047,
|
||||
0.009278226,
|
||||
-0.0282169,
|
||||
0.060756575,
|
||||
-0.020904224,
|
||||
0.01505642,
|
||||
-0.0045534745,
|
||||
0.018723203,
|
||||
0.0035253377,
|
||||
0.011872832,
|
||||
0.042355374,
|
||||
0.017724074,
|
||||
-0.060881015,
|
||||
0.010464869,
|
||||
-0.015852634,
|
||||
-0.03751531,
|
||||
0.022855468,
|
||||
-0.037866883,
|
||||
0.05328077,
|
||||
-0.0320521,
|
||||
-0.030350016,
|
||||
-0.010912554,
|
||||
-0.012704745,
|
||||
0.0076516517,
|
||||
0.0014142905,
|
||||
0.011725254,
|
||||
0.0067488,
|
||||
-0.008221275,
|
||||
0.01648301,
|
||||
-0.013712469,
|
||||
0.0129476935,
|
||||
0.028405288,
|
||||
-0.011489568,
|
||||
-0.006695754,
|
||||
-0.07523671,
|
||||
0.0012562524,
|
||||
-0.051538818,
|
||||
0.017514601,
|
||||
0.03280067,
|
||||
-0.018965578,
|
||||
0.009017527,
|
||||
-0.052108284,
|
||||
0.0017074383,
|
||||
0.00056099903,
|
||||
0.008343997,
|
||||
-0.01674154,
|
||||
-0.012425597,
|
||||
-0.00041037227,
|
||||
0.1104793,
|
||||
-0.015096156,
|
||||
0.014880369,
|
||||
-0.0098567465,
|
||||
0.024937985,
|
||||
0.0112802675,
|
||||
-0.0010737488,
|
||||
-0.06354736,
|
||||
-3.862344e-05,
|
||||
-0.024247888
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "nomic-embed-text:latest",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 8,
|
||||
"total_tokens": 8
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,807 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "nomic-embed-text:latest",
|
||||
"input": [
|
||||
"This is batch test file 1"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "nomic-embed-text:latest"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
0.01183041,
|
||||
-0.0065989625,
|
||||
-0.159677,
|
||||
0.011660306,
|
||||
0.055617318,
|
||||
-0.03764695,
|
||||
0.0163666,
|
||||
0.033777084,
|
||||
-0.06433634,
|
||||
-0.08037781,
|
||||
-0.0057114926,
|
||||
0.07607082,
|
||||
0.033490222,
|
||||
0.048497792,
|
||||
-0.048456103,
|
||||
-0.049539,
|
||||
0.059783153,
|
||||
-0.08439946,
|
||||
0.0076269372,
|
||||
-0.0128732305,
|
||||
0.05902644,
|
||||
0.012931591,
|
||||
-0.08323305,
|
||||
-0.00037215627,
|
||||
0.13830419,
|
||||
0.024290211,
|
||||
-0.047809705,
|
||||
0.039520696,
|
||||
-0.06423598,
|
||||
-0.01653946,
|
||||
0.03764018,
|
||||
-0.001062501,
|
||||
0.028489634,
|
||||
-0.025925444,
|
||||
-0.015699588,
|
||||
-0.012715725,
|
||||
0.017358005,
|
||||
-0.007198467,
|
||||
0.059812553,
|
||||
0.028332362,
|
||||
-0.00015984774,
|
||||
0.004483949,
|
||||
0.034580402,
|
||||
-0.054280724,
|
||||
-0.002989754,
|
||||
0.023461882,
|
||||
0.011839507,
|
||||
0.018908013,
|
||||
0.016710319,
|
||||
0.004905327,
|
||||
-0.0107955905,
|
||||
-0.01565778,
|
||||
-0.04169478,
|
||||
0.02510759,
|
||||
0.026486792,
|
||||
0.01054831,
|
||||
0.011289881,
|
||||
0.038714606,
|
||||
-0.0136384675,
|
||||
-0.023249293,
|
||||
0.014086617,
|
||||
0.018654121,
|
||||
-0.07146624,
|
||||
0.047506154,
|
||||
-0.012085512,
|
||||
-0.007589288,
|
||||
-0.009515838,
|
||||
0.0048574316,
|
||||
-0.004600554,
|
||||
0.0031499087,
|
||||
0.06778753,
|
||||
-0.019641325,
|
||||
0.018102348,
|
||||
-0.01726182,
|
||||
-0.003802732,
|
||||
-0.04414122,
|
||||
-0.010491107,
|
||||
-0.065158285,
|
||||
-0.045328394,
|
||||
-0.0019480857,
|
||||
0.052318677,
|
||||
0.0386049,
|
||||
0.020296056,
|
||||
0.044793047,
|
||||
0.08282699,
|
||||
-0.019911101,
|
||||
-0.016511027,
|
||||
-0.0062321154,
|
||||
-0.025036003,
|
||||
0.04578435,
|
||||
0.0019149093,
|
||||
0.025694296,
|
||||
-0.0042011673,
|
||||
-0.018107908,
|
||||
-0.026668591,
|
||||
0.018340195,
|
||||
0.010810087,
|
||||
0.018672433,
|
||||
-0.006774911,
|
||||
-0.0026458725,
|
||||
0.023082372,
|
||||
0.027705511,
|
||||
0.019753877,
|
||||
-0.03543464,
|
||||
-0.0061461334,
|
||||
0.0155549655,
|
||||
-0.019579103,
|
||||
-0.00693201,
|
||||
-0.06635246,
|
||||
-0.015482261,
|
||||
-0.0040295934,
|
||||
0.0006957319,
|
||||
-0.008871345,
|
||||
-0.00842857,
|
||||
0.031484608,
|
||||
-0.010076284,
|
||||
0.06257018,
|
||||
0.0012318427,
|
||||
-0.024530765,
|
||||
0.00015912329,
|
||||
0.0033331378,
|
||||
-0.032083686,
|
||||
-0.007399188,
|
||||
0.07031288,
|
||||
0.033552274,
|
||||
0.061820872,
|
||||
-0.09171231,
|
||||
0.036374647,
|
||||
0.007984676,
|
||||
-0.031679634,
|
||||
0.00598418,
|
||||
-0.0029291043,
|
||||
-0.0049730917,
|
||||
-0.052057285,
|
||||
0.020125173,
|
||||
0.009004486,
|
||||
-0.022456508,
|
||||
-0.012051283,
|
||||
-0.03740793,
|
||||
-0.027594674,
|
||||
-0.02012376,
|
||||
0.011664398,
|
||||
0.04336321,
|
||||
0.061720803,
|
||||
0.041055538,
|
||||
-0.02444171,
|
||||
0.024476659,
|
||||
0.030615946,
|
||||
-0.01689858,
|
||||
0.0091607245,
|
||||
0.0038629547,
|
||||
-0.0019203863,
|
||||
-0.0035829302,
|
||||
0.021674454,
|
||||
0.037874587,
|
||||
-0.057554636,
|
||||
0.014823112,
|
||||
0.0036189007,
|
||||
0.012866306,
|
||||
0.051631145,
|
||||
0.0021970836,
|
||||
-0.033981066,
|
||||
-0.03782387,
|
||||
0.01235394,
|
||||
-0.057634324,
|
||||
-0.07556398,
|
||||
0.008977255,
|
||||
0.07841102,
|
||||
0.060794022,
|
||||
-0.03463157,
|
||||
-0.063551195,
|
||||
-0.064811006,
|
||||
0.010709957,
|
||||
-0.027145889,
|
||||
-0.0837886,
|
||||
-0.035913587,
|
||||
0.017231362,
|
||||
-0.01455278,
|
||||
0.039031487,
|
||||
-0.038145658,
|
||||
0.023733672,
|
||||
-0.019787688,
|
||||
0.020164428,
|
||||
0.023367887,
|
||||
0.0035691075,
|
||||
-0.028722964,
|
||||
0.014704597,
|
||||
-0.019744202,
|
||||
-0.06668101,
|
||||
-0.017812628,
|
||||
-0.009186517,
|
||||
-0.033119973,
|
||||
-0.085967295,
|
||||
-0.080312125,
|
||||
0.013302178,
|
||||
-0.061551016,
|
||||
0.017130975,
|
||||
-0.017442413,
|
||||
0.04742156,
|
||||
-0.013023663,
|
||||
-0.013847287,
|
||||
-0.01880652,
|
||||
-0.07011646,
|
||||
0.018233122,
|
||||
-0.030537246,
|
||||
-0.026766777,
|
||||
-0.012263141,
|
||||
0.014689888,
|
||||
-0.049961388,
|
||||
0.03201573,
|
||||
0.015774516,
|
||||
-0.020335846,
|
||||
-0.038940914,
|
||||
0.0065977564,
|
||||
0.035997562,
|
||||
-0.053227507,
|
||||
0.008883548,
|
||||
-0.039375745,
|
||||
-0.017865263,
|
||||
0.007343183,
|
||||
0.017375462,
|
||||
0.021595728,
|
||||
0.057712954,
|
||||
-0.040693924,
|
||||
-0.000778912,
|
||||
-0.018082067,
|
||||
-0.015103824,
|
||||
-0.024191063,
|
||||
-0.0077742958,
|
||||
-0.034330968,
|
||||
-0.020159615,
|
||||
-0.03245423,
|
||||
0.0020437704,
|
||||
-0.000114842755,
|
||||
-0.029564297,
|
||||
-0.018030599,
|
||||
-0.0031425157,
|
||||
0.053831782,
|
||||
-0.026106073,
|
||||
0.04243461,
|
||||
-0.048363626,
|
||||
0.025711408,
|
||||
-0.008338205,
|
||||
0.0009197218,
|
||||
-0.011072695,
|
||||
0.00031293565,
|
||||
0.0033421176,
|
||||
-0.007302082,
|
||||
0.04127773,
|
||||
-0.0074836435,
|
||||
-0.04299338,
|
||||
-0.002760089,
|
||||
0.019094143,
|
||||
0.039009947,
|
||||
0.03581834,
|
||||
-0.032022007,
|
||||
-0.009045915,
|
||||
-0.03275861,
|
||||
0.017295409,
|
||||
-0.039618656,
|
||||
0.015396318,
|
||||
-0.07593323,
|
||||
0.03475173,
|
||||
0.007710904,
|
||||
-0.009037294,
|
||||
-0.026630195,
|
||||
-0.027383188,
|
||||
0.02212514,
|
||||
-0.035001624,
|
||||
-0.0219445,
|
||||
-0.01212384,
|
||||
-0.0018017493,
|
||||
-0.011781174,
|
||||
-0.051410057,
|
||||
0.026306989,
|
||||
0.006329408,
|
||||
0.010307703,
|
||||
0.01613663,
|
||||
-0.006002573,
|
||||
0.031006144,
|
||||
-0.036049806,
|
||||
-0.018159281,
|
||||
-0.012575659,
|
||||
-0.0048318235,
|
||||
0.048996273,
|
||||
-0.0010814993,
|
||||
0.050774954,
|
||||
-0.027395276,
|
||||
0.0115728015,
|
||||
0.031056559,
|
||||
0.011177566,
|
||||
0.012006755,
|
||||
-0.02556873,
|
||||
0.029484332,
|
||||
-0.009657058,
|
||||
0.009322593,
|
||||
0.022122696,
|
||||
-0.018415872,
|
||||
0.010098681,
|
||||
-0.007367993,
|
||||
-0.023805562,
|
||||
0.035959154,
|
||||
0.028602934,
|
||||
0.030718775,
|
||||
0.01705538,
|
||||
-0.024984695,
|
||||
0.042858277,
|
||||
-0.015449,
|
||||
0.005040281,
|
||||
0.038991883,
|
||||
-0.07141338,
|
||||
-0.002947093,
|
||||
-0.044420503,
|
||||
0.019382862,
|
||||
-0.040407836,
|
||||
0.04245461,
|
||||
0.048940845,
|
||||
0.018063093,
|
||||
0.08591597,
|
||||
-0.035389014,
|
||||
-0.010674617,
|
||||
-0.103511095,
|
||||
-0.008537786,
|
||||
0.010264984,
|
||||
-0.003966177,
|
||||
0.02314327,
|
||||
0.0048719845,
|
||||
0.06199085,
|
||||
-0.00810136,
|
||||
-0.039515182,
|
||||
0.05785542,
|
||||
0.06719427,
|
||||
-0.039108012,
|
||||
-0.050833326,
|
||||
0.05823837,
|
||||
0.017042343,
|
||||
0.005815163,
|
||||
0.039831843,
|
||||
-0.012049576,
|
||||
0.076485425,
|
||||
0.012621482,
|
||||
0.06927575,
|
||||
0.05359866,
|
||||
-0.015146923,
|
||||
0.044284295,
|
||||
-0.062355984,
|
||||
-0.009034613,
|
||||
0.04071826,
|
||||
-0.01236521,
|
||||
0.079400524,
|
||||
0.0017920422,
|
||||
-0.011480363,
|
||||
0.008711773,
|
||||
0.018180674,
|
||||
-0.0030674522,
|
||||
0.0326583,
|
||||
0.03525443,
|
||||
-0.02087537,
|
||||
0.05094025,
|
||||
-0.0037492628,
|
||||
0.009178962,
|
||||
-0.0050435406,
|
||||
-0.01166052,
|
||||
0.0060158456,
|
||||
-0.002493798,
|
||||
0.021641793,
|
||||
0.0019783853,
|
||||
0.023140313,
|
||||
0.046997964,
|
||||
0.0069999313,
|
||||
-0.0552795,
|
||||
-0.020092534,
|
||||
0.06467227,
|
||||
0.044829298,
|
||||
0.013295184,
|
||||
0.0377816,
|
||||
-0.046331275,
|
||||
0.01770082,
|
||||
-0.013348137,
|
||||
0.04617519,
|
||||
0.04468347,
|
||||
-0.03253012,
|
||||
0.015447477,
|
||||
0.030224748,
|
||||
-0.0013485672,
|
||||
-0.03615717,
|
||||
0.008698818,
|
||||
-0.0037734164,
|
||||
0.04494809,
|
||||
0.037184346,
|
||||
-0.011223347,
|
||||
0.0046344185,
|
||||
-0.07529732,
|
||||
0.025554653,
|
||||
-0.015140733,
|
||||
-0.0035430966,
|
||||
0.03661124,
|
||||
0.013250649,
|
||||
-0.055586766,
|
||||
0.027562145,
|
||||
-0.018204745,
|
||||
-0.029428158,
|
||||
-0.0029150618,
|
||||
0.03623637,
|
||||
-0.022476854,
|
||||
-0.0058649112,
|
||||
-0.015735915,
|
||||
-0.019995706,
|
||||
0.032269973,
|
||||
0.017872665,
|
||||
0.028031865,
|
||||
-0.043758772,
|
||||
-0.027188994,
|
||||
-0.058870632,
|
||||
0.024894219,
|
||||
0.015318543,
|
||||
0.06244725,
|
||||
0.021922529,
|
||||
0.000678521,
|
||||
-0.025339983,
|
||||
0.025911404,
|
||||
0.01583583,
|
||||
-0.014407775,
|
||||
-0.037194725,
|
||||
-0.015699212,
|
||||
0.008184332,
|
||||
0.014927899,
|
||||
0.0737949,
|
||||
0.007748195,
|
||||
-0.07158831,
|
||||
-0.039901625,
|
||||
0.031431172,
|
||||
0.011147712,
|
||||
0.020828275,
|
||||
-0.035193726,
|
||||
0.05613746,
|
||||
-0.0022006142,
|
||||
0.008007006,
|
||||
0.001472366,
|
||||
0.019893395,
|
||||
0.044233263,
|
||||
-0.02244468,
|
||||
-0.0665883,
|
||||
0.013832251,
|
||||
0.0026457622,
|
||||
0.09737926,
|
||||
0.09575702,
|
||||
-0.04908296,
|
||||
-0.062802345,
|
||||
-0.0095988205,
|
||||
0.008329187,
|
||||
0.041316554,
|
||||
-0.0222064,
|
||||
0.02813126,
|
||||
0.07059441,
|
||||
-0.02560012,
|
||||
0.044651207,
|
||||
-0.027545268,
|
||||
-0.007889025,
|
||||
0.03391235,
|
||||
0.008170332,
|
||||
0.0067786956,
|
||||
0.0615806,
|
||||
0.044006567,
|
||||
0.0056231483,
|
||||
-0.024909342,
|
||||
0.040038925,
|
||||
-0.037021257,
|
||||
0.0010181392,
|
||||
0.058034208,
|
||||
-0.021651162,
|
||||
-0.06021004,
|
||||
0.014830516,
|
||||
-0.050770685,
|
||||
0.010422301,
|
||||
0.0016205559,
|
||||
-0.03166853,
|
||||
0.014091049,
|
||||
-0.002066098,
|
||||
0.02992549,
|
||||
0.013251145,
|
||||
0.011673487,
|
||||
-0.0430954,
|
||||
-0.048110887,
|
||||
0.01493126,
|
||||
0.006862025,
|
||||
0.04188833,
|
||||
0.011692501,
|
||||
0.0465231,
|
||||
0.010624,
|
||||
0.02873104,
|
||||
0.037793215,
|
||||
0.08978305,
|
||||
0.011727344,
|
||||
0.043248493,
|
||||
-0.033803374,
|
||||
0.011249601,
|
||||
-0.015437648,
|
||||
-0.009372223,
|
||||
-0.005403984,
|
||||
-0.009915787,
|
||||
-0.030847883,
|
||||
-0.00076942804,
|
||||
0.018497106,
|
||||
-0.00030310496,
|
||||
-0.0076847905,
|
||||
-0.0036222623,
|
||||
-0.008554024,
|
||||
-0.07606582,
|
||||
-0.024716768,
|
||||
-0.028077196,
|
||||
-0.024249833,
|
||||
0.027158285,
|
||||
0.0075863106,
|
||||
0.09348848,
|
||||
-0.00034073484,
|
||||
0.039915837,
|
||||
-0.007647916,
|
||||
-0.035295825,
|
||||
0.01611119,
|
||||
0.060429912,
|
||||
0.009458672,
|
||||
0.027763832,
|
||||
-0.025683967,
|
||||
-0.091306895,
|
||||
0.0367077,
|
||||
0.009893541,
|
||||
-5.195292e-05,
|
||||
0.045459133,
|
||||
0.04671114,
|
||||
-0.0023683042,
|
||||
0.017460158,
|
||||
-0.007978136,
|
||||
0.00081788,
|
||||
-0.009908127,
|
||||
0.0049076737,
|
||||
-0.03604046,
|
||||
0.024152907,
|
||||
0.0022956813,
|
||||
0.061990347,
|
||||
-0.061900347,
|
||||
0.0047628507,
|
||||
0.007954329,
|
||||
-0.05227117,
|
||||
0.013897867,
|
||||
-0.0034024485,
|
||||
-0.06788996,
|
||||
0.036198605,
|
||||
-0.014600589,
|
||||
-0.038748026,
|
||||
0.031534728,
|
||||
-0.037783317,
|
||||
-0.057816587,
|
||||
-0.054505207,
|
||||
0.010229355,
|
||||
-0.01668772,
|
||||
-0.013999046,
|
||||
-0.049303915,
|
||||
-0.013006012,
|
||||
-0.020143948,
|
||||
0.0009209327,
|
||||
0.010504151,
|
||||
0.052313875,
|
||||
-0.003835063,
|
||||
0.03984861,
|
||||
-0.05403,
|
||||
0.004036369,
|
||||
0.035671517,
|
||||
-0.009310839,
|
||||
0.01921996,
|
||||
0.015426655,
|
||||
-0.042717084,
|
||||
-0.016548151,
|
||||
-0.03559785,
|
||||
-0.03052737,
|
||||
0.0016032697,
|
||||
0.04009937,
|
||||
0.05516244,
|
||||
-0.009645057,
|
||||
-0.019377265,
|
||||
0.017122837,
|
||||
0.007185355,
|
||||
0.012066883,
|
||||
0.015954316,
|
||||
-0.0029309995,
|
||||
-0.008670052,
|
||||
0.0007600626,
|
||||
-0.0019616315,
|
||||
0.03605449,
|
||||
-0.028704248,
|
||||
-0.057372347,
|
||||
-0.03711266,
|
||||
0.02601168,
|
||||
0.020637576,
|
||||
-0.014288832,
|
||||
0.023694387,
|
||||
-0.018556923,
|
||||
-0.003977263,
|
||||
-0.03251488,
|
||||
-0.04545843,
|
||||
-0.027434839,
|
||||
0.013158248,
|
||||
-0.005281848,
|
||||
-0.03187363,
|
||||
-0.022890532,
|
||||
-0.0063330783,
|
||||
0.040277284,
|
||||
0.017638152,
|
||||
-0.038472284,
|
||||
0.015346814,
|
||||
0.06673371,
|
||||
-0.011651253,
|
||||
-0.06683331,
|
||||
0.008377879,
|
||||
-0.030951817,
|
||||
-0.036013808,
|
||||
0.02394849,
|
||||
0.023321355,
|
||||
0.024521058,
|
||||
-0.03078664,
|
||||
0.014595395,
|
||||
-0.037766363,
|
||||
0.075227626,
|
||||
-0.01933975,
|
||||
0.043791853,
|
||||
-0.025162384,
|
||||
-0.044860955,
|
||||
0.0059519857,
|
||||
0.04085485,
|
||||
0.06551164,
|
||||
-0.05282273,
|
||||
0.0030225238,
|
||||
-0.06850771,
|
||||
-0.062015526,
|
||||
-0.06011264,
|
||||
0.014174797,
|
||||
-0.050894123,
|
||||
0.017077608,
|
||||
0.021088008,
|
||||
0.058029104,
|
||||
0.043224387,
|
||||
-0.004394573,
|
||||
-0.0022478225,
|
||||
-0.006972821,
|
||||
0.02401093,
|
||||
0.022611097,
|
||||
8.550083e-05,
|
||||
0.056450296,
|
||||
0.055112243,
|
||||
-0.034522895,
|
||||
0.06482398,
|
||||
0.08114595,
|
||||
0.022528961,
|
||||
-0.013464262,
|
||||
-0.0029874062,
|
||||
0.005515398,
|
||||
0.026176685,
|
||||
-0.041392956,
|
||||
-0.035894908,
|
||||
-0.052102275,
|
||||
0.032556653,
|
||||
-0.016931413,
|
||||
-0.047386043,
|
||||
0.012574915,
|
||||
0.03802867,
|
||||
0.045309085,
|
||||
0.025728,
|
||||
-0.02505067,
|
||||
0.039530423,
|
||||
-0.065004446,
|
||||
0.017083768,
|
||||
0.0033854055,
|
||||
0.07688453,
|
||||
-0.019878633,
|
||||
-0.0025184979,
|
||||
-0.0027949202,
|
||||
0.052868426,
|
||||
0.054179598,
|
||||
-0.0040608337,
|
||||
-0.0053128796,
|
||||
-0.04103081,
|
||||
-0.049691968,
|
||||
0.06014039,
|
||||
0.04751648,
|
||||
0.015087763,
|
||||
-0.06859484,
|
||||
0.00805693,
|
||||
-0.061754886,
|
||||
0.008819008,
|
||||
-0.027785089,
|
||||
-0.010586925,
|
||||
-0.020496469,
|
||||
-0.029158294,
|
||||
-0.05417346,
|
||||
-0.029509347,
|
||||
-0.025456924,
|
||||
0.041056376,
|
||||
0.0075264946,
|
||||
-0.018885529,
|
||||
0.07735419,
|
||||
0.00489195,
|
||||
0.050696895,
|
||||
-0.0041886116,
|
||||
0.064080104,
|
||||
-0.020775754,
|
||||
-0.017177466,
|
||||
0.0023288913,
|
||||
0.010398866,
|
||||
-0.040627487,
|
||||
-0.034321204,
|
||||
0.016019996,
|
||||
0.028118521,
|
||||
0.014172112,
|
||||
0.08738979,
|
||||
-0.03657629,
|
||||
0.018347794,
|
||||
-0.03947416,
|
||||
-0.01077611,
|
||||
0.00085160177,
|
||||
0.0368259,
|
||||
0.05611389,
|
||||
0.05134766,
|
||||
-0.025541335,
|
||||
-0.0057555106,
|
||||
-0.013793745,
|
||||
-0.05975066,
|
||||
-0.0064870752,
|
||||
-0.053716175,
|
||||
0.024583345,
|
||||
-0.084030546,
|
||||
-0.048775505,
|
||||
-0.059886374,
|
||||
-0.057641674,
|
||||
0.030222055,
|
||||
0.018706435,
|
||||
0.023170326,
|
||||
-0.0064046904,
|
||||
-0.018711446,
|
||||
-0.0029956547,
|
||||
0.0377868,
|
||||
-0.012569718,
|
||||
0.0514249,
|
||||
-0.012415474,
|
||||
-0.018657023,
|
||||
-0.040379863,
|
||||
0.029388199,
|
||||
-0.07378978,
|
||||
0.026212148,
|
||||
0.0056296797,
|
||||
-0.00405927,
|
||||
0.021354636,
|
||||
-0.0822599,
|
||||
0.01597725,
|
||||
0.07648158,
|
||||
-0.006006045,
|
||||
-0.014829594,
|
||||
-0.021541826,
|
||||
0.0032610476,
|
||||
0.06906917,
|
||||
-0.05802312,
|
||||
-0.023113884,
|
||||
-0.015534724,
|
||||
0.016758824,
|
||||
0.0030736707,
|
||||
-0.0022294512,
|
||||
-0.026804008,
|
||||
-0.0031566115,
|
||||
-0.0584943
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "nomic-embed-text:latest",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 6,
|
||||
"total_tokens": 6
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is the content of test file 1"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.029406669,
|
||||
0.08920982,
|
||||
-0.11326726,
|
||||
0.0065823817,
|
||||
0.07725067,
|
||||
-0.036890104,
|
||||
0.030436223,
|
||||
0.041454185,
|
||||
-0.049156666,
|
||||
0.018258564,
|
||||
0.14662577,
|
||||
0.01744915,
|
||||
-0.012837422,
|
||||
-0.06889876,
|
||||
-0.039401636,
|
||||
-0.038800705,
|
||||
-0.08963421,
|
||||
-0.059656583,
|
||||
0.001375945,
|
||||
0.045138627,
|
||||
0.042796962,
|
||||
0.053700265,
|
||||
-0.035706885,
|
||||
0.010138017,
|
||||
0.060920056,
|
||||
0.017344126,
|
||||
-0.05633907,
|
||||
0.063370295,
|
||||
0.0021257724,
|
||||
-0.083796844,
|
||||
0.050487563,
|
||||
0.047987595,
|
||||
0.069071226,
|
||||
0.049588464,
|
||||
0.117036626,
|
||||
0.05339311,
|
||||
0.10129953,
|
||||
-0.048230153,
|
||||
-0.014987975,
|
||||
0.0250915,
|
||||
0.031392053,
|
||||
-0.008863942,
|
||||
0.0073650074,
|
||||
-0.0009767569,
|
||||
-0.016403567,
|
||||
0.015523393,
|
||||
-0.010998956,
|
||||
-0.014870063,
|
||||
0.0061682137,
|
||||
-0.0017961137,
|
||||
-0.022682818,
|
||||
0.018210242,
|
||||
-0.07757007,
|
||||
-0.0015845516,
|
||||
0.069547005,
|
||||
0.000419109,
|
||||
0.038414054,
|
||||
0.005823485,
|
||||
-0.028931383,
|
||||
0.07009549,
|
||||
-0.0018009909,
|
||||
0.033516172,
|
||||
-0.014593847,
|
||||
0.03922457,
|
||||
0.08240545,
|
||||
-0.050596908,
|
||||
-0.039732855,
|
||||
-0.024425076,
|
||||
-0.015055329,
|
||||
-0.11705068,
|
||||
-0.15979129,
|
||||
-0.008256823,
|
||||
-0.0100719705,
|
||||
0.03266482,
|
||||
0.0029998205,
|
||||
0.0316428,
|
||||
-0.094554916,
|
||||
0.017661797,
|
||||
0.058996264,
|
||||
-0.119718134,
|
||||
-0.027414676,
|
||||
-0.09155906,
|
||||
0.040038,
|
||||
0.01091849,
|
||||
-0.029446004,
|
||||
0.10225186,
|
||||
0.06583262,
|
||||
-0.003439552,
|
||||
-0.009694834,
|
||||
0.016906522,
|
||||
0.023685955,
|
||||
-0.032616187,
|
||||
-0.010238839,
|
||||
0.07891618,
|
||||
-0.007330681,
|
||||
0.05238444,
|
||||
0.00943625,
|
||||
0.042121,
|
||||
0.08491511,
|
||||
0.049208272,
|
||||
-0.01868227,
|
||||
-0.013585418,
|
||||
0.06727199,
|
||||
0.084571496,
|
||||
-0.103213035,
|
||||
-0.08387524,
|
||||
0.03641842,
|
||||
-0.047227863,
|
||||
0.057315867,
|
||||
-0.04463932,
|
||||
0.006783099,
|
||||
-0.08934107,
|
||||
-0.015040418,
|
||||
-0.08107057,
|
||||
0.013285569,
|
||||
-0.060907867,
|
||||
-0.042128306,
|
||||
0.057306163,
|
||||
-0.058711898,
|
||||
0.04628304,
|
||||
0.070194095,
|
||||
-0.041729517,
|
||||
-0.0338408,
|
||||
-0.012369257,
|
||||
-0.044708908,
|
||||
-0.059450094,
|
||||
0.08251312,
|
||||
-3.443368e-33,
|
||||
0.0121309515,
|
||||
-0.11084454,
|
||||
-0.020510655,
|
||||
0.10916455,
|
||||
0.033683147,
|
||||
-0.02845083,
|
||||
0.024345158,
|
||||
0.034192592,
|
||||
-0.08367815,
|
||||
0.0064610844,
|
||||
-0.00912456,
|
||||
-0.0663567,
|
||||
-0.0028754657,
|
||||
0.008272698,
|
||||
-0.09166764,
|
||||
0.0089771375,
|
||||
-0.03963948,
|
||||
0.019947624,
|
||||
-0.01321528,
|
||||
-0.019034218,
|
||||
0.051933073,
|
||||
0.028107261,
|
||||
-0.039153125,
|
||||
-0.080395184,
|
||||
-0.050503474,
|
||||
0.02060341,
|
||||
-0.012718284,
|
||||
-0.046732575,
|
||||
0.017907938,
|
||||
-0.0028334607,
|
||||
-0.011695137,
|
||||
-0.05667005,
|
||||
-0.043894444,
|
||||
0.034919597,
|
||||
0.022352098,
|
||||
0.046777196,
|
||||
0.045085873,
|
||||
-0.008840106,
|
||||
-0.06373453,
|
||||
0.036720857,
|
||||
0.012829601,
|
||||
-0.035169926,
|
||||
0.046209145,
|
||||
-0.014361767,
|
||||
0.03706697,
|
||||
-0.056797564,
|
||||
-0.06310496,
|
||||
0.010818958,
|
||||
0.047810175,
|
||||
0.0029118094,
|
||||
-0.003235893,
|
||||
0.061511047,
|
||||
0.072056666,
|
||||
-0.03286638,
|
||||
0.005070082,
|
||||
0.021947902,
|
||||
-0.017779002,
|
||||
-0.022738373,
|
||||
-0.021926457,
|
||||
0.047074158,
|
||||
0.010847615,
|
||||
0.05539702,
|
||||
-0.07119971,
|
||||
0.033833236,
|
||||
0.012342855,
|
||||
-0.047586687,
|
||||
-0.026776271,
|
||||
-0.09885727,
|
||||
0.10053448,
|
||||
0.036877092,
|
||||
-0.07049897,
|
||||
-0.059692938,
|
||||
0.016129492,
|
||||
-0.0016443401,
|
||||
-0.026804024,
|
||||
-0.013527272,
|
||||
-0.015385511,
|
||||
0.055627547,
|
||||
-0.060485132,
|
||||
-0.055540122,
|
||||
-0.04329072,
|
||||
-0.07097361,
|
||||
-0.04857043,
|
||||
-0.03726256,
|
||||
-0.09059366,
|
||||
-0.036855534,
|
||||
0.024561211,
|
||||
-0.10113953,
|
||||
0.056738112,
|
||||
-0.10995085,
|
||||
0.042282794,
|
||||
0.014222368,
|
||||
-0.07067843,
|
||||
-0.05902307,
|
||||
0.06426122,
|
||||
1.6036318e-33,
|
||||
0.037851896,
|
||||
0.032911286,
|
||||
-0.04029648,
|
||||
-0.00049357174,
|
||||
0.028011942,
|
||||
0.048672136,
|
||||
0.07279598,
|
||||
-0.027471887,
|
||||
-0.02847654,
|
||||
0.114492,
|
||||
0.001777095,
|
||||
-0.009519909,
|
||||
0.0025862327,
|
||||
-0.056408145,
|
||||
0.023462169,
|
||||
-0.006209674,
|
||||
-0.010567065,
|
||||
-0.05877587,
|
||||
-0.032393616,
|
||||
0.011836781,
|
||||
-0.038905054,
|
||||
0.05516299,
|
||||
0.09564333,
|
||||
0.028543225,
|
||||
-0.023832332,
|
||||
-0.0015711841,
|
||||
0.047049087,
|
||||
0.03128219,
|
||||
0.02811091,
|
||||
0.007177092,
|
||||
0.055283513,
|
||||
0.06574452,
|
||||
-0.1020208,
|
||||
0.021213628,
|
||||
0.020237882,
|
||||
-0.10449357,
|
||||
0.09608935,
|
||||
-0.06253181,
|
||||
0.015293753,
|
||||
0.042053986,
|
||||
0.06105009,
|
||||
0.0909162,
|
||||
0.018404186,
|
||||
0.031023262,
|
||||
0.03562763,
|
||||
0.112073965,
|
||||
0.10124763,
|
||||
-0.007683015,
|
||||
0.013140281,
|
||||
-0.042280227,
|
||||
0.051135287,
|
||||
-0.02950743,
|
||||
0.027794402,
|
||||
-0.010734668,
|
||||
-0.011067552,
|
||||
0.058104575,
|
||||
-0.009284788,
|
||||
0.056184508,
|
||||
-0.040822964,
|
||||
0.010282754,
|
||||
0.0374409,
|
||||
0.054198533,
|
||||
-0.061418086,
|
||||
0.030569963,
|
||||
0.0023648597,
|
||||
-0.054184474,
|
||||
-0.020570045,
|
||||
0.012422129,
|
||||
0.025696559,
|
||||
-0.007607385,
|
||||
-0.026194826,
|
||||
-0.024159024,
|
||||
0.0012979766,
|
||||
-0.07461716,
|
||||
0.051458035,
|
||||
-0.004183808,
|
||||
-0.040804464,
|
||||
-0.023975441,
|
||||
0.009455526,
|
||||
-0.0018798193,
|
||||
0.03668693,
|
||||
-0.019319497,
|
||||
-0.06195781,
|
||||
0.06456675,
|
||||
0.040328216,
|
||||
-0.010790134,
|
||||
0.013190221,
|
||||
0.09067539,
|
||||
-0.0051480443,
|
||||
0.013312647,
|
||||
-0.029548675,
|
||||
0.07769003,
|
||||
0.0027328292,
|
||||
0.04533781,
|
||||
-0.0017606319,
|
||||
-1.661594e-08,
|
||||
-0.040610366,
|
||||
-0.09883059,
|
||||
-0.05522113,
|
||||
-0.02916469,
|
||||
-0.019305382,
|
||||
0.088138185,
|
||||
-0.038325552,
|
||||
-0.03327639,
|
||||
-0.012629364,
|
||||
0.006948921,
|
||||
0.010438818,
|
||||
0.026771523,
|
||||
-0.040855426,
|
||||
-0.03958403,
|
||||
-0.051137064,
|
||||
-0.016159322,
|
||||
-0.020525131,
|
||||
-0.023726366,
|
||||
-0.013322245,
|
||||
-0.008097836,
|
||||
0.028000915,
|
||||
0.02806969,
|
||||
0.015645925,
|
||||
-0.0043166955,
|
||||
0.0054488196,
|
||||
0.06720413,
|
||||
0.068473674,
|
||||
0.07172716,
|
||||
-0.06339439,
|
||||
-0.02540609,
|
||||
0.08468492,
|
||||
0.041936778,
|
||||
0.021067144,
|
||||
-0.07596481,
|
||||
0.017143335,
|
||||
0.1260291,
|
||||
0.121315174,
|
||||
0.08431059,
|
||||
0.040587336,
|
||||
0.036687315,
|
||||
-0.04717,
|
||||
-0.022659328,
|
||||
-0.006820436,
|
||||
0.005210712,
|
||||
-0.033785924,
|
||||
-0.08449115,
|
||||
-0.0844501,
|
||||
-0.03192747,
|
||||
-0.036649443,
|
||||
-0.13791409,
|
||||
-0.036417518,
|
||||
-0.00080547476,
|
||||
-0.047578912,
|
||||
0.038795993,
|
||||
-0.06757743,
|
||||
0.016941966,
|
||||
0.036312684,
|
||||
0.0125779435,
|
||||
-0.058240637,
|
||||
0.004471269,
|
||||
0.03226526,
|
||||
0.09821741,
|
||||
0.053010236,
|
||||
-0.016268
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 8,
|
||||
"total_tokens": 8
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is batch test file 1"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.009745733,
|
||||
0.03363038,
|
||||
-0.10852256,
|
||||
0.026609829,
|
||||
-0.0060599064,
|
||||
-0.020473678,
|
||||
0.0692486,
|
||||
0.032276765,
|
||||
-0.11532835,
|
||||
-0.0005207133,
|
||||
0.11814916,
|
||||
0.0119809555,
|
||||
0.03685765,
|
||||
-0.10744223,
|
||||
-0.046515625,
|
||||
0.0015449532,
|
||||
-0.06319664,
|
||||
-0.04640812,
|
||||
-0.037318822,
|
||||
-0.025718328,
|
||||
-0.00026058854,
|
||||
-0.011890766,
|
||||
-0.050925612,
|
||||
0.014111713,
|
||||
0.029467698,
|
||||
0.006379121,
|
||||
-0.012013293,
|
||||
-0.0024293982,
|
||||
-0.044318773,
|
||||
-0.08100101,
|
||||
0.02009568,
|
||||
0.055713937,
|
||||
0.078816675,
|
||||
0.054973654,
|
||||
0.20367871,
|
||||
-0.004309458,
|
||||
0.03877001,
|
||||
0.03825522,
|
||||
-0.002538199,
|
||||
-0.0007973801,
|
||||
0.044761047,
|
||||
-0.054529082,
|
||||
-0.008856888,
|
||||
-0.04078078,
|
||||
0.011367262,
|
||||
-0.022404457,
|
||||
-0.06209053,
|
||||
0.02558725,
|
||||
-0.0034454092,
|
||||
-0.03743928,
|
||||
-0.062026348,
|
||||
-0.030812219,
|
||||
-0.034592565,
|
||||
-0.014926672,
|
||||
0.018588377,
|
||||
0.013435887,
|
||||
0.08169151,
|
||||
0.053658403,
|
||||
-0.03557856,
|
||||
0.033325985,
|
||||
-0.01637577,
|
||||
-0.0222152,
|
||||
-0.039247517,
|
||||
0.00094368146,
|
||||
0.10228945,
|
||||
-0.04305617,
|
||||
-0.052200828,
|
||||
-0.02007385,
|
||||
0.054805383,
|
||||
-0.08231377,
|
||||
-0.14736547,
|
||||
0.048954617,
|
||||
-0.0212168,
|
||||
0.02872658,
|
||||
-0.0671409,
|
||||
0.021436114,
|
||||
-0.023599947,
|
||||
0.03677982,
|
||||
0.010577411,
|
||||
-0.0966004,
|
||||
-0.06367233,
|
||||
-0.10277648,
|
||||
0.0273993,
|
||||
-0.06292906,
|
||||
-0.046344172,
|
||||
0.039919835,
|
||||
0.02682899,
|
||||
0.025460077,
|
||||
-0.013083559,
|
||||
-0.002667712,
|
||||
-0.016529463,
|
||||
0.012605053,
|
||||
-0.0064383023,
|
||||
0.015841383,
|
||||
-0.01710707,
|
||||
0.12320292,
|
||||
-0.0077660284,
|
||||
0.05845043,
|
||||
0.07362552,
|
||||
0.038426086,
|
||||
0.004742023,
|
||||
-0.0155985365,
|
||||
0.01418979,
|
||||
0.07865995,
|
||||
-0.026352523,
|
||||
-0.037174653,
|
||||
0.06787817,
|
||||
-0.060126718,
|
||||
0.06111402,
|
||||
-0.034931272,
|
||||
-0.009446326,
|
||||
-0.006150886,
|
||||
0.02892313,
|
||||
-0.09361577,
|
||||
0.0335364,
|
||||
-0.09088912,
|
||||
0.009241144,
|
||||
0.07092964,
|
||||
-0.08954648,
|
||||
0.04494549,
|
||||
0.040462427,
|
||||
-0.04167353,
|
||||
0.0076030386,
|
||||
-0.0066417656,
|
||||
-0.07275736,
|
||||
-0.043690544,
|
||||
0.07685007,
|
||||
-1.0508795e-33,
|
||||
-0.019583685,
|
||||
-0.13087204,
|
||||
-0.03574564,
|
||||
0.070223756,
|
||||
0.08133056,
|
||||
-0.009436003,
|
||||
0.046778366,
|
||||
0.03478148,
|
||||
-0.09441185,
|
||||
-0.040857755,
|
||||
-0.02127058,
|
||||
-0.106959894,
|
||||
0.024023255,
|
||||
0.022780996,
|
||||
-0.09042505,
|
||||
-0.035755932,
|
||||
0.011359196,
|
||||
0.050059184,
|
||||
0.0050815986,
|
||||
-0.07676938,
|
||||
0.05453651,
|
||||
0.04191775,
|
||||
-0.009206564,
|
||||
-0.022437057,
|
||||
-0.04617258,
|
||||
-0.038608693,
|
||||
-0.00036489012,
|
||||
-0.025092375,
|
||||
0.039146807,
|
||||
-0.0072839926,
|
||||
0.03675482,
|
||||
-0.011301064,
|
||||
-0.08863303,
|
||||
0.059421506,
|
||||
0.015851071,
|
||||
0.033407707,
|
||||
0.056883834,
|
||||
-0.01203776,
|
||||
0.027333334,
|
||||
-0.009560535,
|
||||
-0.05030555,
|
||||
-0.009787559,
|
||||
0.023205005,
|
||||
-0.007937716,
|
||||
0.003991047,
|
||||
-0.036422852,
|
||||
-0.06979188,
|
||||
0.046075627,
|
||||
0.056377746,
|
||||
0.0071927872,
|
||||
-0.00020658698,
|
||||
0.017678235,
|
||||
0.023745935,
|
||||
-0.0031295705,
|
||||
0.016370842,
|
||||
0.027585855,
|
||||
-0.03440131,
|
||||
-0.05594279,
|
||||
0.036442764,
|
||||
0.03577988,
|
||||
-0.005324585,
|
||||
0.015240975,
|
||||
-0.09071462,
|
||||
0.072764605,
|
||||
0.02343818,
|
||||
-0.093097225,
|
||||
0.05842133,
|
||||
-0.061913762,
|
||||
0.045556016,
|
||||
0.07639311,
|
||||
-0.035199754,
|
||||
-0.009256856,
|
||||
0.038682748,
|
||||
-0.040795818,
|
||||
0.017686425,
|
||||
-0.025513103,
|
||||
0.06860537,
|
||||
0.085520275,
|
||||
-0.1023457,
|
||||
-0.0036474275,
|
||||
-0.014826131,
|
||||
-0.05045756,
|
||||
-0.09065474,
|
||||
-0.076476775,
|
||||
-0.008538021,
|
||||
-0.04111943,
|
||||
-0.035473913,
|
||||
-0.061549038,
|
||||
0.114327826,
|
||||
-0.09601482,
|
||||
0.022990143,
|
||||
0.0022396755,
|
||||
-0.023026146,
|
||||
-0.028128328,
|
||||
0.07969127,
|
||||
-4.1765383e-34,
|
||||
0.07866384,
|
||||
0.11484068,
|
||||
0.016687382,
|
||||
0.009315677,
|
||||
0.01664128,
|
||||
0.024303248,
|
||||
0.046507504,
|
||||
-0.043804675,
|
||||
-0.09136995,
|
||||
0.106353745,
|
||||
-0.06948852,
|
||||
0.018747667,
|
||||
0.0053492193,
|
||||
-0.033229355,
|
||||
0.042339083,
|
||||
-0.0017468681,
|
||||
0.05323157,
|
||||
0.0058223205,
|
||||
-0.05331342,
|
||||
0.016506517,
|
||||
-0.02325185,
|
||||
0.097519755,
|
||||
-0.0045558517,
|
||||
0.08866843,
|
||||
-0.028221445,
|
||||
-0.012007969,
|
||||
-0.009742725,
|
||||
0.061458003,
|
||||
0.01574456,
|
||||
-0.00039456616,
|
||||
0.02444834,
|
||||
0.065891184,
|
||||
-0.054779086,
|
||||
0.04863689,
|
||||
0.043890025,
|
||||
-0.062467597,
|
||||
0.07615393,
|
||||
0.0067509366,
|
||||
0.019150084,
|
||||
0.06994535,
|
||||
0.027900916,
|
||||
0.08902746,
|
||||
-0.027433047,
|
||||
0.031390887,
|
||||
0.02271287,
|
||||
0.08119532,
|
||||
0.06855678,
|
||||
0.0023552915,
|
||||
-0.06764184,
|
||||
0.00704173,
|
||||
-0.034521427,
|
||||
-0.053785548,
|
||||
-0.03075216,
|
||||
0.007947864,
|
||||
-0.025317406,
|
||||
-0.040664013,
|
||||
0.036144093,
|
||||
0.017730465,
|
||||
-0.040179063,
|
||||
0.013665757,
|
||||
0.004815376,
|
||||
0.009095556,
|
||||
0.0072483593,
|
||||
0.012753351,
|
||||
-0.047865536,
|
||||
-0.046072423,
|
||||
-0.014048283,
|
||||
0.031082962,
|
||||
-0.034945205,
|
||||
-0.023550391,
|
||||
0.033062257,
|
||||
-0.022966444,
|
||||
0.007744228,
|
||||
0.015939556,
|
||||
-0.0012224894,
|
||||
0.0010534802,
|
||||
-0.015109,
|
||||
-0.021597888,
|
||||
-0.029862719,
|
||||
0.03983828,
|
||||
0.062536344,
|
||||
0.0106168175,
|
||||
-0.027220478,
|
||||
0.02410377,
|
||||
-0.0023566757,
|
||||
0.085310005,
|
||||
0.04843323,
|
||||
0.090823516,
|
||||
0.005126319,
|
||||
0.020297319,
|
||||
-0.01739127,
|
||||
0.047677357,
|
||||
0.11080086,
|
||||
0.030030197,
|
||||
0.029773563,
|
||||
-1.5454503e-08,
|
||||
-0.03580758,
|
||||
-0.12177604,
|
||||
0.019753791,
|
||||
0.05854353,
|
||||
-0.01590761,
|
||||
0.085781366,
|
||||
-0.09558486,
|
||||
-0.0016744126,
|
||||
0.00773199,
|
||||
-0.04790156,
|
||||
0.01175936,
|
||||
0.006536077,
|
||||
-0.032027386,
|
||||
0.0031026274,
|
||||
-0.07580574,
|
||||
-0.039700802,
|
||||
-0.00170645,
|
||||
-0.070955865,
|
||||
0.043680355,
|
||||
0.029966798,
|
||||
0.0039943648,
|
||||
0.031923376,
|
||||
0.08119928,
|
||||
0.038820695,
|
||||
0.013302812,
|
||||
0.041675337,
|
||||
0.044349737,
|
||||
0.060403902,
|
||||
-0.1058191,
|
||||
-0.05287386,
|
||||
0.050275758,
|
||||
0.039101604,
|
||||
0.0599918,
|
||||
-0.025067834,
|
||||
-0.019554066,
|
||||
0.06748813,
|
||||
0.12508559,
|
||||
0.059007537,
|
||||
-0.019899847,
|
||||
-0.030194808,
|
||||
-0.046559453,
|
||||
0.034567222,
|
||||
-0.021644907,
|
||||
-0.03327634,
|
||||
-0.0075667608,
|
||||
-0.100658834,
|
||||
-0.0639619,
|
||||
-0.055270903,
|
||||
-0.0111757815,
|
||||
-0.11671873,
|
||||
-0.07208087,
|
||||
0.023208033,
|
||||
0.027215267,
|
||||
0.063635156,
|
||||
-0.05858023,
|
||||
0.020345282,
|
||||
0.018325811,
|
||||
-0.0036095325,
|
||||
0.006916675,
|
||||
0.06541716,
|
||||
0.009575581,
|
||||
0.046839867,
|
||||
0.0070611075,
|
||||
-0.09470841
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 6,
|
||||
"total_tokens": 6
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,807 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "nomic-embed-text:latest",
|
||||
"input": [
|
||||
"This is the content of test file 1"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "nomic-embed-text:latest"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
0.011639302,
|
||||
0.015066345,
|
||||
-0.1572681,
|
||||
-0.044965014,
|
||||
0.045302268,
|
||||
0.012036585,
|
||||
0.036542512,
|
||||
0.005725059,
|
||||
-0.052068613,
|
||||
-0.023581833,
|
||||
-0.0041714185,
|
||||
0.047297083,
|
||||
0.0044818125,
|
||||
0.0073796143,
|
||||
-0.06833552,
|
||||
-0.020871542,
|
||||
0.030256433,
|
||||
-0.052156504,
|
||||
-0.016426736,
|
||||
-0.045092978,
|
||||
-0.012395779,
|
||||
-0.017792987,
|
||||
-0.08013035,
|
||||
-0.025271736,
|
||||
0.110538565,
|
||||
0.019197728,
|
||||
-0.05617586,
|
||||
0.045342237,
|
||||
-0.100888394,
|
||||
-0.015532925,
|
||||
0.04541118,
|
||||
-0.048470836,
|
||||
0.014533936,
|
||||
-0.04054472,
|
||||
0.004343861,
|
||||
-0.03328387,
|
||||
0.038874496,
|
||||
0.034725353,
|
||||
0.022646122,
|
||||
0.024648184,
|
||||
-0.02911675,
|
||||
-0.0140342,
|
||||
-0.010215055,
|
||||
-0.00092316914,
|
||||
0.015458233,
|
||||
0.0042022346,
|
||||
-0.0118241655,
|
||||
0.031950384,
|
||||
0.021818206,
|
||||
-0.0009401939,
|
||||
0.0028767833,
|
||||
0.022676043,
|
||||
-0.027578428,
|
||||
0.017072845,
|
||||
0.055942602,
|
||||
0.008372957,
|
||||
-8.234923e-05,
|
||||
-0.0076934453,
|
||||
0.017103186,
|
||||
-0.07049976,
|
||||
0.0470288,
|
||||
0.030520689,
|
||||
-0.0853248,
|
||||
0.031003723,
|
||||
0.005461848,
|
||||
-0.039933346,
|
||||
-0.028195756,
|
||||
0.02583814,
|
||||
-0.020038705,
|
||||
0.027421296,
|
||||
0.09211795,
|
||||
-0.0021492639,
|
||||
0.009114191,
|
||||
-0.02858135,
|
||||
-0.0130490055,
|
||||
-0.047928475,
|
||||
0.021908045,
|
||||
-0.03816779,
|
||||
-0.040175024,
|
||||
-0.008988226,
|
||||
0.062123742,
|
||||
0.032331105,
|
||||
0.013500545,
|
||||
0.014699184,
|
||||
0.06949165,
|
||||
-0.029347114,
|
||||
-0.036963575,
|
||||
-0.025804758,
|
||||
0.006973289,
|
||||
0.03219541,
|
||||
0.014725156,
|
||||
0.032485504,
|
||||
0.025228832,
|
||||
0.015978005,
|
||||
-0.0036463195,
|
||||
0.009395636,
|
||||
-0.0030804265,
|
||||
0.026493264,
|
||||
-0.01026861,
|
||||
0.0003747859,
|
||||
0.017421532,
|
||||
0.015864568,
|
||||
0.0254427,
|
||||
-0.021840125,
|
||||
0.006622214,
|
||||
0.018735437,
|
||||
-0.008662971,
|
||||
8.567802e-05,
|
||||
-0.08026379,
|
||||
-0.003987451,
|
||||
-0.0022324976,
|
||||
0.030920457,
|
||||
-0.03272228,
|
||||
-0.025135856,
|
||||
0.015818166,
|
||||
-0.030539474,
|
||||
0.084593095,
|
||||
-0.0072662015,
|
||||
-0.04306349,
|
||||
0.016708883,
|
||||
-0.02148629,
|
||||
-0.021512741,
|
||||
0.011571002,
|
||||
0.07055689,
|
||||
0.016873637,
|
||||
0.017103009,
|
||||
-0.058425475,
|
||||
0.009254332,
|
||||
0.018121762,
|
||||
-0.029209172,
|
||||
-0.017481297,
|
||||
0.005129311,
|
||||
-1.4240719e-05,
|
||||
-0.02815651,
|
||||
-0.012156167,
|
||||
-0.011126637,
|
||||
0.012530035,
|
||||
0.017916953,
|
||||
-0.049299978,
|
||||
-0.023406321,
|
||||
-0.025908781,
|
||||
0.01868743,
|
||||
0.013128073,
|
||||
0.030577261,
|
||||
0.03492911,
|
||||
-0.026720364,
|
||||
0.044888426,
|
||||
0.015100583,
|
||||
-0.02517811,
|
||||
-0.018026695,
|
||||
0.04455666,
|
||||
0.006026339,
|
||||
0.006132853,
|
||||
0.013067113,
|
||||
0.013630368,
|
||||
-0.06992026,
|
||||
0.015714098,
|
||||
-0.025995858,
|
||||
0.00764349,
|
||||
0.06502328,
|
||||
0.00921131,
|
||||
0.00039559926,
|
||||
-0.014907944,
|
||||
0.033250615,
|
||||
-0.022297438,
|
||||
-0.022631606,
|
||||
-0.009259513,
|
||||
0.07459313,
|
||||
0.07961807,
|
||||
-0.04546293,
|
||||
-0.04984229,
|
||||
-0.056986727,
|
||||
-0.021624641,
|
||||
-0.009604434,
|
||||
-0.050308105,
|
||||
-0.029882062,
|
||||
0.02890167,
|
||||
-0.016925206,
|
||||
0.03357617,
|
||||
-0.013084858,
|
||||
0.032628123,
|
||||
0.004407517,
|
||||
0.028667213,
|
||||
0.02581998,
|
||||
-0.008354794,
|
||||
-0.045190092,
|
||||
0.017996402,
|
||||
-0.021489577,
|
||||
-0.049504388,
|
||||
0.003702337,
|
||||
-0.023653682,
|
||||
-0.007418799,
|
||||
-0.09230719,
|
||||
-0.0666417,
|
||||
0.01368294,
|
||||
-0.07626095,
|
||||
0.016283033,
|
||||
-0.025274273,
|
||||
0.046640623,
|
||||
-0.03306251,
|
||||
-0.019547738,
|
||||
-0.02728644,
|
||||
-0.038707435,
|
||||
-0.0075380616,
|
||||
-0.03706684,
|
||||
-0.014038333,
|
||||
-0.019394161,
|
||||
-0.010599687,
|
||||
-0.0057279305,
|
||||
0.01753179,
|
||||
0.037147418,
|
||||
-0.01409748,
|
||||
-0.028078519,
|
||||
0.028943742,
|
||||
0.044126343,
|
||||
-0.024797611,
|
||||
-0.02061766,
|
||||
-0.041145287,
|
||||
-0.0016994113,
|
||||
-0.035794605,
|
||||
0.022767134,
|
||||
0.007715133,
|
||||
0.033083446,
|
||||
-0.06898011,
|
||||
0.0077070463,
|
||||
0.0039297407,
|
||||
-0.038172692,
|
||||
-0.032068398,
|
||||
-0.043679804,
|
||||
-0.0292851,
|
||||
-0.020715753,
|
||||
-0.05462352,
|
||||
0.011206036,
|
||||
0.020920858,
|
||||
-0.007133438,
|
||||
-0.006820509,
|
||||
-0.016809242,
|
||||
0.06488191,
|
||||
-0.0150427865,
|
||||
0.040075593,
|
||||
-0.047243405,
|
||||
0.05071197,
|
||||
0.015879754,
|
||||
-0.0006090825,
|
||||
0.0067252424,
|
||||
0.0052318904,
|
||||
0.0038148144,
|
||||
-0.032034587,
|
||||
0.032176103,
|
||||
0.040441014,
|
||||
-0.03223476,
|
||||
0.0034279015,
|
||||
0.04811163,
|
||||
0.058563426,
|
||||
0.025335358,
|
||||
-0.03077014,
|
||||
-0.0060142917,
|
||||
-0.025248509,
|
||||
0.024592392,
|
||||
-0.03674746,
|
||||
0.024663158,
|
||||
-0.060253005,
|
||||
0.009173809,
|
||||
-0.004111937,
|
||||
-0.063402615,
|
||||
-0.01951628,
|
||||
-0.039490156,
|
||||
0.018258424,
|
||||
-0.043015976,
|
||||
-0.015164487,
|
||||
-0.017073318,
|
||||
0.027809769,
|
||||
-0.021215433,
|
||||
0.007797112,
|
||||
0.008731678,
|
||||
-0.036673818,
|
||||
0.012786695,
|
||||
0.028968208,
|
||||
-0.030241087,
|
||||
0.020865943,
|
||||
-0.026749771,
|
||||
0.033981804,
|
||||
0.010454427,
|
||||
0.023153242,
|
||||
0.020885227,
|
||||
-0.0056243115,
|
||||
0.0117305005,
|
||||
-0.015051302,
|
||||
0.013582618,
|
||||
0.03807434,
|
||||
0.010856497,
|
||||
0.020801183,
|
||||
0.011158894,
|
||||
0.036391996,
|
||||
0.019670399,
|
||||
0.007724792,
|
||||
0.06660602,
|
||||
-0.011434749,
|
||||
0.0057949307,
|
||||
-0.015963648,
|
||||
-0.019779123,
|
||||
0.005820883,
|
||||
0.02833991,
|
||||
0.055220414,
|
||||
0.010273399,
|
||||
-0.016092837,
|
||||
0.03503124,
|
||||
-0.034432467,
|
||||
0.023686841,
|
||||
0.022379564,
|
||||
-0.07128316,
|
||||
0.012263694,
|
||||
-0.015228141,
|
||||
0.0032988787,
|
||||
-0.029930541,
|
||||
0.041881878,
|
||||
0.03506383,
|
||||
0.020228907,
|
||||
0.0438159,
|
||||
-0.038998622,
|
||||
0.0033828963,
|
||||
-0.082220346,
|
||||
-0.021915225,
|
||||
-0.00014996591,
|
||||
0.02804432,
|
||||
0.020062406,
|
||||
0.012756022,
|
||||
0.034497134,
|
||||
-0.02747778,
|
||||
-0.047376838,
|
||||
0.064383976,
|
||||
0.070425786,
|
||||
-0.05746651,
|
||||
-0.028404344,
|
||||
0.026372714,
|
||||
0.03306257,
|
||||
0.0073155067,
|
||||
0.051485326,
|
||||
0.0068675145,
|
||||
0.040136788,
|
||||
0.045383066,
|
||||
0.034149066,
|
||||
0.02086147,
|
||||
0.0009087964,
|
||||
0.037278313,
|
||||
-0.081617154,
|
||||
-0.032882202,
|
||||
0.02157909,
|
||||
0.021868218,
|
||||
0.07965252,
|
||||
-0.0027324036,
|
||||
-0.0022803254,
|
||||
0.014258049,
|
||||
-0.0020600832,
|
||||
0.00047349077,
|
||||
0.04002713,
|
||||
0.04263055,
|
||||
-0.009511693,
|
||||
0.06796055,
|
||||
-0.02155429,
|
||||
0.043834608,
|
||||
-0.029989557,
|
||||
0.009623121,
|
||||
-0.026068889,
|
||||
0.021337777,
|
||||
0.011070724,
|
||||
-0.020380916,
|
||||
-0.0023191955,
|
||||
0.046481982,
|
||||
0.039304417,
|
||||
-0.0045394786,
|
||||
0.003737432,
|
||||
0.034863517,
|
||||
0.053514365,
|
||||
0.035962798,
|
||||
0.04095995,
|
||||
-0.080873586,
|
||||
0.0112584885,
|
||||
-0.0145209655,
|
||||
0.023800805,
|
||||
0.04855744,
|
||||
0.0037306463,
|
||||
0.03949077,
|
||||
0.042007603,
|
||||
0.00916003,
|
||||
-0.012223143,
|
||||
0.022103913,
|
||||
-0.017077385,
|
||||
0.035043065,
|
||||
0.0052557834,
|
||||
-0.039841656,
|
||||
0.0020140728,
|
||||
-0.057917137,
|
||||
0.03641347,
|
||||
0.017727314,
|
||||
-0.030229636,
|
||||
0.026509946,
|
||||
0.010324972,
|
||||
-0.028184937,
|
||||
0.017539727,
|
||||
-0.021746434,
|
||||
0.0031611102,
|
||||
-0.008564719,
|
||||
0.026577024,
|
||||
-0.0073260553,
|
||||
0.012139988,
|
||||
-0.039608642,
|
||||
-0.062452354,
|
||||
0.03773313,
|
||||
0.002820345,
|
||||
0.017331626,
|
||||
-0.008981819,
|
||||
-0.02020533,
|
||||
-0.057272766,
|
||||
-0.014693149,
|
||||
0.033687364,
|
||||
0.038407385,
|
||||
-0.020838683,
|
||||
0.038617346,
|
||||
-0.03282725,
|
||||
0.0065172473,
|
||||
0.031010486,
|
||||
-0.0017651296,
|
||||
-0.02163586,
|
||||
-0.008899588,
|
||||
-0.026506478,
|
||||
0.03540833,
|
||||
0.07076032,
|
||||
-0.016357146,
|
||||
-0.08069671,
|
||||
-0.042310607,
|
||||
-0.012363274,
|
||||
0.03790111,
|
||||
0.007565661,
|
||||
-0.037524316,
|
||||
0.07095513,
|
||||
0.010869782,
|
||||
0.0032129285,
|
||||
-0.033399966,
|
||||
0.038155936,
|
||||
0.034415327,
|
||||
-0.052643284,
|
||||
-0.05567196,
|
||||
-0.03225739,
|
||||
0.008719539,
|
||||
0.14483878,
|
||||
0.071855366,
|
||||
-0.026637336,
|
||||
-0.04281552,
|
||||
-0.02133026,
|
||||
0.020932574,
|
||||
0.023442162,
|
||||
0.0018492922,
|
||||
0.03244938,
|
||||
0.08237317,
|
||||
-0.03321164,
|
||||
0.051374298,
|
||||
-0.018296566,
|
||||
-0.009659297,
|
||||
0.031976808,
|
||||
-0.010097727,
|
||||
0.010057915,
|
||||
0.051651575,
|
||||
0.0199425,
|
||||
0.019540219,
|
||||
-0.020617861,
|
||||
0.03563907,
|
||||
-0.036343392,
|
||||
0.032987807,
|
||||
0.06027452,
|
||||
-0.017668264,
|
||||
-0.044425867,
|
||||
0.015104213,
|
||||
-0.07373515,
|
||||
0.01810383,
|
||||
0.031706426,
|
||||
-0.046879865,
|
||||
0.0036537861,
|
||||
-0.031956047,
|
||||
0.03578955,
|
||||
0.027828328,
|
||||
0.021754785,
|
||||
-0.062319316,
|
||||
-0.035861533,
|
||||
0.023409521,
|
||||
-0.011718964,
|
||||
0.012511818,
|
||||
0.019975103,
|
||||
0.03046746,
|
||||
0.019306395,
|
||||
0.008897869,
|
||||
0.022976985,
|
||||
0.08666871,
|
||||
0.034413245,
|
||||
0.007698169,
|
||||
-0.013328631,
|
||||
0.026807705,
|
||||
-0.039164156,
|
||||
0.0001842902,
|
||||
0.008939378,
|
||||
0.053093646,
|
||||
0.0054843645,
|
||||
-0.0048546907,
|
||||
0.006646481,
|
||||
-0.036913976,
|
||||
-0.02434218,
|
||||
-0.007819763,
|
||||
-0.034326635,
|
||||
-0.09425071,
|
||||
-0.035864092,
|
||||
-0.027039077,
|
||||
0.0018631782,
|
||||
-0.011367168,
|
||||
0.03460308,
|
||||
0.06908907,
|
||||
0.0006993122,
|
||||
0.029187243,
|
||||
0.013981396,
|
||||
-0.034905925,
|
||||
0.009661519,
|
||||
0.016402403,
|
||||
0.013219478,
|
||||
0.025419146,
|
||||
-0.013838593,
|
||||
-0.09521828,
|
||||
0.04690183,
|
||||
0.008306249,
|
||||
-0.04494361,
|
||||
0.07675296,
|
||||
0.08630913,
|
||||
0.0027291386,
|
||||
0.047438163,
|
||||
-0.03291628,
|
||||
-0.017013406,
|
||||
0.008466675,
|
||||
0.0068329596,
|
||||
-0.047961134,
|
||||
-0.0060370415,
|
||||
0.017779041,
|
||||
0.05304337,
|
||||
-0.07138653,
|
||||
-0.013791788,
|
||||
0.01667366,
|
||||
-0.026808698,
|
||||
0.012813507,
|
||||
-0.029537767,
|
||||
-0.07048566,
|
||||
0.026801381,
|
||||
-0.021863695,
|
||||
-0.08986038,
|
||||
0.04256004,
|
||||
-0.042580713,
|
||||
-0.050321113,
|
||||
-0.02441381,
|
||||
0.024967946,
|
||||
-0.03307329,
|
||||
0.023765154,
|
||||
-0.042465124,
|
||||
-0.022590572,
|
||||
-0.050977908,
|
||||
0.02002681,
|
||||
-0.01659008,
|
||||
-0.0016500946,
|
||||
0.007923218,
|
||||
0.023085529,
|
||||
-0.028293792,
|
||||
-0.0070867077,
|
||||
-0.002519201,
|
||||
0.014844528,
|
||||
0.012927241,
|
||||
0.013701682,
|
||||
-0.048480112,
|
||||
0.017051037,
|
||||
-0.048594326,
|
||||
-0.03374255,
|
||||
0.015788445,
|
||||
0.01736624,
|
||||
0.02363127,
|
||||
-0.043622795,
|
||||
-0.04752542,
|
||||
0.05619384,
|
||||
-0.009064419,
|
||||
0.013587886,
|
||||
0.031963795,
|
||||
0.0055674682,
|
||||
0.00821165,
|
||||
-0.007879534,
|
||||
-0.025519967,
|
||||
0.030929072,
|
||||
-0.03054716,
|
||||
-0.028717758,
|
||||
-0.01304714,
|
||||
0.025171572,
|
||||
-0.004879199,
|
||||
-0.001190343,
|
||||
-0.010213315,
|
||||
0.01971419,
|
||||
-0.032143768,
|
||||
-0.008055433,
|
||||
-0.045028396,
|
||||
0.0050284,
|
||||
0.008977255,
|
||||
0.007132238,
|
||||
-0.052949388,
|
||||
0.011562612,
|
||||
-0.0043699676,
|
||||
0.06377099,
|
||||
-0.010715953,
|
||||
-0.027962748,
|
||||
0.0025381946,
|
||||
0.065418504,
|
||||
0.015951851,
|
||||
-0.10228855,
|
||||
-0.0038436814,
|
||||
-0.015558708,
|
||||
-0.035604823,
|
||||
0.039515387,
|
||||
-0.011977611,
|
||||
0.008272532,
|
||||
-0.047362626,
|
||||
0.029810345,
|
||||
-0.026100902,
|
||||
0.080183394,
|
||||
-0.029716058,
|
||||
-0.008065036,
|
||||
-0.0019149253,
|
||||
-0.029152166,
|
||||
0.030865246,
|
||||
0.028290713,
|
||||
0.059991617,
|
||||
-0.0539013,
|
||||
0.037941493,
|
||||
-0.046701435,
|
||||
-0.056897625,
|
||||
-0.050652288,
|
||||
0.0022519496,
|
||||
-0.044697277,
|
||||
0.018704673,
|
||||
0.024128519,
|
||||
0.06013336,
|
||||
0.057803143,
|
||||
-0.011098817,
|
||||
0.004350433,
|
||||
-0.046533823,
|
||||
0.011547173,
|
||||
0.039410993,
|
||||
0.010503389,
|
||||
0.058373533,
|
||||
0.04097013,
|
||||
-0.04243095,
|
||||
0.09540366,
|
||||
0.07546867,
|
||||
0.057422172,
|
||||
-0.0150666535,
|
||||
-0.00072658417,
|
||||
-0.0055776117,
|
||||
0.03369649,
|
||||
-0.07023698,
|
||||
-0.041115183,
|
||||
-0.06924242,
|
||||
0.0061645363,
|
||||
-0.00047588223,
|
||||
-0.03563763,
|
||||
0.011595489,
|
||||
-0.0034681638,
|
||||
0.02738642,
|
||||
0.026109103,
|
||||
-0.018220779,
|
||||
0.026244855,
|
||||
-0.067560904,
|
||||
0.026338186,
|
||||
0.016787479,
|
||||
0.065061815,
|
||||
-0.0032663948,
|
||||
-0.040305886,
|
||||
0.017459001,
|
||||
0.036517326,
|
||||
0.055479337,
|
||||
0.00085552345,
|
||||
-0.0372879,
|
||||
-0.06509678,
|
||||
-0.038734257,
|
||||
0.052903496,
|
||||
0.033298932,
|
||||
0.039541215,
|
||||
-0.09552075,
|
||||
-0.0096350545,
|
||||
-0.08214571,
|
||||
-0.024635889,
|
||||
0.012038027,
|
||||
0.00089192577,
|
||||
-0.03183621,
|
||||
-0.011991485,
|
||||
-0.03902091,
|
||||
-0.0127780195,
|
||||
-0.01724641,
|
||||
0.051544886,
|
||||
-0.0018517342,
|
||||
-0.023545155,
|
||||
0.046582974,
|
||||
0.00838307,
|
||||
0.030676562,
|
||||
0.00019708494,
|
||||
0.045098882,
|
||||
-0.031479437,
|
||||
-0.013706887,
|
||||
0.021959703,
|
||||
0.0020392945,
|
||||
-0.06168245,
|
||||
-0.03649696,
|
||||
0.035295885,
|
||||
0.02590806,
|
||||
-0.010051864,
|
||||
0.06865142,
|
||||
-0.017345844,
|
||||
0.01564999,
|
||||
-0.00623685,
|
||||
0.010844825,
|
||||
-0.013015856,
|
||||
0.022496467,
|
||||
0.07649363,
|
||||
0.036356304,
|
||||
-0.040345356,
|
||||
0.00293154,
|
||||
-0.01804687,
|
||||
-0.03515604,
|
||||
0.022299029,
|
||||
-0.03676945,
|
||||
0.07276787,
|
||||
-0.04430659,
|
||||
-0.03392204,
|
||||
-0.030020125,
|
||||
-0.022968723,
|
||||
0.029162299,
|
||||
-0.0033855392,
|
||||
0.021752143,
|
||||
0.017534897,
|
||||
-0.023780832,
|
||||
0.027371254,
|
||||
0.017058812,
|
||||
-0.0004049258,
|
||||
0.03990323,
|
||||
-0.008081489,
|
||||
-0.013143231,
|
||||
-0.06439464,
|
||||
0.018572995,
|
||||
-0.046607014,
|
||||
0.027462576,
|
||||
0.014255841,
|
||||
-0.02674485,
|
||||
0.023134982,
|
||||
-0.070987545,
|
||||
0.00939401,
|
||||
0.023703443,
|
||||
-0.009809178,
|
||||
0.022829901,
|
||||
-0.040908735,
|
||||
0.0064307996,
|
||||
0.11391804,
|
||||
-0.051118158,
|
||||
0.020216303,
|
||||
-0.02172353,
|
||||
0.04750726,
|
||||
0.018758802,
|
||||
-0.0051700706,
|
||||
-0.02455834,
|
||||
0.005184222,
|
||||
-0.036763046
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "nomic-embed-text:latest",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 8,
|
||||
"total_tokens": 8
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is the content of test file 2"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.014871168,
|
||||
0.094365,
|
||||
-0.098275684,
|
||||
0.016189486,
|
||||
0.072296426,
|
||||
-0.039229725,
|
||||
0.007638039,
|
||||
0.035811495,
|
||||
-0.03784589,
|
||||
0.022591105,
|
||||
0.15810202,
|
||||
0.009195058,
|
||||
-0.029846655,
|
||||
-0.06448414,
|
||||
-0.01898075,
|
||||
-0.02023675,
|
||||
-0.07593923,
|
||||
-0.04666322,
|
||||
0.010769107,
|
||||
0.033283222,
|
||||
0.06951838,
|
||||
0.039086174,
|
||||
-0.009640043,
|
||||
-0.008601025,
|
||||
0.039979056,
|
||||
0.02799972,
|
||||
-0.06578151,
|
||||
0.08029443,
|
||||
0.0101568075,
|
||||
-0.07898879,
|
||||
0.048795786,
|
||||
0.057297125,
|
||||
0.025737243,
|
||||
0.03572965,
|
||||
0.11485981,
|
||||
0.030900626,
|
||||
0.118485495,
|
||||
-0.041167885,
|
||||
-0.019413618,
|
||||
-0.0009897926,
|
||||
0.03717747,
|
||||
-0.012367201,
|
||||
-0.0026639055,
|
||||
0.015703445,
|
||||
-0.0046827365,
|
||||
0.023138778,
|
||||
0.012855939,
|
||||
-0.029367425,
|
||||
0.00042996072,
|
||||
-0.003222942,
|
||||
-0.055509202,
|
||||
0.012830617,
|
||||
-0.06941755,
|
||||
-0.011024706,
|
||||
0.07149942,
|
||||
0.021040803,
|
||||
0.0409756,
|
||||
0.010087916,
|
||||
-0.015326204,
|
||||
0.06633094,
|
||||
0.024846299,
|
||||
0.030543685,
|
||||
-0.036063526,
|
||||
0.04786587,
|
||||
0.08074621,
|
||||
-0.051489003,
|
||||
-0.03944393,
|
||||
-0.025607359,
|
||||
-0.030061793,
|
||||
-0.119378455,
|
||||
-0.14597124,
|
||||
-0.0019379344,
|
||||
0.008393092,
|
||||
0.023913048,
|
||||
0.028285578,
|
||||
0.017838098,
|
||||
-0.10575887,
|
||||
0.008080291,
|
||||
0.06388723,
|
||||
-0.12506105,
|
||||
-0.02536782,
|
||||
-0.11007926,
|
||||
0.051198784,
|
||||
0.007446184,
|
||||
-0.030837545,
|
||||
0.09254253,
|
||||
0.05638562,
|
||||
-0.0155668175,
|
||||
-0.031867314,
|
||||
0.018337138,
|
||||
0.02442871,
|
||||
-0.042078987,
|
||||
0.0038125275,
|
||||
0.089955,
|
||||
-0.008119613,
|
||||
0.040103614,
|
||||
0.011012824,
|
||||
0.044628628,
|
||||
0.0791957,
|
||||
0.054247666,
|
||||
-0.027651828,
|
||||
-0.03190785,
|
||||
0.041443683,
|
||||
0.041629724,
|
||||
-0.077835254,
|
||||
-0.09937542,
|
||||
0.029904107,
|
||||
-0.05434366,
|
||||
0.07058962,
|
||||
-0.04535761,
|
||||
0.03365359,
|
||||
-0.061656676,
|
||||
-0.018105442,
|
||||
-0.07228336,
|
||||
0.035377987,
|
||||
-0.03161877,
|
||||
-0.020589713,
|
||||
0.058485094,
|
||||
-0.049225487,
|
||||
0.03934316,
|
||||
0.08550028,
|
||||
-0.029991213,
|
||||
-0.05576064,
|
||||
-0.029334918,
|
||||
-0.053031918,
|
||||
-0.061839186,
|
||||
0.08176057,
|
||||
-3.3282106e-33,
|
||||
0.00018265574,
|
||||
-0.09808404,
|
||||
-0.00554673,
|
||||
0.13180184,
|
||||
0.026467713,
|
||||
-0.03976283,
|
||||
0.010410568,
|
||||
0.022475285,
|
||||
-0.07190717,
|
||||
0.005138454,
|
||||
-0.021325583,
|
||||
-0.1046733,
|
||||
0.0020021838,
|
||||
0.023773609,
|
||||
-0.057499945,
|
||||
-0.011727483,
|
||||
-0.020912478,
|
||||
0.026353713,
|
||||
0.01779019,
|
||||
-0.0148312645,
|
||||
0.064687304,
|
||||
0.045060385,
|
||||
-0.029312065,
|
||||
-0.08633001,
|
||||
-0.026792597,
|
||||
0.014552106,
|
||||
0.004505434,
|
||||
-0.06774755,
|
||||
0.034052122,
|
||||
0.013713737,
|
||||
-0.0075813113,
|
||||
-0.059718475,
|
||||
-0.016189422,
|
||||
0.044314116,
|
||||
0.026844766,
|
||||
0.026430624,
|
||||
0.024091395,
|
||||
-0.0032406747,
|
||||
-0.075288124,
|
||||
0.032822173,
|
||||
0.027104331,
|
||||
-0.026295068,
|
||||
0.04316082,
|
||||
-0.010091815,
|
||||
0.034184698,
|
||||
-0.08266358,
|
||||
-0.020962045,
|
||||
-0.00719584,
|
||||
0.068549044,
|
||||
0.005033586,
|
||||
0.0017975906,
|
||||
0.06465498,
|
||||
0.05990613,
|
||||
-0.012483792,
|
||||
0.024451919,
|
||||
0.021659598,
|
||||
-0.0046074707,
|
||||
-0.004559902,
|
||||
0.002713282,
|
||||
0.062373567,
|
||||
0.0035651235,
|
||||
0.06017224,
|
||||
-0.062707886,
|
||||
0.039937016,
|
||||
-0.0064443815,
|
||||
-0.041358124,
|
||||
-0.045459975,
|
||||
-0.1090475,
|
||||
0.08058783,
|
||||
0.055110224,
|
||||
-0.05126053,
|
||||
-0.05976516,
|
||||
0.037940193,
|
||||
0.015456569,
|
||||
-0.024956519,
|
||||
-0.037877902,
|
||||
-0.006799,
|
||||
0.031685203,
|
||||
-0.036858797,
|
||||
-0.055584695,
|
||||
-0.048513155,
|
||||
-0.07101657,
|
||||
-0.041681714,
|
||||
-0.04429727,
|
||||
-0.09584418,
|
||||
-0.060873836,
|
||||
0.008867621,
|
||||
-0.106438614,
|
||||
0.040050562,
|
||||
-0.084729105,
|
||||
0.018111277,
|
||||
0.010153493,
|
||||
-0.08883196,
|
||||
-0.063969284,
|
||||
0.08611972,
|
||||
1.4074298e-33,
|
||||
0.03433739,
|
||||
0.037653737,
|
||||
-0.05348675,
|
||||
0.0015385789,
|
||||
0.026684077,
|
||||
0.026603375,
|
||||
0.07006387,
|
||||
-0.034265522,
|
||||
-0.018221779,
|
||||
0.10960259,
|
||||
0.013464475,
|
||||
-0.008325532,
|
||||
0.019438146,
|
||||
-0.039553005,
|
||||
0.03469477,
|
||||
-0.0123773115,
|
||||
-0.013288484,
|
||||
-0.048081715,
|
||||
-0.019539693,
|
||||
-0.0033996427,
|
||||
-0.024453517,
|
||||
0.061505664,
|
||||
0.119236834,
|
||||
0.026294904,
|
||||
-0.01607055,
|
||||
-0.011499089,
|
||||
0.04267117,
|
||||
0.0295908,
|
||||
0.022084564,
|
||||
0.007893738,
|
||||
0.052055445,
|
||||
0.05781507,
|
||||
-0.13408813,
|
||||
0.01778491,
|
||||
0.021400984,
|
||||
-0.12113228,
|
||||
0.10535695,
|
||||
-0.07358604,
|
||||
-0.013651957,
|
||||
0.04049295,
|
||||
0.054150987,
|
||||
0.0987462,
|
||||
0.0110208625,
|
||||
0.040327504,
|
||||
0.034936633,
|
||||
0.10400846,
|
||||
0.12958324,
|
||||
-0.024531014,
|
||||
0.002284699,
|
||||
-0.044239815,
|
||||
0.049778443,
|
||||
-0.055788964,
|
||||
0.015235888,
|
||||
0.0034493478,
|
||||
-0.02607555,
|
||||
0.060282644,
|
||||
-0.028004775,
|
||||
0.040875163,
|
||||
-0.023749253,
|
||||
0.002289086,
|
||||
0.04982698,
|
||||
0.046928305,
|
||||
-0.064160004,
|
||||
0.013701618,
|
||||
0.015511878,
|
||||
-0.054725982,
|
||||
-0.0459802,
|
||||
0.03258067,
|
||||
0.027034523,
|
||||
0.01643672,
|
||||
-0.041782584,
|
||||
-0.03698569,
|
||||
-0.023043923,
|
||||
-0.07073365,
|
||||
0.028486207,
|
||||
0.0017764921,
|
||||
-0.03352676,
|
||||
-0.009977863,
|
||||
0.024488676,
|
||||
-0.01789395,
|
||||
0.029737154,
|
||||
-0.026266927,
|
||||
-0.03567072,
|
||||
0.07469971,
|
||||
0.028393274,
|
||||
-0.029625034,
|
||||
-0.01053128,
|
||||
0.09147493,
|
||||
-0.018718474,
|
||||
0.0012933073,
|
||||
-0.021214467,
|
||||
0.07475739,
|
||||
-0.007773536,
|
||||
0.048597455,
|
||||
0.005216022,
|
||||
-1.6914717e-08,
|
||||
-0.05724563,
|
||||
-0.0938908,
|
||||
-0.034359876,
|
||||
-0.037500683,
|
||||
-0.020235153,
|
||||
0.06142227,
|
||||
-0.042273093,
|
||||
-0.008759724,
|
||||
-0.009908796,
|
||||
0.016232042,
|
||||
-0.014239323,
|
||||
0.024709346,
|
||||
-0.030538557,
|
||||
-0.05391127,
|
||||
-0.051778477,
|
||||
0.01277344,
|
||||
0.0036140021,
|
||||
-0.012569925,
|
||||
-0.025041323,
|
||||
-0.0203936,
|
||||
0.025865255,
|
||||
0.010908398,
|
||||
0.027834684,
|
||||
0.009661084,
|
||||
-0.006598172,
|
||||
0.07860872,
|
||||
0.054516125,
|
||||
0.042956624,
|
||||
-0.06275145,
|
||||
-0.025701547,
|
||||
0.08085865,
|
||||
0.030041302,
|
||||
0.02248997,
|
||||
-0.0840195,
|
||||
0.00029938898,
|
||||
0.10966559,
|
||||
0.118907265,
|
||||
0.063014604,
|
||||
0.037847042,
|
||||
0.032069027,
|
||||
-0.05345487,
|
||||
-0.022730324,
|
||||
0.0071888734,
|
||||
0.037573762,
|
||||
-0.020178014,
|
||||
-0.090167634,
|
||||
-0.07191704,
|
||||
-0.02604166,
|
||||
-0.043885063,
|
||||
-0.14087014,
|
||||
-0.017230472,
|
||||
-0.012063355,
|
||||
-0.046736836,
|
||||
0.039048597,
|
||||
-0.060394738,
|
||||
0.022166032,
|
||||
0.025670663,
|
||||
0.022949725,
|
||||
-0.06707243,
|
||||
-0.014654702,
|
||||
0.057985142,
|
||||
0.10511708,
|
||||
0.05698323,
|
||||
-0.017205814
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 8,
|
||||
"total_tokens": 8
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is batch test file 0"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
0.020637129,
|
||||
0.048547756,
|
||||
-0.12516363,
|
||||
0.01991118,
|
||||
-0.006535745,
|
||||
-0.017178575,
|
||||
0.027727997,
|
||||
0.032170568,
|
||||
-0.07302972,
|
||||
0.008939002,
|
||||
0.11493648,
|
||||
0.0058907545,
|
||||
0.0058539375,
|
||||
-0.077171296,
|
||||
-0.06883132,
|
||||
0.0039748913,
|
||||
-0.046849657,
|
||||
-0.072902456,
|
||||
-0.010890429,
|
||||
-0.0019311906,
|
||||
-0.011614798,
|
||||
0.003689495,
|
||||
-0.03695609,
|
||||
-0.009029024,
|
||||
0.017461002,
|
||||
-0.004713484,
|
||||
-0.010254731,
|
||||
-0.026636763,
|
||||
-0.026125714,
|
||||
-0.046913657,
|
||||
0.017024228,
|
||||
0.0713477,
|
||||
0.07881179,
|
||||
0.03789051,
|
||||
0.21716279,
|
||||
-0.0077837943,
|
||||
0.04686894,
|
||||
0.020414647,
|
||||
7.314368e-05,
|
||||
0.0103133675,
|
||||
0.059848394,
|
||||
-0.04321678,
|
||||
-0.011937493,
|
||||
-0.021149047,
|
||||
0.021315353,
|
||||
-0.00072822213,
|
||||
-0.046116166,
|
||||
-0.0046820445,
|
||||
0.016943695,
|
||||
-0.03249135,
|
||||
-0.055184096,
|
||||
4.1543382e-05,
|
||||
-0.034172166,
|
||||
-0.023247559,
|
||||
0.020267941,
|
||||
0.012827845,
|
||||
0.065036125,
|
||||
0.07180022,
|
||||
-0.013490698,
|
||||
0.06376413,
|
||||
-0.017730094,
|
||||
-0.01806601,
|
||||
-0.034191083,
|
||||
0.008955718,
|
||||
0.098446764,
|
||||
-0.0061265854,
|
||||
-0.06815829,
|
||||
-0.039525956,
|
||||
0.060588058,
|
||||
-0.094874755,
|
||||
-0.11774928,
|
||||
0.019538416,
|
||||
-0.014697532,
|
||||
0.04773719,
|
||||
-0.061298393,
|
||||
0.030337377,
|
||||
-0.0022184649,
|
||||
0.019007793,
|
||||
0.024370821,
|
||||
-0.07063359,
|
||||
-0.07582954,
|
||||
-0.10816809,
|
||||
0.031845964,
|
||||
-0.057830192,
|
||||
-0.04169559,
|
||||
0.0752806,
|
||||
0.019289386,
|
||||
0.028845867,
|
||||
0.0077010663,
|
||||
0.013930818,
|
||||
-0.067987345,
|
||||
0.012679873,
|
||||
-0.07907268,
|
||||
0.0143718915,
|
||||
-0.021433424,
|
||||
0.11880779,
|
||||
-0.016258432,
|
||||
0.07099568,
|
||||
0.035778854,
|
||||
0.028776454,
|
||||
0.013304291,
|
||||
-0.05192297,
|
||||
0.026758345,
|
||||
0.10282426,
|
||||
-0.003306269,
|
||||
-0.03239622,
|
||||
0.083044045,
|
||||
-0.0412691,
|
||||
0.043435257,
|
||||
-0.043423533,
|
||||
-0.013239603,
|
||||
-0.0029038454,
|
||||
0.038365215,
|
||||
-0.10401672,
|
||||
0.012744224,
|
||||
-0.122984126,
|
||||
-0.008942817,
|
||||
0.06162198,
|
||||
-0.120285526,
|
||||
0.043005254,
|
||||
0.04814879,
|
||||
-0.036352232,
|
||||
-0.003885529,
|
||||
-0.018503373,
|
||||
-0.088186465,
|
||||
-0.0031517749,
|
||||
0.09290919,
|
||||
-1.1695094e-33,
|
||||
-0.015589721,
|
||||
-0.13189551,
|
||||
0.008088751,
|
||||
0.06899503,
|
||||
0.07353927,
|
||||
-0.030646399,
|
||||
0.05110342,
|
||||
0.03081624,
|
||||
-0.07850498,
|
||||
-0.021147482,
|
||||
0.00017823944,
|
||||
-0.10502706,
|
||||
0.030078856,
|
||||
0.02572523,
|
||||
-0.068158925,
|
||||
-0.025015576,
|
||||
0.021830637,
|
||||
0.049748335,
|
||||
0.01520941,
|
||||
-0.080153145,
|
||||
0.06796621,
|
||||
0.021865685,
|
||||
-0.034017574,
|
||||
-0.030821111,
|
||||
-0.048006665,
|
||||
0.0005615041,
|
||||
-0.0137883695,
|
||||
-0.04500587,
|
||||
0.015368256,
|
||||
-0.0043663937,
|
||||
0.037706476,
|
||||
0.0049090013,
|
||||
-0.06216566,
|
||||
0.03060772,
|
||||
0.030548712,
|
||||
0.029262561,
|
||||
0.020701125,
|
||||
0.0056516766,
|
||||
0.010610447,
|
||||
0.019530762,
|
||||
-0.05664136,
|
||||
-0.022654066,
|
||||
-0.0010107337,
|
||||
-0.020805702,
|
||||
-0.012242364,
|
||||
-0.05591731,
|
||||
-0.049421698,
|
||||
0.024721064,
|
||||
0.05803342,
|
||||
0.010474127,
|
||||
-0.008790625,
|
||||
0.025362873,
|
||||
0.020258408,
|
||||
0.004368581,
|
||||
-0.01018003,
|
||||
0.012385932,
|
||||
-0.037656736,
|
||||
-0.05642639,
|
||||
0.020923307,
|
||||
0.022813153,
|
||||
-0.005735433,
|
||||
0.015326356,
|
||||
-0.108707875,
|
||||
0.048076265,
|
||||
0.023256551,
|
||||
-0.10311626,
|
||||
0.061980195,
|
||||
-0.07340407,
|
||||
0.051583096,
|
||||
0.07360003,
|
||||
-0.029443117,
|
||||
-0.014564469,
|
||||
0.042043358,
|
||||
-0.020252181,
|
||||
0.0147808045,
|
||||
-0.0285806,
|
||||
0.07891856,
|
||||
0.056849223,
|
||||
-0.106308356,
|
||||
0.0197874,
|
||||
0.0269322,
|
||||
-0.04749746,
|
||||
-0.066681586,
|
||||
-0.10474516,
|
||||
0.012599429,
|
||||
-0.056163482,
|
||||
-0.04901015,
|
||||
-0.04571026,
|
||||
0.09704481,
|
||||
-0.105899766,
|
||||
0.044303197,
|
||||
-0.020125533,
|
||||
-0.0368709,
|
||||
-0.015417924,
|
||||
0.042297333,
|
||||
-8.289866e-35,
|
||||
0.07415767,
|
||||
0.10998298,
|
||||
-0.016995763,
|
||||
0.01066263,
|
||||
-0.0012327223,
|
||||
0.028000232,
|
||||
0.0714317,
|
||||
-0.02320065,
|
||||
-0.07778205,
|
||||
0.11864239,
|
||||
-0.016559754,
|
||||
0.037961867,
|
||||
0.02930022,
|
||||
-0.008237686,
|
||||
0.059777655,
|
||||
0.008086454,
|
||||
0.02075205,
|
||||
0.025284613,
|
||||
-0.055471037,
|
||||
0.0073576584,
|
||||
-0.013398135,
|
||||
0.11896543,
|
||||
-0.014611002,
|
||||
0.07691816,
|
||||
-0.019711656,
|
||||
-0.01920917,
|
||||
-0.004744884,
|
||||
0.08173054,
|
||||
0.019665759,
|
||||
-0.013193461,
|
||||
0.06215852,
|
||||
0.07420406,
|
||||
-0.073212065,
|
||||
0.036052067,
|
||||
0.07328616,
|
||||
-0.057373393,
|
||||
0.08346425,
|
||||
0.018834447,
|
||||
0.03309735,
|
||||
0.041197047,
|
||||
0.033917964,
|
||||
0.09151449,
|
||||
-0.051731598,
|
||||
0.049615093,
|
||||
0.01124018,
|
||||
0.06661862,
|
||||
0.07268375,
|
||||
-0.013245848,
|
||||
-0.039673895,
|
||||
-0.012173254,
|
||||
0.0017787582,
|
||||
-0.05746287,
|
||||
-0.013884767,
|
||||
0.020205025,
|
||||
-0.029692367,
|
||||
-0.031010685,
|
||||
0.0149556715,
|
||||
0.026381323,
|
||||
-0.025382591,
|
||||
0.0074336748,
|
||||
-0.00949915,
|
||||
0.015655186,
|
||||
-0.0012397208,
|
||||
-0.0032508406,
|
||||
-0.046632554,
|
||||
-0.0030316226,
|
||||
-0.007273208,
|
||||
0.064231135,
|
||||
-0.034431897,
|
||||
-0.06433184,
|
||||
0.045421343,
|
||||
-0.010773523,
|
||||
-0.017881984,
|
||||
0.010312532,
|
||||
-0.024369273,
|
||||
-0.008478495,
|
||||
-0.02457377,
|
||||
-0.0263535,
|
||||
-0.027263613,
|
||||
0.047060315,
|
||||
0.08128726,
|
||||
0.0045517692,
|
||||
-0.010821656,
|
||||
0.026526682,
|
||||
0.018961033,
|
||||
0.059243083,
|
||||
0.001561823,
|
||||
0.09838158,
|
||||
0.00822081,
|
||||
0.008796511,
|
||||
-0.0060577285,
|
||||
0.028892087,
|
||||
0.08253284,
|
||||
0.049560018,
|
||||
0.023363132,
|
||||
-1.498271e-08,
|
||||
-0.036891207,
|
||||
-0.10629833,
|
||||
0.030452948,
|
||||
0.049268734,
|
||||
-0.0030453752,
|
||||
0.07413954,
|
||||
-0.07043819,
|
||||
-0.034285706,
|
||||
-0.009679971,
|
||||
-0.046219327,
|
||||
0.013510038,
|
||||
-0.018686565,
|
||||
-0.048570327,
|
||||
0.0028313443,
|
||||
-0.06190722,
|
||||
-0.053201936,
|
||||
0.0060967463,
|
||||
-0.043467365,
|
||||
0.042226154,
|
||||
0.03455835,
|
||||
-0.0375257,
|
||||
0.023590367,
|
||||
0.054896712,
|
||||
0.029878648,
|
||||
0.019286606,
|
||||
0.026097741,
|
||||
0.06938145,
|
||||
0.06272366,
|
||||
-0.09566521,
|
||||
-0.07481147,
|
||||
0.025204772,
|
||||
0.039396077,
|
||||
0.036375154,
|
||||
-0.01104443,
|
||||
-0.028223084,
|
||||
0.111878626,
|
||||
0.13400707,
|
||||
0.06680113,
|
||||
-0.011737675,
|
||||
-0.03585406,
|
||||
-0.07978788,
|
||||
0.032793757,
|
||||
-0.0021075818,
|
||||
-0.028365146,
|
||||
-0.042218164,
|
||||
-0.08132239,
|
||||
-0.0753423,
|
||||
-0.043771427,
|
||||
-0.015633285,
|
||||
-0.14193884,
|
||||
-0.055949364,
|
||||
0.025526602,
|
||||
-0.023186589,
|
||||
0.061106257,
|
||||
-0.056208834,
|
||||
0.00838827,
|
||||
0.014720396,
|
||||
-0.014650135,
|
||||
-0.012830787,
|
||||
0.08434067,
|
||||
0.024660436,
|
||||
0.05366935,
|
||||
0.005782819,
|
||||
-0.10599063
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 6,
|
||||
"total_tokens": 6
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,807 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "nomic-embed-text:latest",
|
||||
"input": [
|
||||
"This is batch test file 0"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "nomic-embed-text:latest"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
0.04614301,
|
||||
-0.020081447,
|
||||
-0.13696706,
|
||||
0.014573554,
|
||||
0.0701535,
|
||||
-0.023059264,
|
||||
0.0017123591,
|
||||
0.046829354,
|
||||
-0.04367561,
|
||||
-0.114361376,
|
||||
-0.035698596,
|
||||
0.06475817,
|
||||
0.05364872,
|
||||
0.038444433,
|
||||
-0.01979135,
|
||||
-0.037200496,
|
||||
0.036657624,
|
||||
-0.06543346,
|
||||
0.019384816,
|
||||
0.014107363,
|
||||
0.08575862,
|
||||
0.005253997,
|
||||
-0.068923116,
|
||||
0.010090121,
|
||||
0.12664902,
|
||||
0.007504762,
|
||||
-0.01953818,
|
||||
0.050352264,
|
||||
-0.054028552,
|
||||
-0.032810874,
|
||||
0.06410688,
|
||||
0.02482149,
|
||||
0.013947014,
|
||||
-0.018964177,
|
||||
-0.014869268,
|
||||
-0.014962293,
|
||||
0.015406188,
|
||||
0.0019375562,
|
||||
0.045115244,
|
||||
0.01663003,
|
||||
0.011144363,
|
||||
-0.0072048977,
|
||||
0.03155388,
|
||||
-0.044834215,
|
||||
-0.0060860706,
|
||||
0.04020486,
|
||||
0.018086052,
|
||||
0.04788624,
|
||||
-0.00983748,
|
||||
0.013905991,
|
||||
-0.015586391,
|
||||
-0.04333209,
|
||||
-0.025767654,
|
||||
0.0115056895,
|
||||
0.023292946,
|
||||
7.8543904e-05,
|
||||
-0.025338432,
|
||||
0.027270807,
|
||||
-0.033308506,
|
||||
-0.0034880606,
|
||||
0.027414253,
|
||||
0.023092583,
|
||||
-0.046051297,
|
||||
0.05614708,
|
||||
0.0013318001,
|
||||
-0.009060849,
|
||||
-0.025585877,
|
||||
0.01975632,
|
||||
0.005582998,
|
||||
0.029287277,
|
||||
0.04566754,
|
||||
-0.025097856,
|
||||
-0.031588476,
|
||||
0.010089205,
|
||||
-0.032345522,
|
||||
-0.05282097,
|
||||
-0.027767532,
|
||||
-0.064588815,
|
||||
-0.048720118,
|
||||
-0.019109437,
|
||||
0.018205147,
|
||||
0.023525203,
|
||||
0.030500842,
|
||||
0.062187836,
|
||||
0.056703616,
|
||||
-0.012734468,
|
||||
-0.0023994881,
|
||||
-0.01470007,
|
||||
-0.014610555,
|
||||
0.041148573,
|
||||
0.02209264,
|
||||
0.016696744,
|
||||
0.01664216,
|
||||
-0.042584907,
|
||||
-0.030513879,
|
||||
0.009484068,
|
||||
0.038292237,
|
||||
0.049507294,
|
||||
-0.008864681,
|
||||
-0.026641846,
|
||||
-0.00806868,
|
||||
0.015242631,
|
||||
0.03148721,
|
||||
-0.029101137,
|
||||
-0.001128117,
|
||||
0.047483873,
|
||||
-0.021579307,
|
||||
-0.0061297114,
|
||||
-0.051103026,
|
||||
-0.01100252,
|
||||
-0.007417349,
|
||||
0.033126004,
|
||||
-0.03208908,
|
||||
-0.014004128,
|
||||
0.0055860616,
|
||||
-0.004471519,
|
||||
0.040237978,
|
||||
-0.011666332,
|
||||
-0.03375841,
|
||||
0.010431967,
|
||||
0.015548171,
|
||||
-0.003738259,
|
||||
0.006507693,
|
||||
0.044239193,
|
||||
0.022051405,
|
||||
0.0369485,
|
||||
-0.08473572,
|
||||
0.050257385,
|
||||
0.021561263,
|
||||
-0.038263254,
|
||||
-0.0028757958,
|
||||
0.004459847,
|
||||
-0.0499833,
|
||||
-0.05051039,
|
||||
0.038672045,
|
||||
0.027102912,
|
||||
-0.038302545,
|
||||
-0.04273586,
|
||||
-0.008564898,
|
||||
0.010148107,
|
||||
-0.033453222,
|
||||
0.025933161,
|
||||
0.034907244,
|
||||
0.05704188,
|
||||
0.047914036,
|
||||
-0.033055704,
|
||||
0.037968747,
|
||||
0.021832222,
|
||||
-0.021085719,
|
||||
0.020705225,
|
||||
0.0013380332,
|
||||
0.0033825892,
|
||||
-0.004659198,
|
||||
0.03569596,
|
||||
0.035501115,
|
||||
-0.07247981,
|
||||
0.024580602,
|
||||
-0.016031673,
|
||||
0.0043628234,
|
||||
0.044260535,
|
||||
0.022414433,
|
||||
-0.024638122,
|
||||
-0.027389847,
|
||||
-0.015699405,
|
||||
-0.0736989,
|
||||
-0.07402259,
|
||||
0.021693923,
|
||||
0.08675446,
|
||||
0.07019457,
|
||||
-0.010507776,
|
||||
-0.053982176,
|
||||
-0.050338153,
|
||||
0.022691121,
|
||||
-0.009254433,
|
||||
-0.08471297,
|
||||
-0.02192142,
|
||||
0.01370606,
|
||||
-0.007591457,
|
||||
0.04464477,
|
||||
-0.041420456,
|
||||
0.014799598,
|
||||
-0.017481469,
|
||||
0.025636235,
|
||||
0.021010395,
|
||||
-0.007866782,
|
||||
-0.044533994,
|
||||
0.02992151,
|
||||
-0.01817989,
|
||||
-0.046332225,
|
||||
-0.038017664,
|
||||
-0.010766631,
|
||||
-0.053923517,
|
||||
-0.06885444,
|
||||
-0.083982274,
|
||||
0.0044967085,
|
||||
-0.059554394,
|
||||
-0.012864852,
|
||||
-0.006990424,
|
||||
0.04381485,
|
||||
-0.019732013,
|
||||
0.0047561186,
|
||||
0.012573004,
|
||||
-0.084608465,
|
||||
0.044956904,
|
||||
-0.043024026,
|
||||
-0.008965278,
|
||||
-0.018597735,
|
||||
0.023019703,
|
||||
-0.058244467,
|
||||
0.03946037,
|
||||
0.0070161144,
|
||||
-0.0072559468,
|
||||
-0.038439214,
|
||||
0.007182057,
|
||||
0.04479635,
|
||||
-0.04825045,
|
||||
0.020951761,
|
||||
-0.04177098,
|
||||
-0.015445904,
|
||||
0.0024602767,
|
||||
0.019107025,
|
||||
0.01879466,
|
||||
0.015647886,
|
||||
-0.03868733,
|
||||
0.0010552967,
|
||||
-0.031725515,
|
||||
0.003677792,
|
||||
-0.008869332,
|
||||
0.016350579,
|
||||
-0.025660282,
|
||||
-0.0033997998,
|
||||
-0.053593792,
|
||||
0.01300085,
|
||||
0.014504953,
|
||||
-0.04167999,
|
||||
-0.013626902,
|
||||
-0.013473784,
|
||||
0.04477186,
|
||||
-0.039079364,
|
||||
0.045125194,
|
||||
-0.038965665,
|
||||
0.041032016,
|
||||
-0.021128332,
|
||||
-0.00079906755,
|
||||
0.0105881365,
|
||||
-0.023438545,
|
||||
-0.009942863,
|
||||
-0.028716002,
|
||||
0.031107599,
|
||||
0.017214399,
|
||||
-0.027654208,
|
||||
0.013554825,
|
||||
0.019173222,
|
||||
0.055249233,
|
||||
0.00617875,
|
||||
-0.01951432,
|
||||
-0.008078177,
|
||||
-0.045365352,
|
||||
-0.013410786,
|
||||
-0.06576592,
|
||||
0.0258962,
|
||||
-0.04870149,
|
||||
0.028375717,
|
||||
0.02127929,
|
||||
0.0074190334,
|
||||
-0.015849123,
|
||||
-0.050413407,
|
||||
0.027654368,
|
||||
-0.01447592,
|
||||
-0.0046318294,
|
||||
0.003353468,
|
||||
0.02456171,
|
||||
-0.006699941,
|
||||
-0.04072025,
|
||||
0.030406132,
|
||||
0.003700867,
|
||||
0.04991202,
|
||||
0.043061696,
|
||||
-0.0014100377,
|
||||
0.038879305,
|
||||
-0.02551224,
|
||||
-0.03253989,
|
||||
0.002335025,
|
||||
0.0066566374,
|
||||
0.013019894,
|
||||
-0.017884579,
|
||||
0.03333752,
|
||||
-0.005329557,
|
||||
0.025703372,
|
||||
0.01243421,
|
||||
0.0015536154,
|
||||
-0.0011326541,
|
||||
-0.02956871,
|
||||
0.010284604,
|
||||
-0.0017640645,
|
||||
0.030444842,
|
||||
0.024831444,
|
||||
-0.015894072,
|
||||
-0.017051669,
|
||||
-0.012481152,
|
||||
-0.021874228,
|
||||
0.032245617,
|
||||
0.029441461,
|
||||
-0.019289171,
|
||||
0.015042458,
|
||||
-0.048809912,
|
||||
0.045543794,
|
||||
-0.025887825,
|
||||
-0.0017429133,
|
||||
0.050035972,
|
||||
-0.094813764,
|
||||
-0.026645338,
|
||||
-0.03496652,
|
||||
0.02787559,
|
||||
-0.009335962,
|
||||
0.050203443,
|
||||
0.007864018,
|
||||
0.008651598,
|
||||
0.07439614,
|
||||
-0.04608253,
|
||||
0.008741113,
|
||||
-0.096183665,
|
||||
0.01909248,
|
||||
0.02903942,
|
||||
-0.020657493,
|
||||
0.03056416,
|
||||
0.025593685,
|
||||
0.05326756,
|
||||
-0.035201855,
|
||||
-0.0042431992,
|
||||
0.047313657,
|
||||
0.04643017,
|
||||
-0.015038583,
|
||||
-0.03623203,
|
||||
0.06975197,
|
||||
0.02893981,
|
||||
-0.012065428,
|
||||
0.03489605,
|
||||
-0.02045082,
|
||||
0.08106463,
|
||||
0.03046569,
|
||||
0.05845714,
|
||||
0.038226783,
|
||||
-0.0039640213,
|
||||
0.020310445,
|
||||
-0.044298742,
|
||||
-0.011063444,
|
||||
0.031646963,
|
||||
-0.016750742,
|
||||
0.06093846,
|
||||
-0.0024345908,
|
||||
0.0137670245,
|
||||
0.01068818,
|
||||
0.00028172386,
|
||||
0.024276268,
|
||||
0.007246687,
|
||||
0.017009424,
|
||||
-0.0058112424,
|
||||
0.055742696,
|
||||
0.0020487534,
|
||||
0.0041393945,
|
||||
-0.002708682,
|
||||
-0.0111793615,
|
||||
-0.016895374,
|
||||
0.009005465,
|
||||
0.025580926,
|
||||
-0.015197682,
|
||||
0.0152440565,
|
||||
0.049733438,
|
||||
0.00909726,
|
||||
-0.04997614,
|
||||
-0.054340348,
|
||||
0.047531743,
|
||||
0.052675292,
|
||||
0.0002477018,
|
||||
0.017530492,
|
||||
-0.04548658,
|
||||
0.0034042338,
|
||||
-0.027109472,
|
||||
0.0540901,
|
||||
0.05400029,
|
||||
-0.039156314,
|
||||
-0.010473526,
|
||||
0.036758192,
|
||||
0.012307892,
|
||||
-0.043290082,
|
||||
0.021435479,
|
||||
-0.013614977,
|
||||
0.047010962,
|
||||
0.061001405,
|
||||
0.0067372657,
|
||||
0.0227589,
|
||||
-0.0519168,
|
||||
0.012738339,
|
||||
-0.027280986,
|
||||
0.0012095303,
|
||||
0.053970173,
|
||||
0.011808772,
|
||||
-0.06391073,
|
||||
0.049324006,
|
||||
-0.016165268,
|
||||
-0.035052363,
|
||||
0.011938826,
|
||||
0.033804722,
|
||||
-0.033935532,
|
||||
0.014987266,
|
||||
-0.03362387,
|
||||
-0.022560425,
|
||||
0.05126289,
|
||||
-0.008983691,
|
||||
0.05116898,
|
||||
-0.053964064,
|
||||
-0.038813572,
|
||||
-0.06834585,
|
||||
0.0425859,
|
||||
0.029469976,
|
||||
0.06586096,
|
||||
0.056827266,
|
||||
-0.028529037,
|
||||
-0.022799347,
|
||||
0.03930842,
|
||||
0.009058165,
|
||||
0.0029452725,
|
||||
-0.046222363,
|
||||
-0.015354657,
|
||||
0.020766245,
|
||||
0.00544761,
|
||||
0.054154783,
|
||||
0.024518205,
|
||||
-0.0762551,
|
||||
-0.03815425,
|
||||
0.030558256,
|
||||
-0.014623021,
|
||||
0.04429291,
|
||||
-0.02593325,
|
||||
0.06950066,
|
||||
-0.011652937,
|
||||
0.00784224,
|
||||
0.010082946,
|
||||
0.02640965,
|
||||
0.044778366,
|
||||
-0.017441178,
|
||||
-0.042124864,
|
||||
0.030845765,
|
||||
-0.047991402,
|
||||
0.1127873,
|
||||
0.11150797,
|
||||
-0.0745599,
|
||||
-0.059560712,
|
||||
-0.00808373,
|
||||
-0.008904883,
|
||||
0.047381986,
|
||||
-0.03259649,
|
||||
-0.0034343451,
|
||||
0.043409795,
|
||||
-0.011778097,
|
||||
0.017888952,
|
||||
-0.042976636,
|
||||
-0.014014427,
|
||||
0.013991117,
|
||||
0.008008242,
|
||||
-0.005016844,
|
||||
0.053890087,
|
||||
0.056538153,
|
||||
0.016641492,
|
||||
-0.011209175,
|
||||
0.005071369,
|
||||
-0.031119458,
|
||||
-0.012060056,
|
||||
0.047321502,
|
||||
-0.01410517,
|
||||
-0.06337502,
|
||||
0.057011377,
|
||||
-0.046111424,
|
||||
-0.022285707,
|
||||
0.00068395643,
|
||||
-0.01453697,
|
||||
0.0030104562,
|
||||
0.031148981,
|
||||
0.029581407,
|
||||
0.007647941,
|
||||
0.011242783,
|
||||
-0.026178291,
|
||||
-0.05194385,
|
||||
0.037139274,
|
||||
0.026292743,
|
||||
0.01298006,
|
||||
0.023150109,
|
||||
0.06221823,
|
||||
-0.024437338,
|
||||
0.056873403,
|
||||
0.027463028,
|
||||
0.07723492,
|
||||
0.0019251422,
|
||||
0.042778768,
|
||||
-0.026794884,
|
||||
-0.016140813,
|
||||
-0.037990715,
|
||||
0.0015520528,
|
||||
0.01605836,
|
||||
-0.012476547,
|
||||
-0.01679565,
|
||||
0.027481532,
|
||||
0.018949807,
|
||||
0.010083091,
|
||||
-0.01057625,
|
||||
-0.024935285,
|
||||
-0.031943906,
|
||||
-0.051917356,
|
||||
-0.04344679,
|
||||
-0.04837223,
|
||||
-0.009939983,
|
||||
0.040695325,
|
||||
0.024695948,
|
||||
0.063317895,
|
||||
-0.0018597379,
|
||||
0.016552558,
|
||||
-0.047521863,
|
||||
-0.07224005,
|
||||
0.042071674,
|
||||
0.016915316,
|
||||
0.014148548,
|
||||
0.01878253,
|
||||
-0.026108567,
|
||||
-0.06437781,
|
||||
0.021399872,
|
||||
0.011175348,
|
||||
0.0033761705,
|
||||
-0.004680718,
|
||||
0.03344319,
|
||||
0.0031177911,
|
||||
0.053175025,
|
||||
0.028025331,
|
||||
-0.0069551654,
|
||||
-0.034634676,
|
||||
-0.012221638,
|
||||
-0.035786934,
|
||||
0.04296345,
|
||||
-0.01631924,
|
||||
0.060271725,
|
||||
-0.04230959,
|
||||
-0.0064216405,
|
||||
0.0013953961,
|
||||
-0.041444454,
|
||||
-0.008569435,
|
||||
-0.01984154,
|
||||
-0.061582044,
|
||||
0.049848285,
|
||||
-0.010022811,
|
||||
-0.07785035,
|
||||
-0.006366211,
|
||||
-0.012778517,
|
||||
-0.037107654,
|
||||
-0.034078293,
|
||||
-0.0019027964,
|
||||
-0.018393178,
|
||||
-0.031273652,
|
||||
-0.030624373,
|
||||
-0.047289733,
|
||||
-0.055507194,
|
||||
0.0149980355,
|
||||
0.009802669,
|
||||
0.05346352,
|
||||
0.011616594,
|
||||
0.040882636,
|
||||
-0.05801636,
|
||||
-0.018325027,
|
||||
0.033699974,
|
||||
-0.015700053,
|
||||
-0.018874831,
|
||||
0.00975098,
|
||||
-0.028787887,
|
||||
-0.010430304,
|
||||
-0.019937277,
|
||||
-0.025684841,
|
||||
-0.017275153,
|
||||
0.048182886,
|
||||
0.040767677,
|
||||
-0.006017042,
|
||||
-0.012711738,
|
||||
-0.0010345151,
|
||||
0.015744662,
|
||||
0.023162043,
|
||||
0.02130765,
|
||||
-0.0024493549,
|
||||
0.015457228,
|
||||
0.037933253,
|
||||
-0.031316977,
|
||||
0.06891338,
|
||||
0.005748761,
|
||||
-0.07730445,
|
||||
-0.032125294,
|
||||
0.036361482,
|
||||
0.0061598606,
|
||||
0.018043444,
|
||||
0.038325332,
|
||||
-0.036203355,
|
||||
-0.0123121375,
|
||||
-0.022851182,
|
||||
-0.035532467,
|
||||
-0.041686766,
|
||||
0.03709366,
|
||||
-0.0017735043,
|
||||
-0.018472947,
|
||||
-0.045957465,
|
||||
-0.023627242,
|
||||
0.01808581,
|
||||
0.015027068,
|
||||
-0.042559687,
|
||||
-0.009885546,
|
||||
0.057179235,
|
||||
-0.03215653,
|
||||
-0.048862357,
|
||||
-0.012386838,
|
||||
-0.021847295,
|
||||
-0.044682942,
|
||||
0.040646516,
|
||||
0.00038476288,
|
||||
0.005513208,
|
||||
-0.03062349,
|
||||
0.011521192,
|
||||
-0.035988722,
|
||||
0.061369143,
|
||||
-0.020910813,
|
||||
0.075483516,
|
||||
-0.045259465,
|
||||
-0.02859422,
|
||||
0.015579937,
|
||||
0.0075254533,
|
||||
0.038143836,
|
||||
-0.045940828,
|
||||
0.027484732,
|
||||
-0.091758996,
|
||||
-0.048610084,
|
||||
-0.095563754,
|
||||
0.0004537795,
|
||||
-0.05040322,
|
||||
0.02240349,
|
||||
0.046084013,
|
||||
0.04480506,
|
||||
0.037050348,
|
||||
1.0597447e-05,
|
||||
-0.018571958,
|
||||
0.009857055,
|
||||
0.021747472,
|
||||
0.031625595,
|
||||
-0.03629067,
|
||||
0.037058298,
|
||||
0.041504655,
|
||||
-0.03894645,
|
||||
0.046530657,
|
||||
0.08956203,
|
||||
0.05101704,
|
||||
0.005822723,
|
||||
-0.014409921,
|
||||
0.0050498573,
|
||||
0.039041325,
|
||||
-0.010459366,
|
||||
-0.022216242,
|
||||
-0.07559245,
|
||||
0.019515479,
|
||||
-0.010434134,
|
||||
-0.040965218,
|
||||
0.006768683,
|
||||
0.021648958,
|
||||
0.059341215,
|
||||
0.0044922573,
|
||||
-0.011139294,
|
||||
0.023696495,
|
||||
-0.04251101,
|
||||
0.028621383,
|
||||
0.005927879,
|
||||
0.05084491,
|
||||
-0.01525845,
|
||||
0.03151167,
|
||||
0.008018476,
|
||||
0.05309983,
|
||||
0.059823282,
|
||||
-0.02189311,
|
||||
0.010798892,
|
||||
0.0027545195,
|
||||
-0.024435053,
|
||||
0.042531513,
|
||||
0.028011957,
|
||||
0.0147431465,
|
||||
-0.062116392,
|
||||
0.032930456,
|
||||
-0.03597175,
|
||||
0.002567075,
|
||||
-0.030825771,
|
||||
-0.0070259375,
|
||||
0.007989939,
|
||||
-0.027159046,
|
||||
-0.0714439,
|
||||
-0.020082822,
|
||||
-0.018486606,
|
||||
0.01108784,
|
||||
-0.012602704,
|
||||
-0.0012252157,
|
||||
0.06443626,
|
||||
0.036829114,
|
||||
0.04501229,
|
||||
0.0022744364,
|
||||
0.058829524,
|
||||
-0.008902569,
|
||||
-0.010049271,
|
||||
-0.0064951205,
|
||||
-0.014354489,
|
||||
-0.044668842,
|
||||
-0.025392724,
|
||||
0.015202658,
|
||||
0.020321742,
|
||||
-0.01176466,
|
||||
0.09413702,
|
||||
-0.0319812,
|
||||
0.03219725,
|
||||
-0.040439297,
|
||||
-0.019967683,
|
||||
0.0164714,
|
||||
0.019272799,
|
||||
0.02388655,
|
||||
0.017886775,
|
||||
-0.03603167,
|
||||
-0.023737542,
|
||||
-0.01898098,
|
||||
-0.04790894,
|
||||
-0.036694597,
|
||||
-0.02994124,
|
||||
0.034576166,
|
||||
-0.05921917,
|
||||
-0.022381892,
|
||||
-0.051536635,
|
||||
-0.05452498,
|
||||
0.053339027,
|
||||
0.019327087,
|
||||
0.012448543,
|
||||
-0.018923279,
|
||||
-0.0019192714,
|
||||
-0.01976354,
|
||||
0.032581042,
|
||||
-0.00695812,
|
||||
0.033768184,
|
||||
-0.028018538,
|
||||
-0.023666212,
|
||||
-0.017496848,
|
||||
0.023191998,
|
||||
-0.0502938,
|
||||
0.01670451,
|
||||
0.0058311033,
|
||||
0.012473936,
|
||||
0.023568941,
|
||||
-0.06854558,
|
||||
0.0073930174,
|
||||
0.07903637,
|
||||
-0.024922114,
|
||||
-0.026363779,
|
||||
-0.006970082,
|
||||
-0.007723444,
|
||||
0.074576765,
|
||||
-0.032073244,
|
||||
-0.013143484,
|
||||
-0.010095435,
|
||||
0.018318929,
|
||||
0.008086707,
|
||||
-0.01570327,
|
||||
-0.046567768,
|
||||
0.0038824868,
|
||||
-0.027711825
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "nomic-embed-text:latest",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 6,
|
||||
"total_tokens": 6
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,34 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "all-minilm:l6-v2",
|
||||
"created": 1759793684,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama3.2:3b-instruct-fp16",
|
||||
"created": 1759791776,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,25 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "all-minilm:l6-v2",
|
||||
"created": 1759785110,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,70 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "nomic-embed-text:latest",
|
||||
"created": 1755204798,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama-guard3:8b",
|
||||
"created": 1755125995,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "all-minilm:l6-v2",
|
||||
"created": 1753804403,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama3.2:3b-instruct-fp16",
|
||||
"created": 1752697170,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "all-minilm:latest",
|
||||
"created": 1752691712,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama-guard3:1b",
|
||||
"created": 1752267588,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,15 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,25 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama3.2:3b-instruct-fp16",
|
||||
"created": 1759791776,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,528 @@
|
|||
{
|
||||
"test_id": null,
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "https://api.fireworks.ai/inference/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-1-dev-fp8",
|
||||
"created": 1729532889,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": false,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-kontext-max",
|
||||
"created": 1750714611,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-kontext-pro",
|
||||
"created": 1750488264,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/sentientfoundation-serverless/models/dobby-mini-unhinged-plus-llama-3-1-8b",
|
||||
"created": 1748467427,
|
||||
"object": "model",
|
||||
"owned_by": "sentientfoundation-serverless",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/sentientfoundation/models/dobby-unhinged-llama-3-3-70b-new",
|
||||
"created": 1739563474,
|
||||
"object": "model",
|
||||
"owned_by": "sentientfoundation",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/gpt-oss-120b",
|
||||
"created": 1754345600,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-30b-a3b-thinking-2507",
|
||||
"created": 1753916446,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507",
|
||||
"created": 1753124424,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-235b-a22b-thinking-2507",
|
||||
"created": 1753455434,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3-0324",
|
||||
"created": 1742827220,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/kimi-k2-instruct",
|
||||
"created": 1752259096,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/gpt-oss-20b",
|
||||
"created": 1754345466,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/kimi-k2-instruct-0905",
|
||||
"created": 1757018994,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3",
|
||||
"created": 1735576668,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p3-70b-instruct",
|
||||
"created": 1733442103,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-235b-a22b",
|
||||
"created": 1745885249,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/glm-4p5-air",
|
||||
"created": 1754089426,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-r1",
|
||||
"created": 1737397673,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-r1-basic",
|
||||
"created": 1742306746,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3p1",
|
||||
"created": 1755758988,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-1-schnell-fp8",
|
||||
"created": 1729535376,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": false,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p1-405b-instruct",
|
||||
"created": 1721428386,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama4-scout-instruct-basic",
|
||||
"created": 1743878279,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": true,
|
||||
"context_length": 1048576
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-30b-a3b",
|
||||
"created": 1745878133,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p1-70b-instruct",
|
||||
"created": 1721287357,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-r1-0528",
|
||||
"created": 1748456377,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/mixtral-8x22b-instruct",
|
||||
"created": 1713375508,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 65536
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-embedding-8b",
|
||||
"created": 1755707090,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "EMBEDDING_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 40960
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-30b-a3b-instruct-2507",
|
||||
"created": 1753808388,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama4-maverick-instruct-basic",
|
||||
"created": 1743878495,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": true,
|
||||
"context_length": 1048576
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/tvergho-87e44d/models/debatecards-70b-ft-3epoch-dpo-v2",
|
||||
"created": 1743381121,
|
||||
"object": "model",
|
||||
"owned_by": "tvergho-87e44d",
|
||||
"kind": "HF_PEFT_ADDON",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen2p5-vl-32b-instruct",
|
||||
"created": 1743392739,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": false,
|
||||
"context_length": 128000
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-coder-30b-a3b-instruct",
|
||||
"created": 1754063588,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3p1-terminus",
|
||||
"created": 1758586241,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p1-8b-instruct",
|
||||
"created": 1721692808,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct",
|
||||
"created": 1753211090,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/glm-4p5",
|
||||
"created": 1753809636,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,527 @@
|
|||
{
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "https://api.fireworks.ai/inference/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-1-dev-fp8",
|
||||
"created": 1729532889,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": false,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-kontext-max",
|
||||
"created": 1750714611,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-kontext-pro",
|
||||
"created": 1750488264,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/sentientfoundation-serverless/models/dobby-mini-unhinged-plus-llama-3-1-8b",
|
||||
"created": 1748467427,
|
||||
"object": "model",
|
||||
"owned_by": "sentientfoundation-serverless",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/sentientfoundation/models/dobby-unhinged-llama-3-3-70b-new",
|
||||
"created": 1739563474,
|
||||
"object": "model",
|
||||
"owned_by": "sentientfoundation",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/gpt-oss-120b",
|
||||
"created": 1754345600,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-30b-a3b-thinking-2507",
|
||||
"created": 1753916446,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507",
|
||||
"created": 1753124424,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-235b-a22b-thinking-2507",
|
||||
"created": 1753455434,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-embedding-8b",
|
||||
"created": 1755707090,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 40960
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3-0324",
|
||||
"created": 1742827220,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/kimi-k2-instruct",
|
||||
"created": 1752259096,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/gpt-oss-20b",
|
||||
"created": 1754345466,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama4-maverick-instruct-basic",
|
||||
"created": 1743878495,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": true,
|
||||
"context_length": 1048576
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/kimi-k2-instruct-0905",
|
||||
"created": 1757018994,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3",
|
||||
"created": 1735576668,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p3-70b-instruct",
|
||||
"created": 1733442103,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-235b-a22b",
|
||||
"created": 1745885249,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/glm-4p5-air",
|
||||
"created": 1754089426,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-r1",
|
||||
"created": 1737397673,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p1-8b-instruct",
|
||||
"created": 1721692808,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-r1-basic",
|
||||
"created": 1742306746,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3p1",
|
||||
"created": 1755758988,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/flux-1-schnell-fp8",
|
||||
"created": 1729535376,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "FLUMINA_BASE_MODEL",
|
||||
"supports_chat": false,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p1-405b-instruct",
|
||||
"created": 1721428386,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama4-scout-instruct-basic",
|
||||
"created": 1743878279,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": true,
|
||||
"context_length": 1048576
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-30b-a3b",
|
||||
"created": 1745878133,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/llama-v3p1-70b-instruct",
|
||||
"created": 1721287357,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-r1-0528",
|
||||
"created": 1748456377,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/mixtral-8x22b-instruct",
|
||||
"created": 1713375508,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 65536
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-30b-a3b-instruct-2507",
|
||||
"created": 1753808388,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen2p5-vl-32b-instruct",
|
||||
"created": 1743392739,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": true,
|
||||
"supports_tools": false,
|
||||
"context_length": 128000
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-coder-30b-a3b-instruct",
|
||||
"created": 1754063588,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/tvergho-87e44d/models/debatecards-70b-ft-3epoch-dpo-v2",
|
||||
"created": 1743381121,
|
||||
"object": "model",
|
||||
"owned_by": "tvergho-87e44d",
|
||||
"kind": "HF_PEFT_ADDON",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": false
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/deepseek-v3p1-terminus",
|
||||
"created": 1758586241,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 163840
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct",
|
||||
"created": 1753211090,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 262144
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "accounts/fireworks/models/glm-4p5",
|
||||
"created": 1753809636,
|
||||
"object": "model",
|
||||
"owned_by": "fireworks",
|
||||
"kind": "HF_BASE_MODEL",
|
||||
"supports_chat": true,
|
||||
"supports_image_input": false,
|
||||
"supports_tools": true,
|
||||
"context_length": 131072
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,69 @@
|
|||
{
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/models",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/v1/models",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": [
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "nomic-embed-text:latest",
|
||||
"created": 1755204798,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama-guard3:8b",
|
||||
"created": 1755125995,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "all-minilm:l6-v2",
|
||||
"created": 1753804403,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama3.2:3b-instruct-fp16",
|
||||
"created": 1752697170,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "all-minilm:latest",
|
||||
"created": 1752691712,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
},
|
||||
{
|
||||
"__type__": "openai.types.model.Model",
|
||||
"__data__": {
|
||||
"id": "llama-guard3:1b",
|
||||
"created": 1752267588,
|
||||
"object": "model",
|
||||
"owned_by": "library"
|
||||
}
|
||||
}
|
||||
],
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_retrieve_contents[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is the content of test file 2"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.014871168,
|
||||
0.094365,
|
||||
-0.098275684,
|
||||
0.016189486,
|
||||
0.072296426,
|
||||
-0.039229725,
|
||||
0.007638039,
|
||||
0.035811495,
|
||||
-0.03784589,
|
||||
0.022591105,
|
||||
0.15810202,
|
||||
0.009195058,
|
||||
-0.029846655,
|
||||
-0.06448414,
|
||||
-0.01898075,
|
||||
-0.02023675,
|
||||
-0.07593923,
|
||||
-0.04666322,
|
||||
0.010769107,
|
||||
0.033283222,
|
||||
0.06951838,
|
||||
0.039086174,
|
||||
-0.009640043,
|
||||
-0.008601025,
|
||||
0.039979056,
|
||||
0.02799972,
|
||||
-0.06578151,
|
||||
0.08029443,
|
||||
0.0101568075,
|
||||
-0.07898879,
|
||||
0.048795786,
|
||||
0.057297125,
|
||||
0.025737243,
|
||||
0.03572965,
|
||||
0.11485981,
|
||||
0.030900626,
|
||||
0.118485495,
|
||||
-0.041167885,
|
||||
-0.019413618,
|
||||
-0.0009897926,
|
||||
0.03717747,
|
||||
-0.012367201,
|
||||
-0.0026639055,
|
||||
0.015703445,
|
||||
-0.0046827365,
|
||||
0.023138778,
|
||||
0.012855939,
|
||||
-0.029367425,
|
||||
0.00042996072,
|
||||
-0.003222942,
|
||||
-0.055509202,
|
||||
0.012830617,
|
||||
-0.06941755,
|
||||
-0.011024706,
|
||||
0.07149942,
|
||||
0.021040803,
|
||||
0.0409756,
|
||||
0.010087916,
|
||||
-0.015326204,
|
||||
0.06633094,
|
||||
0.024846299,
|
||||
0.030543685,
|
||||
-0.036063526,
|
||||
0.04786587,
|
||||
0.08074621,
|
||||
-0.051489003,
|
||||
-0.03944393,
|
||||
-0.025607359,
|
||||
-0.030061793,
|
||||
-0.119378455,
|
||||
-0.14597124,
|
||||
-0.0019379344,
|
||||
0.008393092,
|
||||
0.023913048,
|
||||
0.028285578,
|
||||
0.017838098,
|
||||
-0.10575887,
|
||||
0.008080291,
|
||||
0.06388723,
|
||||
-0.12506105,
|
||||
-0.02536782,
|
||||
-0.11007926,
|
||||
0.051198784,
|
||||
0.007446184,
|
||||
-0.030837545,
|
||||
0.09254253,
|
||||
0.05638562,
|
||||
-0.0155668175,
|
||||
-0.031867314,
|
||||
0.018337138,
|
||||
0.02442871,
|
||||
-0.042078987,
|
||||
0.0038125275,
|
||||
0.089955,
|
||||
-0.008119613,
|
||||
0.040103614,
|
||||
0.011012824,
|
||||
0.044628628,
|
||||
0.0791957,
|
||||
0.054247666,
|
||||
-0.027651828,
|
||||
-0.03190785,
|
||||
0.041443683,
|
||||
0.041629724,
|
||||
-0.077835254,
|
||||
-0.09937542,
|
||||
0.029904107,
|
||||
-0.05434366,
|
||||
0.07058962,
|
||||
-0.04535761,
|
||||
0.03365359,
|
||||
-0.061656676,
|
||||
-0.018105442,
|
||||
-0.07228336,
|
||||
0.035377987,
|
||||
-0.03161877,
|
||||
-0.020589713,
|
||||
0.058485094,
|
||||
-0.049225487,
|
||||
0.03934316,
|
||||
0.08550028,
|
||||
-0.029991213,
|
||||
-0.05576064,
|
||||
-0.029334918,
|
||||
-0.053031918,
|
||||
-0.061839186,
|
||||
0.08176057,
|
||||
-3.3282106e-33,
|
||||
0.00018265574,
|
||||
-0.09808404,
|
||||
-0.00554673,
|
||||
0.13180184,
|
||||
0.026467713,
|
||||
-0.03976283,
|
||||
0.010410568,
|
||||
0.022475285,
|
||||
-0.07190717,
|
||||
0.005138454,
|
||||
-0.021325583,
|
||||
-0.1046733,
|
||||
0.0020021838,
|
||||
0.023773609,
|
||||
-0.057499945,
|
||||
-0.011727483,
|
||||
-0.020912478,
|
||||
0.026353713,
|
||||
0.01779019,
|
||||
-0.0148312645,
|
||||
0.064687304,
|
||||
0.045060385,
|
||||
-0.029312065,
|
||||
-0.08633001,
|
||||
-0.026792597,
|
||||
0.014552106,
|
||||
0.004505434,
|
||||
-0.06774755,
|
||||
0.034052122,
|
||||
0.013713737,
|
||||
-0.0075813113,
|
||||
-0.059718475,
|
||||
-0.016189422,
|
||||
0.044314116,
|
||||
0.026844766,
|
||||
0.026430624,
|
||||
0.024091395,
|
||||
-0.0032406747,
|
||||
-0.075288124,
|
||||
0.032822173,
|
||||
0.027104331,
|
||||
-0.026295068,
|
||||
0.04316082,
|
||||
-0.010091815,
|
||||
0.034184698,
|
||||
-0.08266358,
|
||||
-0.020962045,
|
||||
-0.00719584,
|
||||
0.068549044,
|
||||
0.005033586,
|
||||
0.0017975906,
|
||||
0.06465498,
|
||||
0.05990613,
|
||||
-0.012483792,
|
||||
0.024451919,
|
||||
0.021659598,
|
||||
-0.0046074707,
|
||||
-0.004559902,
|
||||
0.002713282,
|
||||
0.062373567,
|
||||
0.0035651235,
|
||||
0.06017224,
|
||||
-0.062707886,
|
||||
0.039937016,
|
||||
-0.0064443815,
|
||||
-0.041358124,
|
||||
-0.045459975,
|
||||
-0.1090475,
|
||||
0.08058783,
|
||||
0.055110224,
|
||||
-0.05126053,
|
||||
-0.05976516,
|
||||
0.037940193,
|
||||
0.015456569,
|
||||
-0.024956519,
|
||||
-0.037877902,
|
||||
-0.006799,
|
||||
0.031685203,
|
||||
-0.036858797,
|
||||
-0.055584695,
|
||||
-0.048513155,
|
||||
-0.07101657,
|
||||
-0.041681714,
|
||||
-0.04429727,
|
||||
-0.09584418,
|
||||
-0.060873836,
|
||||
0.008867621,
|
||||
-0.106438614,
|
||||
0.040050562,
|
||||
-0.084729105,
|
||||
0.018111277,
|
||||
0.010153493,
|
||||
-0.08883196,
|
||||
-0.063969284,
|
||||
0.08611972,
|
||||
1.4074298e-33,
|
||||
0.03433739,
|
||||
0.037653737,
|
||||
-0.05348675,
|
||||
0.0015385789,
|
||||
0.026684077,
|
||||
0.026603375,
|
||||
0.07006387,
|
||||
-0.034265522,
|
||||
-0.018221779,
|
||||
0.10960259,
|
||||
0.013464475,
|
||||
-0.008325532,
|
||||
0.019438146,
|
||||
-0.039553005,
|
||||
0.03469477,
|
||||
-0.0123773115,
|
||||
-0.013288484,
|
||||
-0.048081715,
|
||||
-0.019539693,
|
||||
-0.0033996427,
|
||||
-0.024453517,
|
||||
0.061505664,
|
||||
0.119236834,
|
||||
0.026294904,
|
||||
-0.01607055,
|
||||
-0.011499089,
|
||||
0.04267117,
|
||||
0.0295908,
|
||||
0.022084564,
|
||||
0.007893738,
|
||||
0.052055445,
|
||||
0.05781507,
|
||||
-0.13408813,
|
||||
0.01778491,
|
||||
0.021400984,
|
||||
-0.12113228,
|
||||
0.10535695,
|
||||
-0.07358604,
|
||||
-0.013651957,
|
||||
0.04049295,
|
||||
0.054150987,
|
||||
0.0987462,
|
||||
0.0110208625,
|
||||
0.040327504,
|
||||
0.034936633,
|
||||
0.10400846,
|
||||
0.12958324,
|
||||
-0.024531014,
|
||||
0.002284699,
|
||||
-0.044239815,
|
||||
0.049778443,
|
||||
-0.055788964,
|
||||
0.015235888,
|
||||
0.0034493478,
|
||||
-0.02607555,
|
||||
0.060282644,
|
||||
-0.028004775,
|
||||
0.040875163,
|
||||
-0.023749253,
|
||||
0.002289086,
|
||||
0.04982698,
|
||||
0.046928305,
|
||||
-0.064160004,
|
||||
0.013701618,
|
||||
0.015511878,
|
||||
-0.054725982,
|
||||
-0.0459802,
|
||||
0.03258067,
|
||||
0.027034523,
|
||||
0.01643672,
|
||||
-0.041782584,
|
||||
-0.03698569,
|
||||
-0.023043923,
|
||||
-0.07073365,
|
||||
0.028486207,
|
||||
0.0017764921,
|
||||
-0.03352676,
|
||||
-0.009977863,
|
||||
0.024488676,
|
||||
-0.01789395,
|
||||
0.029737154,
|
||||
-0.026266927,
|
||||
-0.03567072,
|
||||
0.07469971,
|
||||
0.028393274,
|
||||
-0.029625034,
|
||||
-0.01053128,
|
||||
0.09147493,
|
||||
-0.018718474,
|
||||
0.0012933073,
|
||||
-0.021214467,
|
||||
0.07475739,
|
||||
-0.007773536,
|
||||
0.048597455,
|
||||
0.005216022,
|
||||
-1.6914717e-08,
|
||||
-0.05724563,
|
||||
-0.0938908,
|
||||
-0.034359876,
|
||||
-0.037500683,
|
||||
-0.020235153,
|
||||
0.06142227,
|
||||
-0.042273093,
|
||||
-0.008759724,
|
||||
-0.009908796,
|
||||
0.016232042,
|
||||
-0.014239323,
|
||||
0.024709346,
|
||||
-0.030538557,
|
||||
-0.05391127,
|
||||
-0.051778477,
|
||||
0.01277344,
|
||||
0.0036140021,
|
||||
-0.012569925,
|
||||
-0.025041323,
|
||||
-0.0203936,
|
||||
0.025865255,
|
||||
0.010908398,
|
||||
0.027834684,
|
||||
0.009661084,
|
||||
-0.006598172,
|
||||
0.07860872,
|
||||
0.054516125,
|
||||
0.042956624,
|
||||
-0.06275145,
|
||||
-0.025701547,
|
||||
0.08085865,
|
||||
0.030041302,
|
||||
0.02248997,
|
||||
-0.0840195,
|
||||
0.00029938898,
|
||||
0.10966559,
|
||||
0.118907265,
|
||||
0.063014604,
|
||||
0.037847042,
|
||||
0.032069027,
|
||||
-0.05345487,
|
||||
-0.022730324,
|
||||
0.0071888734,
|
||||
0.037573762,
|
||||
-0.020178014,
|
||||
-0.090167634,
|
||||
-0.07191704,
|
||||
-0.02604166,
|
||||
-0.043885063,
|
||||
-0.14087014,
|
||||
-0.017230472,
|
||||
-0.012063355,
|
||||
-0.046736836,
|
||||
0.039048597,
|
||||
-0.060394738,
|
||||
0.022166032,
|
||||
0.025670663,
|
||||
0.022949725,
|
||||
-0.06707243,
|
||||
-0.014654702,
|
||||
0.057985142,
|
||||
0.10511708,
|
||||
0.05698323,
|
||||
-0.017205814
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 8,
|
||||
"total_tokens": 8
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_retrieve_contents[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is the content of test file 1"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.029406669,
|
||||
0.08920982,
|
||||
-0.11326726,
|
||||
0.0065823817,
|
||||
0.07725067,
|
||||
-0.036890104,
|
||||
0.030436223,
|
||||
0.041454185,
|
||||
-0.049156666,
|
||||
0.018258564,
|
||||
0.14662577,
|
||||
0.01744915,
|
||||
-0.012837422,
|
||||
-0.06889876,
|
||||
-0.039401636,
|
||||
-0.038800705,
|
||||
-0.08963421,
|
||||
-0.059656583,
|
||||
0.001375945,
|
||||
0.045138627,
|
||||
0.042796962,
|
||||
0.053700265,
|
||||
-0.035706885,
|
||||
0.010138017,
|
||||
0.060920056,
|
||||
0.017344126,
|
||||
-0.05633907,
|
||||
0.063370295,
|
||||
0.0021257724,
|
||||
-0.083796844,
|
||||
0.050487563,
|
||||
0.047987595,
|
||||
0.069071226,
|
||||
0.049588464,
|
||||
0.117036626,
|
||||
0.05339311,
|
||||
0.10129953,
|
||||
-0.048230153,
|
||||
-0.014987975,
|
||||
0.0250915,
|
||||
0.031392053,
|
||||
-0.008863942,
|
||||
0.0073650074,
|
||||
-0.0009767569,
|
||||
-0.016403567,
|
||||
0.015523393,
|
||||
-0.010998956,
|
||||
-0.014870063,
|
||||
0.0061682137,
|
||||
-0.0017961137,
|
||||
-0.022682818,
|
||||
0.018210242,
|
||||
-0.07757007,
|
||||
-0.0015845516,
|
||||
0.069547005,
|
||||
0.000419109,
|
||||
0.038414054,
|
||||
0.005823485,
|
||||
-0.028931383,
|
||||
0.07009549,
|
||||
-0.0018009909,
|
||||
0.033516172,
|
||||
-0.014593847,
|
||||
0.03922457,
|
||||
0.08240545,
|
||||
-0.050596908,
|
||||
-0.039732855,
|
||||
-0.024425076,
|
||||
-0.015055329,
|
||||
-0.11705068,
|
||||
-0.15979129,
|
||||
-0.008256823,
|
||||
-0.0100719705,
|
||||
0.03266482,
|
||||
0.0029998205,
|
||||
0.0316428,
|
||||
-0.094554916,
|
||||
0.017661797,
|
||||
0.058996264,
|
||||
-0.119718134,
|
||||
-0.027414676,
|
||||
-0.09155906,
|
||||
0.040038,
|
||||
0.01091849,
|
||||
-0.029446004,
|
||||
0.10225186,
|
||||
0.06583262,
|
||||
-0.003439552,
|
||||
-0.009694834,
|
||||
0.016906522,
|
||||
0.023685955,
|
||||
-0.032616187,
|
||||
-0.010238839,
|
||||
0.07891618,
|
||||
-0.007330681,
|
||||
0.05238444,
|
||||
0.00943625,
|
||||
0.042121,
|
||||
0.08491511,
|
||||
0.049208272,
|
||||
-0.01868227,
|
||||
-0.013585418,
|
||||
0.06727199,
|
||||
0.084571496,
|
||||
-0.103213035,
|
||||
-0.08387524,
|
||||
0.03641842,
|
||||
-0.047227863,
|
||||
0.057315867,
|
||||
-0.04463932,
|
||||
0.006783099,
|
||||
-0.08934107,
|
||||
-0.015040418,
|
||||
-0.08107057,
|
||||
0.013285569,
|
||||
-0.060907867,
|
||||
-0.042128306,
|
||||
0.057306163,
|
||||
-0.058711898,
|
||||
0.04628304,
|
||||
0.070194095,
|
||||
-0.041729517,
|
||||
-0.0338408,
|
||||
-0.012369257,
|
||||
-0.044708908,
|
||||
-0.059450094,
|
||||
0.08251312,
|
||||
-3.443368e-33,
|
||||
0.0121309515,
|
||||
-0.11084454,
|
||||
-0.020510655,
|
||||
0.10916455,
|
||||
0.033683147,
|
||||
-0.02845083,
|
||||
0.024345158,
|
||||
0.034192592,
|
||||
-0.08367815,
|
||||
0.0064610844,
|
||||
-0.00912456,
|
||||
-0.0663567,
|
||||
-0.0028754657,
|
||||
0.008272698,
|
||||
-0.09166764,
|
||||
0.0089771375,
|
||||
-0.03963948,
|
||||
0.019947624,
|
||||
-0.01321528,
|
||||
-0.019034218,
|
||||
0.051933073,
|
||||
0.028107261,
|
||||
-0.039153125,
|
||||
-0.080395184,
|
||||
-0.050503474,
|
||||
0.02060341,
|
||||
-0.012718284,
|
||||
-0.046732575,
|
||||
0.017907938,
|
||||
-0.0028334607,
|
||||
-0.011695137,
|
||||
-0.05667005,
|
||||
-0.043894444,
|
||||
0.034919597,
|
||||
0.022352098,
|
||||
0.046777196,
|
||||
0.045085873,
|
||||
-0.008840106,
|
||||
-0.06373453,
|
||||
0.036720857,
|
||||
0.012829601,
|
||||
-0.035169926,
|
||||
0.046209145,
|
||||
-0.014361767,
|
||||
0.03706697,
|
||||
-0.056797564,
|
||||
-0.06310496,
|
||||
0.010818958,
|
||||
0.047810175,
|
||||
0.0029118094,
|
||||
-0.003235893,
|
||||
0.061511047,
|
||||
0.072056666,
|
||||
-0.03286638,
|
||||
0.005070082,
|
||||
0.021947902,
|
||||
-0.017779002,
|
||||
-0.022738373,
|
||||
-0.021926457,
|
||||
0.047074158,
|
||||
0.010847615,
|
||||
0.05539702,
|
||||
-0.07119971,
|
||||
0.033833236,
|
||||
0.012342855,
|
||||
-0.047586687,
|
||||
-0.026776271,
|
||||
-0.09885727,
|
||||
0.10053448,
|
||||
0.036877092,
|
||||
-0.07049897,
|
||||
-0.059692938,
|
||||
0.016129492,
|
||||
-0.0016443401,
|
||||
-0.026804024,
|
||||
-0.013527272,
|
||||
-0.015385511,
|
||||
0.055627547,
|
||||
-0.060485132,
|
||||
-0.055540122,
|
||||
-0.04329072,
|
||||
-0.07097361,
|
||||
-0.04857043,
|
||||
-0.03726256,
|
||||
-0.09059366,
|
||||
-0.036855534,
|
||||
0.024561211,
|
||||
-0.10113953,
|
||||
0.056738112,
|
||||
-0.10995085,
|
||||
0.042282794,
|
||||
0.014222368,
|
||||
-0.07067843,
|
||||
-0.05902307,
|
||||
0.06426122,
|
||||
1.6036318e-33,
|
||||
0.037851896,
|
||||
0.032911286,
|
||||
-0.04029648,
|
||||
-0.00049357174,
|
||||
0.028011942,
|
||||
0.048672136,
|
||||
0.07279598,
|
||||
-0.027471887,
|
||||
-0.02847654,
|
||||
0.114492,
|
||||
0.001777095,
|
||||
-0.009519909,
|
||||
0.0025862327,
|
||||
-0.056408145,
|
||||
0.023462169,
|
||||
-0.006209674,
|
||||
-0.010567065,
|
||||
-0.05877587,
|
||||
-0.032393616,
|
||||
0.011836781,
|
||||
-0.038905054,
|
||||
0.05516299,
|
||||
0.09564333,
|
||||
0.028543225,
|
||||
-0.023832332,
|
||||
-0.0015711841,
|
||||
0.047049087,
|
||||
0.03128219,
|
||||
0.02811091,
|
||||
0.007177092,
|
||||
0.055283513,
|
||||
0.06574452,
|
||||
-0.1020208,
|
||||
0.021213628,
|
||||
0.020237882,
|
||||
-0.10449357,
|
||||
0.09608935,
|
||||
-0.06253181,
|
||||
0.015293753,
|
||||
0.042053986,
|
||||
0.06105009,
|
||||
0.0909162,
|
||||
0.018404186,
|
||||
0.031023262,
|
||||
0.03562763,
|
||||
0.112073965,
|
||||
0.10124763,
|
||||
-0.007683015,
|
||||
0.013140281,
|
||||
-0.042280227,
|
||||
0.051135287,
|
||||
-0.02950743,
|
||||
0.027794402,
|
||||
-0.010734668,
|
||||
-0.011067552,
|
||||
0.058104575,
|
||||
-0.009284788,
|
||||
0.056184508,
|
||||
-0.040822964,
|
||||
0.010282754,
|
||||
0.0374409,
|
||||
0.054198533,
|
||||
-0.061418086,
|
||||
0.030569963,
|
||||
0.0023648597,
|
||||
-0.054184474,
|
||||
-0.020570045,
|
||||
0.012422129,
|
||||
0.025696559,
|
||||
-0.007607385,
|
||||
-0.026194826,
|
||||
-0.024159024,
|
||||
0.0012979766,
|
||||
-0.07461716,
|
||||
0.051458035,
|
||||
-0.004183808,
|
||||
-0.040804464,
|
||||
-0.023975441,
|
||||
0.009455526,
|
||||
-0.0018798193,
|
||||
0.03668693,
|
||||
-0.019319497,
|
||||
-0.06195781,
|
||||
0.06456675,
|
||||
0.040328216,
|
||||
-0.010790134,
|
||||
0.013190221,
|
||||
0.09067539,
|
||||
-0.0051480443,
|
||||
0.013312647,
|
||||
-0.029548675,
|
||||
0.07769003,
|
||||
0.0027328292,
|
||||
0.04533781,
|
||||
-0.0017606319,
|
||||
-1.661594e-08,
|
||||
-0.040610366,
|
||||
-0.09883059,
|
||||
-0.05522113,
|
||||
-0.02916469,
|
||||
-0.019305382,
|
||||
0.088138185,
|
||||
-0.038325552,
|
||||
-0.03327639,
|
||||
-0.012629364,
|
||||
0.006948921,
|
||||
0.010438818,
|
||||
0.026771523,
|
||||
-0.040855426,
|
||||
-0.03958403,
|
||||
-0.051137064,
|
||||
-0.016159322,
|
||||
-0.020525131,
|
||||
-0.023726366,
|
||||
-0.013322245,
|
||||
-0.008097836,
|
||||
0.028000915,
|
||||
0.02806969,
|
||||
0.015645925,
|
||||
-0.0043166955,
|
||||
0.0054488196,
|
||||
0.06720413,
|
||||
0.068473674,
|
||||
0.07172716,
|
||||
-0.06339439,
|
||||
-0.02540609,
|
||||
0.08468492,
|
||||
0.041936778,
|
||||
0.021067144,
|
||||
-0.07596481,
|
||||
0.017143335,
|
||||
0.1260291,
|
||||
0.121315174,
|
||||
0.08431059,
|
||||
0.040587336,
|
||||
0.036687315,
|
||||
-0.04717,
|
||||
-0.022659328,
|
||||
-0.006820436,
|
||||
0.005210712,
|
||||
-0.033785924,
|
||||
-0.08449115,
|
||||
-0.0844501,
|
||||
-0.03192747,
|
||||
-0.036649443,
|
||||
-0.13791409,
|
||||
-0.036417518,
|
||||
-0.00080547476,
|
||||
-0.047578912,
|
||||
0.038795993,
|
||||
-0.06757743,
|
||||
0.016941966,
|
||||
0.036312684,
|
||||
0.0125779435,
|
||||
-0.058240637,
|
||||
0.004471269,
|
||||
0.03226526,
|
||||
0.09821741,
|
||||
0.053010236,
|
||||
-0.016268
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 8,
|
||||
"total_tokens": 8
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_create_and_retrieve[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is batch test file 1"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.009745733,
|
||||
0.03363038,
|
||||
-0.10852256,
|
||||
0.026609829,
|
||||
-0.0060599064,
|
||||
-0.020473678,
|
||||
0.0692486,
|
||||
0.032276765,
|
||||
-0.11532835,
|
||||
-0.0005207133,
|
||||
0.11814916,
|
||||
0.0119809555,
|
||||
0.03685765,
|
||||
-0.10744223,
|
||||
-0.046515625,
|
||||
0.0015449532,
|
||||
-0.06319664,
|
||||
-0.04640812,
|
||||
-0.037318822,
|
||||
-0.025718328,
|
||||
-0.00026058854,
|
||||
-0.011890766,
|
||||
-0.050925612,
|
||||
0.014111713,
|
||||
0.029467698,
|
||||
0.006379121,
|
||||
-0.012013293,
|
||||
-0.0024293982,
|
||||
-0.044318773,
|
||||
-0.08100101,
|
||||
0.02009568,
|
||||
0.055713937,
|
||||
0.078816675,
|
||||
0.054973654,
|
||||
0.20367871,
|
||||
-0.004309458,
|
||||
0.03877001,
|
||||
0.03825522,
|
||||
-0.002538199,
|
||||
-0.0007973801,
|
||||
0.044761047,
|
||||
-0.054529082,
|
||||
-0.008856888,
|
||||
-0.04078078,
|
||||
0.011367262,
|
||||
-0.022404457,
|
||||
-0.06209053,
|
||||
0.02558725,
|
||||
-0.0034454092,
|
||||
-0.03743928,
|
||||
-0.062026348,
|
||||
-0.030812219,
|
||||
-0.034592565,
|
||||
-0.014926672,
|
||||
0.018588377,
|
||||
0.013435887,
|
||||
0.08169151,
|
||||
0.053658403,
|
||||
-0.03557856,
|
||||
0.033325985,
|
||||
-0.01637577,
|
||||
-0.0222152,
|
||||
-0.039247517,
|
||||
0.00094368146,
|
||||
0.10228945,
|
||||
-0.04305617,
|
||||
-0.052200828,
|
||||
-0.02007385,
|
||||
0.054805383,
|
||||
-0.08231377,
|
||||
-0.14736547,
|
||||
0.048954617,
|
||||
-0.0212168,
|
||||
0.02872658,
|
||||
-0.0671409,
|
||||
0.021436114,
|
||||
-0.023599947,
|
||||
0.03677982,
|
||||
0.010577411,
|
||||
-0.0966004,
|
||||
-0.06367233,
|
||||
-0.10277648,
|
||||
0.0273993,
|
||||
-0.06292906,
|
||||
-0.046344172,
|
||||
0.039919835,
|
||||
0.02682899,
|
||||
0.025460077,
|
||||
-0.013083559,
|
||||
-0.002667712,
|
||||
-0.016529463,
|
||||
0.012605053,
|
||||
-0.0064383023,
|
||||
0.015841383,
|
||||
-0.01710707,
|
||||
0.12320292,
|
||||
-0.0077660284,
|
||||
0.05845043,
|
||||
0.07362552,
|
||||
0.038426086,
|
||||
0.004742023,
|
||||
-0.0155985365,
|
||||
0.01418979,
|
||||
0.07865995,
|
||||
-0.026352523,
|
||||
-0.037174653,
|
||||
0.06787817,
|
||||
-0.060126718,
|
||||
0.06111402,
|
||||
-0.034931272,
|
||||
-0.009446326,
|
||||
-0.006150886,
|
||||
0.02892313,
|
||||
-0.09361577,
|
||||
0.0335364,
|
||||
-0.09088912,
|
||||
0.009241144,
|
||||
0.07092964,
|
||||
-0.08954648,
|
||||
0.04494549,
|
||||
0.040462427,
|
||||
-0.04167353,
|
||||
0.0076030386,
|
||||
-0.0066417656,
|
||||
-0.07275736,
|
||||
-0.043690544,
|
||||
0.07685007,
|
||||
-1.0508795e-33,
|
||||
-0.019583685,
|
||||
-0.13087204,
|
||||
-0.03574564,
|
||||
0.070223756,
|
||||
0.08133056,
|
||||
-0.009436003,
|
||||
0.046778366,
|
||||
0.03478148,
|
||||
-0.09441185,
|
||||
-0.040857755,
|
||||
-0.02127058,
|
||||
-0.106959894,
|
||||
0.024023255,
|
||||
0.022780996,
|
||||
-0.09042505,
|
||||
-0.035755932,
|
||||
0.011359196,
|
||||
0.050059184,
|
||||
0.0050815986,
|
||||
-0.07676938,
|
||||
0.05453651,
|
||||
0.04191775,
|
||||
-0.009206564,
|
||||
-0.022437057,
|
||||
-0.04617258,
|
||||
-0.038608693,
|
||||
-0.00036489012,
|
||||
-0.025092375,
|
||||
0.039146807,
|
||||
-0.0072839926,
|
||||
0.03675482,
|
||||
-0.011301064,
|
||||
-0.08863303,
|
||||
0.059421506,
|
||||
0.015851071,
|
||||
0.033407707,
|
||||
0.056883834,
|
||||
-0.01203776,
|
||||
0.027333334,
|
||||
-0.009560535,
|
||||
-0.05030555,
|
||||
-0.009787559,
|
||||
0.023205005,
|
||||
-0.007937716,
|
||||
0.003991047,
|
||||
-0.036422852,
|
||||
-0.06979188,
|
||||
0.046075627,
|
||||
0.056377746,
|
||||
0.0071927872,
|
||||
-0.00020658698,
|
||||
0.017678235,
|
||||
0.023745935,
|
||||
-0.0031295705,
|
||||
0.016370842,
|
||||
0.027585855,
|
||||
-0.03440131,
|
||||
-0.05594279,
|
||||
0.036442764,
|
||||
0.03577988,
|
||||
-0.005324585,
|
||||
0.015240975,
|
||||
-0.09071462,
|
||||
0.072764605,
|
||||
0.02343818,
|
||||
-0.093097225,
|
||||
0.05842133,
|
||||
-0.061913762,
|
||||
0.045556016,
|
||||
0.07639311,
|
||||
-0.035199754,
|
||||
-0.009256856,
|
||||
0.038682748,
|
||||
-0.040795818,
|
||||
0.017686425,
|
||||
-0.025513103,
|
||||
0.06860537,
|
||||
0.085520275,
|
||||
-0.1023457,
|
||||
-0.0036474275,
|
||||
-0.014826131,
|
||||
-0.05045756,
|
||||
-0.09065474,
|
||||
-0.076476775,
|
||||
-0.008538021,
|
||||
-0.04111943,
|
||||
-0.035473913,
|
||||
-0.061549038,
|
||||
0.114327826,
|
||||
-0.09601482,
|
||||
0.022990143,
|
||||
0.0022396755,
|
||||
-0.023026146,
|
||||
-0.028128328,
|
||||
0.07969127,
|
||||
-4.1765383e-34,
|
||||
0.07866384,
|
||||
0.11484068,
|
||||
0.016687382,
|
||||
0.009315677,
|
||||
0.01664128,
|
||||
0.024303248,
|
||||
0.046507504,
|
||||
-0.043804675,
|
||||
-0.09136995,
|
||||
0.106353745,
|
||||
-0.06948852,
|
||||
0.018747667,
|
||||
0.0053492193,
|
||||
-0.033229355,
|
||||
0.042339083,
|
||||
-0.0017468681,
|
||||
0.05323157,
|
||||
0.0058223205,
|
||||
-0.05331342,
|
||||
0.016506517,
|
||||
-0.02325185,
|
||||
0.097519755,
|
||||
-0.0045558517,
|
||||
0.08866843,
|
||||
-0.028221445,
|
||||
-0.012007969,
|
||||
-0.009742725,
|
||||
0.061458003,
|
||||
0.01574456,
|
||||
-0.00039456616,
|
||||
0.02444834,
|
||||
0.065891184,
|
||||
-0.054779086,
|
||||
0.04863689,
|
||||
0.043890025,
|
||||
-0.062467597,
|
||||
0.07615393,
|
||||
0.0067509366,
|
||||
0.019150084,
|
||||
0.06994535,
|
||||
0.027900916,
|
||||
0.08902746,
|
||||
-0.027433047,
|
||||
0.031390887,
|
||||
0.02271287,
|
||||
0.08119532,
|
||||
0.06855678,
|
||||
0.0023552915,
|
||||
-0.06764184,
|
||||
0.00704173,
|
||||
-0.034521427,
|
||||
-0.053785548,
|
||||
-0.03075216,
|
||||
0.007947864,
|
||||
-0.025317406,
|
||||
-0.040664013,
|
||||
0.036144093,
|
||||
0.017730465,
|
||||
-0.040179063,
|
||||
0.013665757,
|
||||
0.004815376,
|
||||
0.009095556,
|
||||
0.0072483593,
|
||||
0.012753351,
|
||||
-0.047865536,
|
||||
-0.046072423,
|
||||
-0.014048283,
|
||||
0.031082962,
|
||||
-0.034945205,
|
||||
-0.023550391,
|
||||
0.033062257,
|
||||
-0.022966444,
|
||||
0.007744228,
|
||||
0.015939556,
|
||||
-0.0012224894,
|
||||
0.0010534802,
|
||||
-0.015109,
|
||||
-0.021597888,
|
||||
-0.029862719,
|
||||
0.03983828,
|
||||
0.062536344,
|
||||
0.0106168175,
|
||||
-0.027220478,
|
||||
0.02410377,
|
||||
-0.0023566757,
|
||||
0.085310005,
|
||||
0.04843323,
|
||||
0.090823516,
|
||||
0.005126319,
|
||||
0.020297319,
|
||||
-0.01739127,
|
||||
0.047677357,
|
||||
0.11080086,
|
||||
0.030030197,
|
||||
0.029773563,
|
||||
-1.5454503e-08,
|
||||
-0.03580758,
|
||||
-0.12177604,
|
||||
0.019753791,
|
||||
0.05854353,
|
||||
-0.01590761,
|
||||
0.085781366,
|
||||
-0.09558486,
|
||||
-0.0016744126,
|
||||
0.00773199,
|
||||
-0.04790156,
|
||||
0.01175936,
|
||||
0.006536077,
|
||||
-0.032027386,
|
||||
0.0031026274,
|
||||
-0.07580574,
|
||||
-0.039700802,
|
||||
-0.00170645,
|
||||
-0.070955865,
|
||||
0.043680355,
|
||||
0.029966798,
|
||||
0.0039943648,
|
||||
0.031923376,
|
||||
0.08119928,
|
||||
0.038820695,
|
||||
0.013302812,
|
||||
0.041675337,
|
||||
0.044349737,
|
||||
0.060403902,
|
||||
-0.1058191,
|
||||
-0.05287386,
|
||||
0.050275758,
|
||||
0.039101604,
|
||||
0.0599918,
|
||||
-0.025067834,
|
||||
-0.019554066,
|
||||
0.06748813,
|
||||
0.12508559,
|
||||
0.059007537,
|
||||
-0.019899847,
|
||||
-0.030194808,
|
||||
-0.046559453,
|
||||
0.034567222,
|
||||
-0.021644907,
|
||||
-0.03327634,
|
||||
-0.0075667608,
|
||||
-0.100658834,
|
||||
-0.0639619,
|
||||
-0.055270903,
|
||||
-0.0111757815,
|
||||
-0.11671873,
|
||||
-0.07208087,
|
||||
0.023208033,
|
||||
0.027215267,
|
||||
0.063635156,
|
||||
-0.05858023,
|
||||
0.020345282,
|
||||
0.018325811,
|
||||
-0.0036095325,
|
||||
0.006916675,
|
||||
0.06541716,
|
||||
0.009575581,
|
||||
0.046839867,
|
||||
0.0070611075,
|
||||
-0.09470841
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 6,
|
||||
"total_tokens": 6
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,39 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_retrieve_contents[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://localhost:11434/api/ps",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/api/ps",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "ollama._types.ProcessResponse",
|
||||
"__data__": {
|
||||
"models": [
|
||||
{
|
||||
"model": "all-minilm:l6-v2",
|
||||
"name": "all-minilm:l6-v2",
|
||||
"digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef",
|
||||
"expires_at": "2025-10-06T16:41:45.231544-07:00",
|
||||
"size": 590204928,
|
||||
"size_vram": 590204928,
|
||||
"details": {
|
||||
"parent_model": "",
|
||||
"format": "gguf",
|
||||
"family": "bert",
|
||||
"families": [
|
||||
"bert"
|
||||
],
|
||||
"parameter_size": "23M",
|
||||
"quantization_level": "F16"
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_create_and_retrieve[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is batch test file 0"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
0.020637129,
|
||||
0.048547756,
|
||||
-0.12516363,
|
||||
0.01991118,
|
||||
-0.006535745,
|
||||
-0.017178575,
|
||||
0.027727997,
|
||||
0.032170568,
|
||||
-0.07302972,
|
||||
0.008939002,
|
||||
0.11493648,
|
||||
0.0058907545,
|
||||
0.0058539375,
|
||||
-0.077171296,
|
||||
-0.06883132,
|
||||
0.0039748913,
|
||||
-0.046849657,
|
||||
-0.072902456,
|
||||
-0.010890429,
|
||||
-0.0019311906,
|
||||
-0.011614798,
|
||||
0.003689495,
|
||||
-0.03695609,
|
||||
-0.009029024,
|
||||
0.017461002,
|
||||
-0.004713484,
|
||||
-0.010254731,
|
||||
-0.026636763,
|
||||
-0.026125714,
|
||||
-0.046913657,
|
||||
0.017024228,
|
||||
0.0713477,
|
||||
0.07881179,
|
||||
0.03789051,
|
||||
0.21716279,
|
||||
-0.0077837943,
|
||||
0.04686894,
|
||||
0.020414647,
|
||||
7.314368e-05,
|
||||
0.0103133675,
|
||||
0.059848394,
|
||||
-0.04321678,
|
||||
-0.011937493,
|
||||
-0.021149047,
|
||||
0.021315353,
|
||||
-0.00072822213,
|
||||
-0.046116166,
|
||||
-0.0046820445,
|
||||
0.016943695,
|
||||
-0.03249135,
|
||||
-0.055184096,
|
||||
4.1543382e-05,
|
||||
-0.034172166,
|
||||
-0.023247559,
|
||||
0.020267941,
|
||||
0.012827845,
|
||||
0.065036125,
|
||||
0.07180022,
|
||||
-0.013490698,
|
||||
0.06376413,
|
||||
-0.017730094,
|
||||
-0.01806601,
|
||||
-0.034191083,
|
||||
0.008955718,
|
||||
0.098446764,
|
||||
-0.0061265854,
|
||||
-0.06815829,
|
||||
-0.039525956,
|
||||
0.060588058,
|
||||
-0.094874755,
|
||||
-0.11774928,
|
||||
0.019538416,
|
||||
-0.014697532,
|
||||
0.04773719,
|
||||
-0.061298393,
|
||||
0.030337377,
|
||||
-0.0022184649,
|
||||
0.019007793,
|
||||
0.024370821,
|
||||
-0.07063359,
|
||||
-0.07582954,
|
||||
-0.10816809,
|
||||
0.031845964,
|
||||
-0.057830192,
|
||||
-0.04169559,
|
||||
0.0752806,
|
||||
0.019289386,
|
||||
0.028845867,
|
||||
0.0077010663,
|
||||
0.013930818,
|
||||
-0.067987345,
|
||||
0.012679873,
|
||||
-0.07907268,
|
||||
0.0143718915,
|
||||
-0.021433424,
|
||||
0.11880779,
|
||||
-0.016258432,
|
||||
0.07099568,
|
||||
0.035778854,
|
||||
0.028776454,
|
||||
0.013304291,
|
||||
-0.05192297,
|
||||
0.026758345,
|
||||
0.10282426,
|
||||
-0.003306269,
|
||||
-0.03239622,
|
||||
0.083044045,
|
||||
-0.0412691,
|
||||
0.043435257,
|
||||
-0.043423533,
|
||||
-0.013239603,
|
||||
-0.0029038454,
|
||||
0.038365215,
|
||||
-0.10401672,
|
||||
0.012744224,
|
||||
-0.122984126,
|
||||
-0.008942817,
|
||||
0.06162198,
|
||||
-0.120285526,
|
||||
0.043005254,
|
||||
0.04814879,
|
||||
-0.036352232,
|
||||
-0.003885529,
|
||||
-0.018503373,
|
||||
-0.088186465,
|
||||
-0.0031517749,
|
||||
0.09290919,
|
||||
-1.1695094e-33,
|
||||
-0.015589721,
|
||||
-0.13189551,
|
||||
0.008088751,
|
||||
0.06899503,
|
||||
0.07353927,
|
||||
-0.030646399,
|
||||
0.05110342,
|
||||
0.03081624,
|
||||
-0.07850498,
|
||||
-0.021147482,
|
||||
0.00017823944,
|
||||
-0.10502706,
|
||||
0.030078856,
|
||||
0.02572523,
|
||||
-0.068158925,
|
||||
-0.025015576,
|
||||
0.021830637,
|
||||
0.049748335,
|
||||
0.01520941,
|
||||
-0.080153145,
|
||||
0.06796621,
|
||||
0.021865685,
|
||||
-0.034017574,
|
||||
-0.030821111,
|
||||
-0.048006665,
|
||||
0.0005615041,
|
||||
-0.0137883695,
|
||||
-0.04500587,
|
||||
0.015368256,
|
||||
-0.0043663937,
|
||||
0.037706476,
|
||||
0.0049090013,
|
||||
-0.06216566,
|
||||
0.03060772,
|
||||
0.030548712,
|
||||
0.029262561,
|
||||
0.020701125,
|
||||
0.0056516766,
|
||||
0.010610447,
|
||||
0.019530762,
|
||||
-0.05664136,
|
||||
-0.022654066,
|
||||
-0.0010107337,
|
||||
-0.020805702,
|
||||
-0.012242364,
|
||||
-0.05591731,
|
||||
-0.049421698,
|
||||
0.024721064,
|
||||
0.05803342,
|
||||
0.010474127,
|
||||
-0.008790625,
|
||||
0.025362873,
|
||||
0.020258408,
|
||||
0.004368581,
|
||||
-0.01018003,
|
||||
0.012385932,
|
||||
-0.037656736,
|
||||
-0.05642639,
|
||||
0.020923307,
|
||||
0.022813153,
|
||||
-0.005735433,
|
||||
0.015326356,
|
||||
-0.108707875,
|
||||
0.048076265,
|
||||
0.023256551,
|
||||
-0.10311626,
|
||||
0.061980195,
|
||||
-0.07340407,
|
||||
0.051583096,
|
||||
0.07360003,
|
||||
-0.029443117,
|
||||
-0.014564469,
|
||||
0.042043358,
|
||||
-0.020252181,
|
||||
0.0147808045,
|
||||
-0.0285806,
|
||||
0.07891856,
|
||||
0.056849223,
|
||||
-0.106308356,
|
||||
0.0197874,
|
||||
0.0269322,
|
||||
-0.04749746,
|
||||
-0.066681586,
|
||||
-0.10474516,
|
||||
0.012599429,
|
||||
-0.056163482,
|
||||
-0.04901015,
|
||||
-0.04571026,
|
||||
0.09704481,
|
||||
-0.105899766,
|
||||
0.044303197,
|
||||
-0.020125533,
|
||||
-0.0368709,
|
||||
-0.015417924,
|
||||
0.042297333,
|
||||
-8.289866e-35,
|
||||
0.07415767,
|
||||
0.10998298,
|
||||
-0.016995763,
|
||||
0.01066263,
|
||||
-0.0012327223,
|
||||
0.028000232,
|
||||
0.0714317,
|
||||
-0.02320065,
|
||||
-0.07778205,
|
||||
0.11864239,
|
||||
-0.016559754,
|
||||
0.037961867,
|
||||
0.02930022,
|
||||
-0.008237686,
|
||||
0.059777655,
|
||||
0.008086454,
|
||||
0.02075205,
|
||||
0.025284613,
|
||||
-0.055471037,
|
||||
0.0073576584,
|
||||
-0.013398135,
|
||||
0.11896543,
|
||||
-0.014611002,
|
||||
0.07691816,
|
||||
-0.019711656,
|
||||
-0.01920917,
|
||||
-0.004744884,
|
||||
0.08173054,
|
||||
0.019665759,
|
||||
-0.013193461,
|
||||
0.06215852,
|
||||
0.07420406,
|
||||
-0.073212065,
|
||||
0.036052067,
|
||||
0.07328616,
|
||||
-0.057373393,
|
||||
0.08346425,
|
||||
0.018834447,
|
||||
0.03309735,
|
||||
0.041197047,
|
||||
0.033917964,
|
||||
0.09151449,
|
||||
-0.051731598,
|
||||
0.049615093,
|
||||
0.01124018,
|
||||
0.06661862,
|
||||
0.07268375,
|
||||
-0.013245848,
|
||||
-0.039673895,
|
||||
-0.012173254,
|
||||
0.0017787582,
|
||||
-0.05746287,
|
||||
-0.013884767,
|
||||
0.020205025,
|
||||
-0.029692367,
|
||||
-0.031010685,
|
||||
0.0149556715,
|
||||
0.026381323,
|
||||
-0.025382591,
|
||||
0.0074336748,
|
||||
-0.00949915,
|
||||
0.015655186,
|
||||
-0.0012397208,
|
||||
-0.0032508406,
|
||||
-0.046632554,
|
||||
-0.0030316226,
|
||||
-0.007273208,
|
||||
0.064231135,
|
||||
-0.034431897,
|
||||
-0.06433184,
|
||||
0.045421343,
|
||||
-0.010773523,
|
||||
-0.017881984,
|
||||
0.010312532,
|
||||
-0.024369273,
|
||||
-0.008478495,
|
||||
-0.02457377,
|
||||
-0.0263535,
|
||||
-0.027263613,
|
||||
0.047060315,
|
||||
0.08128726,
|
||||
0.0045517692,
|
||||
-0.010821656,
|
||||
0.026526682,
|
||||
0.018961033,
|
||||
0.059243083,
|
||||
0.001561823,
|
||||
0.09838158,
|
||||
0.00822081,
|
||||
0.008796511,
|
||||
-0.0060577285,
|
||||
0.028892087,
|
||||
0.08253284,
|
||||
0.049560018,
|
||||
0.023363132,
|
||||
-1.498271e-08,
|
||||
-0.036891207,
|
||||
-0.10629833,
|
||||
0.030452948,
|
||||
0.049268734,
|
||||
-0.0030453752,
|
||||
0.07413954,
|
||||
-0.07043819,
|
||||
-0.034285706,
|
||||
-0.009679971,
|
||||
-0.046219327,
|
||||
0.013510038,
|
||||
-0.018686565,
|
||||
-0.048570327,
|
||||
0.0028313443,
|
||||
-0.06190722,
|
||||
-0.053201936,
|
||||
0.0060967463,
|
||||
-0.043467365,
|
||||
0.042226154,
|
||||
0.03455835,
|
||||
-0.0375257,
|
||||
0.023590367,
|
||||
0.054896712,
|
||||
0.029878648,
|
||||
0.019286606,
|
||||
0.026097741,
|
||||
0.06938145,
|
||||
0.06272366,
|
||||
-0.09566521,
|
||||
-0.07481147,
|
||||
0.025204772,
|
||||
0.039396077,
|
||||
0.036375154,
|
||||
-0.01104443,
|
||||
-0.028223084,
|
||||
0.111878626,
|
||||
0.13400707,
|
||||
0.06680113,
|
||||
-0.011737675,
|
||||
-0.03585406,
|
||||
-0.07978788,
|
||||
0.032793757,
|
||||
-0.0021075818,
|
||||
-0.028365146,
|
||||
-0.042218164,
|
||||
-0.08132239,
|
||||
-0.0753423,
|
||||
-0.043771427,
|
||||
-0.015633285,
|
||||
-0.14193884,
|
||||
-0.055949364,
|
||||
0.025526602,
|
||||
-0.023186589,
|
||||
0.061106257,
|
||||
-0.056208834,
|
||||
0.00838827,
|
||||
0.014720396,
|
||||
-0.014650135,
|
||||
-0.012830787,
|
||||
0.08434067,
|
||||
0.024660436,
|
||||
0.05366935,
|
||||
0.005782819,
|
||||
-0.10599063
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 6,
|
||||
"total_tokens": 6
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_cancel[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is batch cancel test file 0 with substantial content"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.010706507,
|
||||
0.11740309,
|
||||
-0.070396945,
|
||||
0.036590267,
|
||||
0.03445541,
|
||||
-0.037278067,
|
||||
0.033794403,
|
||||
-0.013823747,
|
||||
-0.032249726,
|
||||
0.039381154,
|
||||
0.09738964,
|
||||
0.043944314,
|
||||
-0.015195914,
|
||||
-0.08339148,
|
||||
-0.12092182,
|
||||
-0.0144716315,
|
||||
-0.06525938,
|
||||
0.008907217,
|
||||
-0.016506711,
|
||||
-0.011929026,
|
||||
-0.0519942,
|
||||
0.07381637,
|
||||
0.028294124,
|
||||
0.056386005,
|
||||
0.028838597,
|
||||
0.02860147,
|
||||
-0.046813786,
|
||||
-0.018329943,
|
||||
-0.037620317,
|
||||
-0.06344129,
|
||||
0.037448265,
|
||||
0.0807444,
|
||||
0.08218735,
|
||||
-0.018610513,
|
||||
0.16465282,
|
||||
0.006478139,
|
||||
0.009052014,
|
||||
0.024081843,
|
||||
0.04604129,
|
||||
-0.016105218,
|
||||
0.050088186,
|
||||
-0.014189308,
|
||||
-0.055208918,
|
||||
-0.024689473,
|
||||
0.009216049,
|
||||
0.0032953622,
|
||||
-0.08004139,
|
||||
-0.050898325,
|
||||
0.030319132,
|
||||
0.0038868543,
|
||||
-0.03242241,
|
||||
-0.008002084,
|
||||
-0.05405017,
|
||||
0.0034951256,
|
||||
0.026613077,
|
||||
-0.03749797,
|
||||
0.074383445,
|
||||
0.05947148,
|
||||
-0.037571322,
|
||||
0.07424358,
|
||||
-0.031258598,
|
||||
-0.010979168,
|
||||
-0.115162514,
|
||||
0.016076973,
|
||||
0.12323825,
|
||||
0.057677355,
|
||||
-0.08872208,
|
||||
-0.028623635,
|
||||
0.05342226,
|
||||
-0.060159575,
|
||||
-0.07479101,
|
||||
-0.01794232,
|
||||
-0.0049816607,
|
||||
0.08948416,
|
||||
-0.042007502,
|
||||
0.0925552,
|
||||
-0.016678093,
|
||||
0.013261441,
|
||||
-0.0068968083,
|
||||
0.00078877964,
|
||||
-0.070652686,
|
||||
-0.14053895,
|
||||
0.054617904,
|
||||
-0.064937904,
|
||||
-0.036082774,
|
||||
0.04364618,
|
||||
0.039191015,
|
||||
0.009325763,
|
||||
0.055350192,
|
||||
0.007441803,
|
||||
-0.04520714,
|
||||
0.0070686075,
|
||||
0.029522296,
|
||||
0.016590035,
|
||||
-0.020568646,
|
||||
0.083674796,
|
||||
0.0076218233,
|
||||
0.006881344,
|
||||
0.013654858,
|
||||
0.03697504,
|
||||
0.04504176,
|
||||
-0.012595865,
|
||||
-0.006368664,
|
||||
-0.006188894,
|
||||
-0.02347456,
|
||||
-0.014876863,
|
||||
0.07330545,
|
||||
-0.008524341,
|
||||
0.03080002,
|
||||
-0.079184264,
|
||||
-0.002168809,
|
||||
-0.04496155,
|
||||
0.02353669,
|
||||
-0.061784163,
|
||||
0.019026963,
|
||||
-0.034334134,
|
||||
0.07823938,
|
||||
0.086644776,
|
||||
-0.100164026,
|
||||
0.00979978,
|
||||
0.043132447,
|
||||
-0.00027732752,
|
||||
-0.007950898,
|
||||
-0.03439145,
|
||||
-0.07176784,
|
||||
-0.010847044,
|
||||
0.10318583,
|
||||
1.28398045e-33,
|
||||
-0.057539165,
|
||||
-0.10064088,
|
||||
-0.036363184,
|
||||
0.070467934,
|
||||
0.12267441,
|
||||
0.023121687,
|
||||
0.036528632,
|
||||
0.043095388,
|
||||
-0.053614546,
|
||||
0.034320176,
|
||||
-0.015772322,
|
||||
-0.07880764,
|
||||
0.019716268,
|
||||
0.017762613,
|
||||
-0.094458655,
|
||||
-0.08139035,
|
||||
0.027233537,
|
||||
0.07888667,
|
||||
-0.024265131,
|
||||
-0.054107342,
|
||||
0.11021126,
|
||||
-0.016241824,
|
||||
-0.05417309,
|
||||
-0.028439889,
|
||||
-0.027373016,
|
||||
-0.01668086,
|
||||
-0.031238388,
|
||||
-0.03203346,
|
||||
0.017995317,
|
||||
-0.011522754,
|
||||
-0.0029258654,
|
||||
0.022844825,
|
||||
-0.019639384,
|
||||
0.05111425,
|
||||
-0.0015511515,
|
||||
0.04084381,
|
||||
0.0043716393,
|
||||
-0.05789265,
|
||||
0.024110112,
|
||||
0.03920258,
|
||||
-0.08151888,
|
||||
-0.008190904,
|
||||
-0.0645496,
|
||||
-0.014420588,
|
||||
0.00016276255,
|
||||
-0.10466175,
|
||||
-0.015631696,
|
||||
-0.054435816,
|
||||
0.03390489,
|
||||
0.042083304,
|
||||
0.041493565,
|
||||
0.033552594,
|
||||
0.027098974,
|
||||
-0.035584476,
|
||||
-0.025616122,
|
||||
0.015369336,
|
||||
0.025080213,
|
||||
-0.047622968,
|
||||
0.0076927147,
|
||||
0.048611037,
|
||||
0.07658855,
|
||||
0.030115629,
|
||||
-0.10192636,
|
||||
0.009031788,
|
||||
-0.026905872,
|
||||
-0.07093241,
|
||||
0.009540495,
|
||||
-0.0967732,
|
||||
0.006907292,
|
||||
0.008907563,
|
||||
-0.036709655,
|
||||
-0.0074325944,
|
||||
0.06927971,
|
||||
-0.044891518,
|
||||
-0.0022573345,
|
||||
-0.05632572,
|
||||
0.03744841,
|
||||
0.026788702,
|
||||
-0.00916575,
|
||||
0.008179489,
|
||||
0.08744597,
|
||||
-0.046512436,
|
||||
-0.061149366,
|
||||
-0.13555244,
|
||||
0.0010608839,
|
||||
-0.06323009,
|
||||
-0.039003603,
|
||||
-0.07015582,
|
||||
0.03916791,
|
||||
-0.07763432,
|
||||
-0.00032964678,
|
||||
-0.026286542,
|
||||
-0.053487364,
|
||||
0.009920836,
|
||||
0.104119115,
|
||||
-1.9471978e-33,
|
||||
0.04772588,
|
||||
0.04490678,
|
||||
-0.04262699,
|
||||
0.03524018,
|
||||
-0.003943472,
|
||||
0.033365145,
|
||||
0.06762878,
|
||||
-0.021556355,
|
||||
-0.043953415,
|
||||
0.023543492,
|
||||
0.005500359,
|
||||
0.03756542,
|
||||
0.025656395,
|
||||
-0.014806406,
|
||||
0.01845547,
|
||||
0.015662882,
|
||||
0.06915146,
|
||||
0.010516805,
|
||||
-0.08958506,
|
||||
0.008974718,
|
||||
-0.035460126,
|
||||
0.05160542,
|
||||
0.01763933,
|
||||
0.067841165,
|
||||
-0.02522728,
|
||||
-0.022180483,
|
||||
-0.085712284,
|
||||
0.061407775,
|
||||
0.07101853,
|
||||
-0.0015686463,
|
||||
0.055281166,
|
||||
0.04126171,
|
||||
-0.04599903,
|
||||
-0.037977487,
|
||||
0.09936549,
|
||||
-0.064348385,
|
||||
0.07501729,
|
||||
0.06690245,
|
||||
0.01264843,
|
||||
0.011582279,
|
||||
0.06661292,
|
||||
0.083571374,
|
||||
-0.05528334,
|
||||
0.03757593,
|
||||
0.043382253,
|
||||
0.059041474,
|
||||
0.056976013,
|
||||
-0.02765602,
|
||||
-0.00018057597,
|
||||
-0.010140114,
|
||||
-0.023275468,
|
||||
-0.040977187,
|
||||
-0.0051338123,
|
||||
0.06462851,
|
||||
-0.015096949,
|
||||
-0.04108323,
|
||||
0.013806998,
|
||||
-0.013243718,
|
||||
-0.04096836,
|
||||
-0.021470992,
|
||||
0.0037039437,
|
||||
0.04606251,
|
||||
0.027378108,
|
||||
-0.009201031,
|
||||
0.024913032,
|
||||
0.027817363,
|
||||
0.011912681,
|
||||
0.072464235,
|
||||
-0.04599433,
|
||||
-0.033524342,
|
||||
0.031872187,
|
||||
-0.0017134893,
|
||||
-0.030329237,
|
||||
0.021338675,
|
||||
0.050125677,
|
||||
-0.006607719,
|
||||
0.005844466,
|
||||
-0.049508642,
|
||||
2.296406e-05,
|
||||
0.033044532,
|
||||
0.07586271,
|
||||
0.0094868485,
|
||||
-0.0023229877,
|
||||
0.063257135,
|
||||
0.0073867897,
|
||||
0.067748606,
|
||||
-0.088573374,
|
||||
0.06831021,
|
||||
0.0047544846,
|
||||
0.08063805,
|
||||
-0.02170177,
|
||||
0.020645779,
|
||||
0.082571074,
|
||||
0.039116666,
|
||||
0.03906674,
|
||||
-1.756136e-08,
|
||||
-0.01928442,
|
||||
-0.123927765,
|
||||
0.0188664,
|
||||
0.03889619,
|
||||
0.003943178,
|
||||
0.017261649,
|
||||
-0.072421774,
|
||||
0.010595731,
|
||||
-0.032426827,
|
||||
-0.07068102,
|
||||
0.027171727,
|
||||
-0.032465994,
|
||||
-0.03428293,
|
||||
0.00012704723,
|
||||
-0.07441139,
|
||||
-0.061249517,
|
||||
0.003310212,
|
||||
-0.030616615,
|
||||
0.037538156,
|
||||
0.013060206,
|
||||
-0.02899822,
|
||||
0.002607385,
|
||||
0.023053044,
|
||||
-0.008261543,
|
||||
0.027366797,
|
||||
0.041916996,
|
||||
0.07509514,
|
||||
0.093088634,
|
||||
-0.05660954,
|
||||
-0.10259794,
|
||||
0.041243467,
|
||||
-0.025973666,
|
||||
0.013900956,
|
||||
0.0023358895,
|
||||
-0.075266555,
|
||||
0.07490993,
|
||||
0.14500652,
|
||||
0.04697599,
|
||||
-0.03860971,
|
||||
0.009254478,
|
||||
-0.06991552,
|
||||
0.011762797,
|
||||
0.02150895,
|
||||
0.010407091,
|
||||
-0.016874894,
|
||||
-0.057741348,
|
||||
-0.075219,
|
||||
-0.07250321,
|
||||
-0.03090426,
|
||||
-0.110799745,
|
||||
-0.024827298,
|
||||
0.0065941666,
|
||||
-0.027638538,
|
||||
0.08827356,
|
||||
-0.044589255,
|
||||
-0.04193462,
|
||||
0.021976525,
|
||||
0.015851181,
|
||||
-0.07105447,
|
||||
0.106275305,
|
||||
0.058465168,
|
||||
0.0026831257,
|
||||
-0.006616897,
|
||||
-0.086507544
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 10,
|
||||
"total_tokens": 10
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,39 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_cancel[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://localhost:11434/api/ps",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/api/ps",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "ollama._types.ProcessResponse",
|
||||
"__data__": {
|
||||
"models": [
|
||||
{
|
||||
"model": "all-minilm:l6-v2",
|
||||
"name": "all-minilm:l6-v2",
|
||||
"digest": "1b226e2802dbb772b5fc32a58f103ca1804ef7501331012de126ab22f67475ef",
|
||||
"expires_at": "2025-10-06T16:40:13.262640-07:00",
|
||||
"size": 590204928,
|
||||
"size_vram": 590204928,
|
||||
"details": {
|
||||
"parent_model": "",
|
||||
"format": "gguf",
|
||||
"family": "bert",
|
||||
"families": [
|
||||
"bert"
|
||||
],
|
||||
"parameter_size": "23M",
|
||||
"quantization_level": "F16"
|
||||
}
|
||||
}
|
||||
]
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,423 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_cancel[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://0.0.0.0:11434/v1/v1/embeddings",
|
||||
"headers": {},
|
||||
"body": {
|
||||
"model": "all-minilm:l6-v2",
|
||||
"input": [
|
||||
"This is batch cancel test file 1 with substantial content"
|
||||
],
|
||||
"encoding_format": "float"
|
||||
},
|
||||
"endpoint": "/v1/embeddings",
|
||||
"model": "all-minilm:l6-v2"
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "openai.types.create_embedding_response.CreateEmbeddingResponse",
|
||||
"__data__": {
|
||||
"data": [
|
||||
{
|
||||
"embedding": [
|
||||
-0.024848156,
|
||||
0.10927085,
|
||||
-0.0545053,
|
||||
0.038470518,
|
||||
0.046556868,
|
||||
-0.034411646,
|
||||
0.04878781,
|
||||
-0.014318634,
|
||||
-0.046015147,
|
||||
0.044597667,
|
||||
0.09629065,
|
||||
0.058968317,
|
||||
-0.007982022,
|
||||
-0.10140896,
|
||||
-0.10389055,
|
||||
-0.019553911,
|
||||
-0.07593768,
|
||||
0.025729634,
|
||||
-0.029175822,
|
||||
-0.02637171,
|
||||
-0.050457876,
|
||||
0.066799924,
|
||||
0.022711048,
|
||||
0.06541894,
|
||||
0.041600663,
|
||||
0.030976223,
|
||||
-0.056684654,
|
||||
-0.0035002322,
|
||||
-0.050632603,
|
||||
-0.08931927,
|
||||
0.040150054,
|
||||
0.06798157,
|
||||
0.08541512,
|
||||
-0.0107848635,
|
||||
0.15392521,
|
||||
0.009335848,
|
||||
0.010962297,
|
||||
0.029146858,
|
||||
0.047823314,
|
||||
-0.026440151,
|
||||
0.04159767,
|
||||
-0.010160618,
|
||||
-0.05779408,
|
||||
-0.039702307,
|
||||
-0.004494967,
|
||||
-0.004617397,
|
||||
-0.08862508,
|
||||
-0.034483556,
|
||||
0.024042498,
|
||||
0.0051155766,
|
||||
-0.0317056,
|
||||
-0.01699217,
|
||||
-0.053024635,
|
||||
0.015636722,
|
||||
0.03557156,
|
||||
-0.039649993,
|
||||
0.081902996,
|
||||
0.06176357,
|
||||
-0.05502012,
|
||||
0.06357122,
|
||||
-0.030193875,
|
||||
-0.012515638,
|
||||
-0.12543206,
|
||||
0.012063709,
|
||||
0.12448795,
|
||||
0.040869392,
|
||||
-0.07753088,
|
||||
-0.021475459,
|
||||
0.04500842,
|
||||
-0.056871727,
|
||||
-0.09496841,
|
||||
-0.01180043,
|
||||
-0.0017254521,
|
||||
0.08008634,
|
||||
-0.047713377,
|
||||
0.08782804,
|
||||
-0.02004271,
|
||||
0.033268984,
|
||||
-0.016207146,
|
||||
-0.010731495,
|
||||
-0.063805684,
|
||||
-0.14302677,
|
||||
0.0575187,
|
||||
-0.06904251,
|
||||
-0.037962824,
|
||||
0.0182337,
|
||||
0.042886198,
|
||||
0.01039097,
|
||||
0.044122625,
|
||||
-0.0020459748,
|
||||
-0.014757414,
|
||||
0.0011372506,
|
||||
0.07999029,
|
||||
0.018020395,
|
||||
-0.018433796,
|
||||
0.07817236,
|
||||
0.012330995,
|
||||
0.007078602,
|
||||
0.03731257,
|
||||
0.03993665,
|
||||
0.039117657,
|
||||
0.0077354074,
|
||||
-0.009170466,
|
||||
-0.018691367,
|
||||
-0.028763011,
|
||||
-0.019665359,
|
||||
0.062140632,
|
||||
-0.020356707,
|
||||
0.038877316,
|
||||
-0.08305566,
|
||||
0.00014209712,
|
||||
-0.05700167,
|
||||
0.021387467,
|
||||
-0.054998472,
|
||||
0.03538585,
|
||||
-0.023105556,
|
||||
0.089621656,
|
||||
0.09418147,
|
||||
-0.08390289,
|
||||
0.009763535,
|
||||
0.043676704,
|
||||
-0.0022283366,
|
||||
0.00070641236,
|
||||
-0.03374215,
|
||||
-0.07274797,
|
||||
-0.034256138,
|
||||
0.09228734,
|
||||
1.2329422e-33,
|
||||
-0.06229734,
|
||||
-0.10348473,
|
||||
-0.05939012,
|
||||
0.07817319,
|
||||
0.12856846,
|
||||
0.03253048,
|
||||
0.03706221,
|
||||
0.03843275,
|
||||
-0.06781762,
|
||||
0.027851813,
|
||||
-0.03286515,
|
||||
-0.07305933,
|
||||
0.011496317,
|
||||
0.016992282,
|
||||
-0.10859345,
|
||||
-0.089275,
|
||||
0.02053902,
|
||||
0.07540007,
|
||||
-0.030434899,
|
||||
-0.057486024,
|
||||
0.1028371,
|
||||
-0.011332772,
|
||||
-0.040277272,
|
||||
-0.022627348,
|
||||
-0.029583039,
|
||||
-0.042487655,
|
||||
-0.01710431,
|
||||
-0.028937005,
|
||||
0.034644134,
|
||||
-0.015131404,
|
||||
-0.005402634,
|
||||
0.0111823045,
|
||||
-0.024323324,
|
||||
0.061144948,
|
||||
-0.0068504023,
|
||||
0.04550556,
|
||||
0.017341396,
|
||||
-0.063010655,
|
||||
0.033939265,
|
||||
0.029030005,
|
||||
-0.07075115,
|
||||
0.0076140417,
|
||||
-0.056033216,
|
||||
-0.01839173,
|
||||
0.006444027,
|
||||
-0.10148905,
|
||||
-0.024238782,
|
||||
-0.045753844,
|
||||
0.029873326,
|
||||
0.03732028,
|
||||
0.05342056,
|
||||
0.024428835,
|
||||
0.03200607,
|
||||
-0.045322895,
|
||||
-0.009412481,
|
||||
0.01895284,
|
||||
0.026068604,
|
||||
-0.043451786,
|
||||
0.017836504,
|
||||
0.060751975,
|
||||
0.0770648,
|
||||
0.037520513,
|
||||
-0.094844334,
|
||||
0.018022675,
|
||||
-0.028010713,
|
||||
-0.05970307,
|
||||
0.0042470302,
|
||||
-0.08537647,
|
||||
0.0025366507,
|
||||
0.0059753954,
|
||||
-0.040670317,
|
||||
-0.008420785,
|
||||
0.070101276,
|
||||
-0.05581281,
|
||||
0.009997155,
|
||||
-0.053269707,
|
||||
0.030278698,
|
||||
0.034753144,
|
||||
-0.0069992156,
|
||||
-0.0018294669,
|
||||
0.052869115,
|
||||
-0.047554925,
|
||||
-0.07009094,
|
||||
-0.12028551,
|
||||
-0.016411684,
|
||||
-0.0558196,
|
||||
-0.026485136,
|
||||
-0.07406597,
|
||||
0.052336086,
|
||||
-0.07966716,
|
||||
-0.009600498,
|
||||
-0.016012779,
|
||||
-0.04670444,
|
||||
0.0040856744,
|
||||
0.13087922,
|
||||
-1.9130171e-33,
|
||||
0.04951988,
|
||||
0.04144521,
|
||||
-0.030660233,
|
||||
0.02966906,
|
||||
-0.0019053655,
|
||||
0.038034633,
|
||||
0.053598672,
|
||||
-0.03873592,
|
||||
-0.050682254,
|
||||
0.0163216,
|
||||
-0.018117629,
|
||||
0.02705123,
|
||||
0.014957701,
|
||||
-0.029251544,
|
||||
0.010732444,
|
||||
0.01150037,
|
||||
0.08527361,
|
||||
0.000666767,
|
||||
-0.09031944,
|
||||
0.007236525,
|
||||
-0.0394124,
|
||||
0.032647807,
|
||||
0.029387591,
|
||||
0.0696317,
|
||||
-0.028400488,
|
||||
-0.019728381,
|
||||
-0.08580391,
|
||||
0.050916594,
|
||||
0.07555233,
|
||||
0.0013333871,
|
||||
0.036405865,
|
||||
0.03485496,
|
||||
-0.035891958,
|
||||
-0.03518406,
|
||||
0.08422707,
|
||||
-0.07100648,
|
||||
0.066512264,
|
||||
0.0566844,
|
||||
0.005254722,
|
||||
0.026210023,
|
||||
0.06271422,
|
||||
0.07715752,
|
||||
-0.042685844,
|
||||
0.029498853,
|
||||
0.048694577,
|
||||
0.06829996,
|
||||
0.05471948,
|
||||
-0.014717811,
|
||||
-0.0084376065,
|
||||
-0.007800526,
|
||||
-0.033968475,
|
||||
-0.035792083,
|
||||
-0.01680357,
|
||||
0.056615632,
|
||||
-0.008940466,
|
||||
-0.044396702,
|
||||
0.033141203,
|
||||
-0.020710811,
|
||||
-0.052891865,
|
||||
-0.012946567,
|
||||
0.013425288,
|
||||
0.045469046,
|
||||
0.02655372,
|
||||
-7.159544e-06,
|
||||
0.033383444,
|
||||
0.012771919,
|
||||
0.0050781234,
|
||||
0.05739414,
|
||||
-0.05292731,
|
||||
-0.009027621,
|
||||
0.019719183,
|
||||
-0.0046205786,
|
||||
-0.012921344,
|
||||
0.021115582,
|
||||
0.063510135,
|
||||
0.006540324,
|
||||
0.008657973,
|
||||
-0.044172782,
|
||||
-0.0010352373,
|
||||
0.025917202,
|
||||
0.07357742,
|
||||
0.012915724,
|
||||
-0.010159995,
|
||||
0.05862044,
|
||||
0.0032137444,
|
||||
0.08368076,
|
||||
-0.06552963,
|
||||
0.06294139,
|
||||
0.004963379,
|
||||
0.08497223,
|
||||
-0.030302247,
|
||||
0.028541481,
|
||||
0.103464715,
|
||||
0.03432187,
|
||||
0.039947473,
|
||||
-1.757192e-08,
|
||||
-0.020163277,
|
||||
-0.12507844,
|
||||
0.015846072,
|
||||
0.038265407,
|
||||
-0.0031526515,
|
||||
0.01804952,
|
||||
-0.0817553,
|
||||
0.030486222,
|
||||
-0.02073271,
|
||||
-0.069118954,
|
||||
0.0252006,
|
||||
-0.016496325,
|
||||
-0.018695008,
|
||||
-0.0063493066,
|
||||
-0.08448383,
|
||||
-0.05474651,
|
||||
0.008191211,
|
||||
-0.04699509,
|
||||
0.03820692,
|
||||
0.019186925,
|
||||
-0.006977571,
|
||||
-0.0002934883,
|
||||
0.030278133,
|
||||
-0.009153849,
|
||||
0.030300315,
|
||||
0.04737054,
|
||||
0.06026962,
|
||||
0.09765302,
|
||||
-0.05529498,
|
||||
-0.09553832,
|
||||
0.06008278,
|
||||
-0.025960611,
|
||||
0.034287665,
|
||||
-0.012333093,
|
||||
-0.07106284,
|
||||
0.05141244,
|
||||
0.14179605,
|
||||
0.04709826,
|
||||
-0.049292527,
|
||||
0.014455253,
|
||||
-0.047851674,
|
||||
0.011403938,
|
||||
0.014072481,
|
||||
0.010494679,
|
||||
-0.0009859774,
|
||||
-0.06089218,
|
||||
-0.07293921,
|
||||
-0.07961594,
|
||||
-0.03404924,
|
||||
-0.10086713,
|
||||
-0.031331882,
|
||||
0.0042822976,
|
||||
-0.0045380252,
|
||||
0.09583955,
|
||||
-0.044172354,
|
||||
-0.034359995,
|
||||
0.023726532,
|
||||
0.02167657,
|
||||
-0.06509328,
|
||||
0.09268318,
|
||||
0.055370033,
|
||||
0.003980954,
|
||||
-0.0053826002,
|
||||
-0.07774321
|
||||
],
|
||||
"index": 0,
|
||||
"object": "embedding"
|
||||
}
|
||||
],
|
||||
"model": "all-minilm:l6-v2",
|
||||
"object": "list",
|
||||
"usage": {
|
||||
"prompt_tokens": 10,
|
||||
"total_tokens": 10
|
||||
}
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
|
@ -0,0 +1,20 @@
|
|||
{
|
||||
"test_id": "tests/integration/vector_io/test_openai_vector_stores.py::test_openai_vector_store_file_batch_create_and_retrieve[client_with_models-ollama/llama3.2:3b-instruct-fp16-None-ollama/all-minilm:l6-v2-None-384]",
|
||||
"request": {
|
||||
"method": "POST",
|
||||
"url": "http://localhost:11434/api/ps",
|
||||
"headers": {},
|
||||
"body": {},
|
||||
"endpoint": "/api/ps",
|
||||
"model": ""
|
||||
},
|
||||
"response": {
|
||||
"body": {
|
||||
"__type__": "ollama._types.ProcessResponse",
|
||||
"__data__": {
|
||||
"models": []
|
||||
}
|
||||
},
|
||||
"is_streaming": false
|
||||
}
|
||||
}
|
Some files were not shown because too many files have changed in this diff Show more
Loading…
Add table
Add a link
Reference in a new issue