mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-16 18:08:09 +00:00
Merge branch 'main' into chunk-metadata
This commit is contained in:
commit
f52eb51555
40 changed files with 272 additions and 722 deletions
4
docs/_static/llama-stack-spec.html
vendored
4
docs/_static/llama-stack-spec.html
vendored
|
@ -14006,6 +14006,10 @@
|
|||
"rewrite_query": {
|
||||
"type": "boolean",
|
||||
"description": "Whether to rewrite the natural language query for vector search (default false)"
|
||||
},
|
||||
"search_mode": {
|
||||
"type": "string",
|
||||
"description": "The search mode to use - \"keyword\", \"vector\", or \"hybrid\" (default \"vector\")"
|
||||
}
|
||||
},
|
||||
"additionalProperties": false,
|
||||
|
|
4
docs/_static/llama-stack-spec.yaml
vendored
4
docs/_static/llama-stack-spec.yaml
vendored
|
@ -9805,6 +9805,10 @@ components:
|
|||
description: >-
|
||||
Whether to rewrite the natural language query for vector search (default
|
||||
false)
|
||||
search_mode:
|
||||
type: string
|
||||
description: >-
|
||||
The search mode to use - "keyword", "vector", or "hybrid" (default "vector")
|
||||
additionalProperties: false
|
||||
required:
|
||||
- query
|
||||
|
|
|
@ -109,6 +109,18 @@ A Model is an instance of a "Resource" (see [Concepts](../concepts/index)) and i
|
|||
|
||||
What's with the `provider_model_id` field? This is an identifier for the model inside the provider's model catalog. Contrast it with `model_id` which is the identifier for the same model for Llama Stack's purposes. For example, you may want to name "llama3.2:vision-11b" as "image_captioning_model" when you use it in your Stack interactions. When omitted, the server will set `provider_model_id` to be the same as `model_id`.
|
||||
|
||||
If you need to conditionally register a model in the configuration, such as only when specific environment variable(s) are set, this can be accomplished by utilizing a special `__disabled__` string as the default value of an environment variable substitution, as shown below:
|
||||
|
||||
```yaml
|
||||
models:
|
||||
- metadata: {}
|
||||
model_id: ${env.INFERENCE_MODEL:__disabled__}
|
||||
provider_id: ollama
|
||||
provider_model_id: ${env.INFERENCE_MODEL:__disabled__}
|
||||
```
|
||||
|
||||
The snippet above will only register this model if the environment variable `INFERENCE_MODEL` is set and non-empty. If the environment variable is not set, the model will not get registered at all.
|
||||
|
||||
## Server Configuration
|
||||
|
||||
The `server` section configures the HTTP server that serves the Llama Stack APIs:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue