mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 09:53:45 +00:00
Completes #3732 by removing runtime URL transformations and requiring users to provide full URLs in configuration. All providers now use 'base_url' consistently and respect the exact URL provided without appending paths like /v1 or /openai/v1 at runtime. Add unit test to enforce URL standardization across remote inference providers (verifies all use 'base_url' field with HttpUrl | None type) BREAKING CHANGE: Users must update configs to include full URL paths (e.g., http://localhost:11434/v1 instead of http://localhost:11434). Signed-off-by: Charlie Doern <cdoern@redhat.com>
25 lines
810 B
Text
25 lines
810 B
Text
---
|
|
description: "Ollama inference provider for running local models through the Ollama runtime."
|
|
sidebar_label: Remote - Ollama
|
|
title: remote::ollama
|
|
---
|
|
|
|
# remote::ollama
|
|
|
|
## Description
|
|
|
|
Ollama inference provider for running local models through the Ollama runtime.
|
|
|
|
## Configuration
|
|
|
|
| Field | Type | Required | Default | Description |
|
|
|-------|------|----------|---------|-------------|
|
|
| `allowed_models` | `list[str] \| None` | No | | List of models that should be registered with the model registry. If None, all models are allowed. |
|
|
| `refresh_models` | `bool` | No | False | Whether to refresh models periodically from the provider |
|
|
| `base_url` | `HttpUrl \| None` | No | http://localhost:11434/v1 | |
|
|
|
|
## Sample Configuration
|
|
|
|
```yaml
|
|
base_url: ${env.OLLAMA_URL:=http://localhost:11434/v1}
|
|
```
|