mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-27 09:01:59 +00:00
add `v1/providers/` which uses PUT to allow users to change their provider configuration this is a follow up to #1429 and related to #1359 a user can call something like: `llama_stack_client.providers.update(api="inference", provider_id="ollama", provider_type="remote::ollama", config={'url': 'http:/localhost:12345'})` or `llama-stack-client providers update inference ollama remote::ollama "{'url': 'http://localhost:12345'}"` this API works by adding a `RequestMiddleware` to the server which checks requests, and if the user is using PUT /v1/providers, the routes are re-registered with the re-initialized provider configurations/methods for the client, `self.impls` is updated to hold the proper methods+configurations this depends on a client PR, the CI will fail until then but succeeded locally Signed-off-by: Charlie Doern <cdoern@redhat.com> |
||
|---|---|---|
| .. | ||
| agents | ||
| batch_inference | ||
| benchmarks | ||
| common | ||
| datasetio | ||
| datasets | ||
| eval | ||
| files | ||
| inference | ||
| inspect | ||
| models | ||
| post_training | ||
| providers | ||
| safety | ||
| scoring | ||
| scoring_functions | ||
| shields | ||
| synthetic_data_generation | ||
| telemetry | ||
| tools | ||
| vector_dbs | ||
| vector_io | ||
| __init__.py | ||
| datatypes.py | ||
| resource.py | ||
| version.py | ||