llama-stack/llama_stack
grs b8f7e1504d
feat: allow the interface on which the server will listen to be configured (#2015)
# What does this PR do?

It may not always be desirable to listen on all interfaces, which is the
default. As an example, by listening instead only on a loopback
interface, the server cannot be reached except from within the host it
is run on. This PR makes this configurable, through a CLI option, an env
var or an entry on the config file.

## Test Plan

I ran a server with and without the added CLI argument to verify that
the argument is used if provided, but the default is as it was before if
not.

Signed-off-by: Gordon Sim <gsim@redhat.com>
2025-05-16 12:59:31 -07:00
..
apis chore: more API validators (#2165) 2025-05-15 11:22:51 -07:00
cli feat: refactor external providers dir (#2049) 2025-05-15 20:17:03 +02:00
distribution feat: allow the interface on which the server will listen to be configured (#2015) 2025-05-16 12:59:31 -07:00
models fix: llama4 tool use prompt fix (#2103) 2025-05-06 22:18:31 -07:00
providers feat: use openai-python for openai inference provider (#2193) 2025-05-16 12:57:56 -07:00
strong_typing chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
templates feat: allow the interface on which the server will listen to be configured (#2015) 2025-05-16 12:59:31 -07:00
ui feat: Adding dark mode, cleaning the UI a small bit, adding a link to the API documentation, and linting the code. (#2182) 2025-05-16 10:48:26 -07:00
__init__.py export LibraryClient 2024-12-13 12:08:00 -08:00
env.py refactor(test): move tools, evals, datasetio, scoring and post training tests (#1401) 2025-03-04 14:53:47 -08:00
log.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00
schema_utils.py chore: enable pyupgrade fixes (#1806) 2025-05-01 14:23:50 -07:00