llama-stack-mirror/llama_stack/providers/inline
ehhuang 9936f33f7e
chore: disable telemetry if otel endpoint isn't set (#3859)
# What does this PR do?

removes error:
ConnectionError: HTTPConnectionPool(host='localhost', port=4318): Max
retries exceeded with url: /v1/traces
(Caused by NewConnectionError('<urllib3.connection.HTTPConnection object
at 0x10fd98e60>: Failed to establish a
         new connection: [Errno 61] Connection refused'))


## Test Plan
uv run llama stack run starter
curl http://localhost:8321/v1/models
observe no error in server logs
2025-10-20 11:42:57 -07:00
..
agents chore!: remove telemetry API usage (#3815) 2025-10-16 10:39:32 -07:00
batches feat(api)!: support extra_body to embeddings and vector_stores APIs (#3794) 2025-10-12 19:01:52 -07:00
datasetio chore(misc): make tests and starter faster (#3042) 2025-08-05 14:55:05 -07:00
eval feat(api)!: BREAKING CHANGE: support passing extra_body through to providers (#3777) 2025-10-10 16:21:44 -07:00
files/localfs feat(tests): make inference_recorder into api_recorder (include tool_invoke) (#3403) 2025-10-09 14:27:51 -07:00
inference feat: Enable setting a default embedding model in the stack (#3803) 2025-10-14 18:25:13 -07:00
ios/inference feat(tools)!: substantial clean up of "Tool" related datatypes (#3627) 2025-10-02 15:12:03 -07:00
post_training fix: update dangling references to llama download command (#3763) 2025-10-09 18:35:02 -07:00
safety chore!: Safety api refactoring to use OpenAIMessageParam (#3796) 2025-10-12 08:01:00 -07:00
scoring feat(api)!: BREAKING CHANGE: support passing extra_body through to providers (#3777) 2025-10-10 16:21:44 -07:00
telemetry chore: disable telemetry if otel endpoint isn't set (#3859) 2025-10-20 11:42:57 -07:00
tool_runtime fix(vector-io): handle missing document_id in insert_chunks (#3521) 2025-10-15 11:02:48 -07:00
vector_io feat: Enable setting a default embedding model in the stack (#3803) 2025-10-14 18:25:13 -07:00
__init__.py impls -> inline, adapters -> remote (#381) 2024-11-06 14:54:05 -08:00