llama-stack-mirror/docs/source/providers/inference
Matthew Farrellee ae804ed5a8 feat: (re-)enable Databricks inference adapter
Databricks inference adapter was broken, would not start, see #3486

- remove deprecated completion / chat_completion endpoints
- enable dynamic model listing w/o refresh, listing is not async
- use SecretStr instead of str for token
- backward incompatible change: for consistency with databricks docs, env DATABRICKS_URL -> DATABRICKS_HOST and DATABRICKS_API_TOKEN -> DATABRICKS_TOKEN
- databricks urls are custom per user/org, add special recorder handling for databricks urls
- add integration test --setup databricks
- enable chat completions tests
- enable embeddings tests
- disable n > 1 tests
- disable embeddings base64 tests
- disable embeddings dimensions tests

note: reasoning models, e.g. gpt oss, fail because databricks has a custom, incompatible response format

test with: ./scripts/integration-tests.sh --stack-config server:ci-tests --setup databricks --subdirs inference --pattern openai

note: databricks needs to be manually added to the ci-tests distro for replay testing
2025-09-20 05:05:05 -04:00
..
index.md feat: add Azure OpenAI inference provider support (#3396) 2025-09-11 13:48:38 +02:00
inline_meta-reference.md docs: auto generated documentation for providers (#2543) 2025-06-30 15:13:20 +02:00
inline_sentence-transformers.md docs: auto generated documentation for providers (#2543) 2025-06-30 15:13:20 +02:00
remote_anthropic.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_azure.md feat: add Azure OpenAI inference provider support (#3396) 2025-09-11 13:48:38 +02:00
remote_bedrock.md fix: use lambda pattern for bedrock config env vars (#3307) 2025-09-05 10:45:11 +02:00
remote_cerebras.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_databricks.md feat: (re-)enable Databricks inference adapter 2025-09-20 05:05:05 -04:00
remote_fireworks.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_gemini.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_groq.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_hf_endpoint.md chore(misc): make tests and starter faster (#3042) 2025-08-05 14:55:05 -07:00
remote_hf_serverless.md chore(misc): make tests and starter faster (#3042) 2025-08-05 14:55:05 -07:00
remote_llama-openai-compat.md docs: auto generated documentation for providers (#2543) 2025-06-30 15:13:20 +02:00
remote_nvidia.md fix: allow default empty vars for conditionals (#2570) 2025-07-01 14:42:05 +02:00
remote_ollama.md feat(registry): make the Stack query providers for model listing (#2862) 2025-07-24 10:39:53 -07:00
remote_openai.md feat(openai): add configurable base_url support with OPENAI_BASE_URL env var (#2919) 2025-07-28 10:16:02 -07:00
remote_passthrough.md docs: auto generated documentation for providers (#2543) 2025-06-30 15:13:20 +02:00
remote_runpod.md feat: consolidate most distros into "starter" (#2516) 2025-07-04 15:58:03 +02:00
remote_sambanova-openai-compat.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_sambanova.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_tgi.md chore(misc): make tests and starter faster (#3042) 2025-08-05 14:55:05 -07:00
remote_together.md feat(starter)!: simplify starter distro; litellm model registry changes (#2916) 2025-07-25 15:02:04 -07:00
remote_vertexai.md feat: Add Google Vertex AI inference provider support (#2841) 2025-08-11 08:22:04 -04:00
remote_vllm.md feat(registry): make the Stack query providers for model listing (#2862) 2025-07-24 10:39:53 -07:00
remote_watsonx.md chore: various watsonx fixes (#3428) 2025-09-16 13:55:10 +02:00