From 2f498ad7764f818cefa65fb7e65c7ddf8e1ae4a3 Mon Sep 17 00:00:00 2001 From: Sajikumar JS Date: Fri, 25 Apr 2025 11:03:21 +0530 Subject: [PATCH] updated the string to watsonx --- README.md | 2 +- .../distributions/remote_hosted_distro/watsonx.md | 10 +++++----- .../providers/remote/inference/watsonx/config.py | 4 ++-- .../providers/remote/inference/watsonx/watsonx.py | 2 +- llama_stack/templates/watsonx/build.yaml | 2 +- llama_stack/templates/watsonx/doc_template.md | 6 +++--- llama_stack/templates/watsonx/watsonx.py | 6 +++--- 7 files changed, 16 insertions(+), 16 deletions(-) diff --git a/README.md b/README.md index 588cbac9d..c2e688763 100644 --- a/README.md +++ b/README.md @@ -119,7 +119,7 @@ Here is a list of the various API providers and available distributions that can | OpenAI | Hosted | | ✅ | | | | | Anthropic | Hosted | | ✅ | | | | | Gemini | Hosted | | ✅ | | | | -| WatsonX | Hosted | | ✅ | | | | +| watsonx | Hosted | | ✅ | | | | ### Distributions diff --git a/docs/source/distributions/remote_hosted_distro/watsonx.md b/docs/source/distributions/remote_hosted_distro/watsonx.md index 248903d73..018dc2a3c 100644 --- a/docs/source/distributions/remote_hosted_distro/watsonx.md +++ b/docs/source/distributions/remote_hosted_distro/watsonx.md @@ -2,7 +2,7 @@ orphan: true --- -# WatsonX Distribution +# watsonx Distribution ```{toctree} :maxdepth: 2 @@ -32,8 +32,8 @@ The `llamastack/distribution-watsonx` distribution consists of the following pro The following environment variables can be configured: - `LLAMASTACK_PORT`: Port for the Llama Stack distribution server (default: `5001`) -- `WATSONX_API_KEY`: Watsonx API Key (default: ``) -- `WATSONX_PROJECT_ID`: Watsonx Project ID (default: ``) +- `WATSONX_API_KEY`: watsonx API Key (default: ``) +- `WATSONX_PROJECT_ID`: watsonx Project ID (default: ``) ### Models @@ -52,10 +52,10 @@ The following models are available by default: ### Prerequisite: API Keys -Make sure you have access to a WatsonX API Key. You can get one by referring [watsonx.ai](https://www.ibm.com/docs/en/masv-and-l/maximo-manage/continuous-delivery?topic=setup-create-watsonx-api-key). +Make sure you have access to a watsonx API Key. You can get one by referring [watsonx.ai](https://www.ibm.com/docs/en/masv-and-l/maximo-manage/continuous-delivery?topic=setup-create-watsonx-api-key). -## Running Llama Stack with WatsonX +## Running Llama Stack with watsonx You can do this via Conda (build code), venv or Docker which has a pre-built image. diff --git a/llama_stack/providers/remote/inference/watsonx/config.py b/llama_stack/providers/remote/inference/watsonx/config.py index 3cb14f4e7..7ee99b7e0 100644 --- a/llama_stack/providers/remote/inference/watsonx/config.py +++ b/llama_stack/providers/remote/inference/watsonx/config.py @@ -22,11 +22,11 @@ class WatsonXProviderDataValidator(BaseModel): class WatsonXConfig(BaseModel): url: str = Field( default_factory=lambda: os.getenv("WATSONX_BASE_URL", "https://us-south.ml.cloud.ibm.com"), - description="A base url for accessing the Watsonx.ai", + description="A base url for accessing the watsonx.ai", ) api_key: Optional[SecretStr] = Field( default_factory=lambda: os.getenv("WATSONX_API_KEY"), - description="The Watsonx API key, only needed of using the hosted service", + description="The watsonx API key, only needed of using the hosted service", ) project_id: Optional[str] = Field( default_factory=lambda: os.getenv("WATSONX_PROJECT_ID"), diff --git a/llama_stack/providers/remote/inference/watsonx/watsonx.py b/llama_stack/providers/remote/inference/watsonx/watsonx.py index 1bb8a6bbf..d5d87ec01 100644 --- a/llama_stack/providers/remote/inference/watsonx/watsonx.py +++ b/llama_stack/providers/remote/inference/watsonx/watsonx.py @@ -50,7 +50,7 @@ class WatsonXInferenceAdapter(Inference, ModelRegistryHelper): def __init__(self, config: WatsonXConfig) -> None: ModelRegistryHelper.__init__(self, MODEL_ENTRIES) - print(f"Initializing WatsonXInferenceAdapter({config.url})...") + print(f"Initializing watsonx InferenceAdapter({config.url})...") self._config = config diff --git a/llama_stack/templates/watsonx/build.yaml b/llama_stack/templates/watsonx/build.yaml index 0b5d14b87..badd643ad 100644 --- a/llama_stack/templates/watsonx/build.yaml +++ b/llama_stack/templates/watsonx/build.yaml @@ -1,6 +1,6 @@ version: '2' distribution_spec: - description: Use WatsonX for running LLM inference + description: Use watsonx for running LLM inference providers: inference: - remote::watsonx diff --git a/llama_stack/templates/watsonx/doc_template.md b/llama_stack/templates/watsonx/doc_template.md index e56a32bbf..af0ae15a8 100644 --- a/llama_stack/templates/watsonx/doc_template.md +++ b/llama_stack/templates/watsonx/doc_template.md @@ -1,7 +1,7 @@ --- orphan: true --- -# WatsonX Distribution +# watsonx Distribution ```{toctree} :maxdepth: 2 @@ -38,10 +38,10 @@ The following models are available by default: ### Prerequisite: API Keys -Make sure you have access to a WatsonX API Key. You can get one by referring [watsonx.ai](https://www.ibm.com/docs/en/masv-and-l/maximo-manage/continuous-delivery?topic=setup-create-watsonx-api-key). +Make sure you have access to a watsonx API Key. You can get one by referring [watsonx.ai](https://www.ibm.com/docs/en/masv-and-l/maximo-manage/continuous-delivery?topic=setup-create-watsonx-api-key). -## Running Llama Stack with WatsonX +## Running Llama Stack with watsonx You can do this via Conda (build code), venv or Docker which has a pre-built image. diff --git a/llama_stack/templates/watsonx/watsonx.py b/llama_stack/templates/watsonx/watsonx.py index b1f14d3b3..d59bb6f20 100644 --- a/llama_stack/templates/watsonx/watsonx.py +++ b/llama_stack/templates/watsonx/watsonx.py @@ -59,7 +59,7 @@ def get_distribution_template() -> DistributionTemplate: return DistributionTemplate( name="watsonx", distro_type="remote_hosted", - description="Use WatsonX for running LLM inference", + description="Use watsonx for running LLM inference", container_image=None, template_path=Path(__file__).parent / "doc_template.md", providers=providers, @@ -80,11 +80,11 @@ def get_distribution_template() -> DistributionTemplate: ), "WATSONX_API_KEY": ( "", - "Watsonx API Key", + "watsonx API Key", ), "WATSONX_PROJECT_ID": ( "", - "Watsonx Project ID", + "watsonx Project ID", ), }, )