diff --git a/llama_stack/providers/remote/inference/centml/centml.py b/llama_stack/providers/remote/inference/centml/centml.py index 93fac26ba..5f14c23f1 100644 --- a/llama_stack/providers/remote/inference/centml/centml.py +++ b/llama_stack/providers/remote/inference/centml/centml.py @@ -57,12 +57,8 @@ from .config import CentMLImplConfig # published model identifiers to llama-stack's `CoreModelId`. MODEL_ALIASES = [ build_model_entry( - "meta-llama/Llama-3.3-70B-Instruct", - CoreModelId.llama3_3_70b_instruct.value, - ), - build_model_entry( - "meta-llama/Llama-3.1-405B-Instruct-FP8", - CoreModelId.llama3_1_405b_instruct.value, + "meta-llama/Llama-3.2-3B-Instruct", + CoreModelId.llama3_2_3b_instruct.value, ), ] diff --git a/llama_stack/providers/remote/inference/centml/config.py b/llama_stack/providers/remote/inference/centml/config.py index bc9711bdb..7bf101b23 100644 --- a/llama_stack/providers/remote/inference/centml/config.py +++ b/llama_stack/providers/remote/inference/centml/config.py @@ -13,7 +13,7 @@ from pydantic import BaseModel, Field, SecretStr @json_schema_type class CentMLImplConfig(BaseModel): url: str = Field( - default="https://api.centml.com/openai/v1", + default="https://api.centml.org/openai/v1", description="The CentML API server URL", ) api_key: Optional[SecretStr] = Field( @@ -24,6 +24,6 @@ class CentMLImplConfig(BaseModel): @classmethod def sample_run_config(cls, **kwargs) -> Dict[str, Any]: return { - "url": "https://api.centml.com/openai/v1", + "url": "https://api.centml.org/openai/v1", "api_key": "${env.CENTML_API_KEY}", }