diff --git a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb index a552ce69d..61b5ab178 100644 --- a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb +++ b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb @@ -513,8 +513,8 @@ " provider_id: code-interpreter\n", " provider_type: inline::code-interpreter\n", " - config: {}\n", - " provider_id: memory-runtime\n", - " provider_type: inline::memory-runtime\n", + " provider_id: rag-runtime\n", + " provider_type: inline::rag-runtime\n", "scoring_fns: []\n", "shields:\n", "- params: null\n", @@ -528,8 +528,8 @@ " toolgroup_id: builtin::websearch\n", "- args: null\n", " mcp_endpoint: null\n", - " provider_id: memory-runtime\n", - " toolgroup_id: builtin::memory\n", + " provider_id: rag-runtime\n", + " toolgroup_id: builtin::rag\n", "- args: null\n", " mcp_endpoint: null\n", " provider_id: code-interpreter\n", @@ -694,8 +694,8 @@ " provider_id: code-interpreter\n", " provider_type: inlin\u001b[1;92me::c\u001b[0mode-interpreter\n", " - config: \u001b[1m{\u001b[0m\u001b[1m}\u001b[0m\n", - " provider_id: memory-runtime\n", - " provider_type: inline::memory-runtime\n", + " provider_id: rag-runtime\n", + " provider_type: inline::rag-runtime\n", "scoring_fns: \u001b[1m[\u001b[0m\u001b[1m]\u001b[0m\n", "shields:\n", "- params: null\n", @@ -709,8 +709,8 @@ " toolgroup_id: builtin::websearch\n", "- args: null\n", " mcp_endpoint: null\n", - " provider_id: memory-runtime\n", - " toolgroup_id: builtin::memory\n", + " provider_id: rag-runtime\n", + " toolgroup_id: builtin::rag\n", "- args: null\n", " mcp_endpoint: null\n", " provider_id: code-interpreter\n", diff --git a/docs/notebooks/Llama_Stack_Building_AI_Applications.ipynb b/docs/notebooks/Llama_Stack_Building_AI_Applications.ipynb index daf37ab53..58b025db4 100644 --- a/docs/notebooks/Llama_Stack_Building_AI_Applications.ipynb +++ b/docs/notebooks/Llama_Stack_Building_AI_Applications.ipynb @@ -884,8 +884,8 @@ " provider_id: code-interpreter\n", " provider_type: inline::code-interpreter\n", " - config: {}\n", - " provider_id: memory-runtime\n", - " provider_type: inline::memory-runtime\n", + " provider_id: rag-runtime\n", + " provider_type: inline::rag-runtime\n", " - config: {}\n", " provider_id: model-context-protocol\n", " provider_type: remote::model-context-protocol\n", @@ -910,8 +910,8 @@ " toolgroup_id: builtin::websearch\n", "- args: null\n", " mcp_endpoint: null\n", - " provider_id: memory-runtime\n", - " toolgroup_id: builtin::memory\n", + " provider_id: rag-runtime\n", + " toolgroup_id: builtin::rag\n", "- args: null\n", " mcp_endpoint: null\n", " provider_id: code-interpreter\n", @@ -1068,8 +1068,8 @@ " provider_id: code-interpreter\n", " provider_type: inlin\u001b[1;92me::c\u001b[0mode-interpreter\n", " - config: \u001b[1m{\u001b[0m\u001b[1m}\u001b[0m\n", - " provider_id: memory-runtime\n", - " provider_type: inline::memory-runtime\n", + " provider_id: rag-runtime\n", + " provider_type: inline::rag-runtime\n", " - config: \u001b[1m{\u001b[0m\u001b[1m}\u001b[0m\n", " provider_id: model-context-protocol\n", " provider_type: remote::model-context-protocol\n", @@ -1094,8 +1094,8 @@ " toolgroup_id: builtin::websearch\n", "- args: null\n", " mcp_endpoint: null\n", - " provider_id: memory-runtime\n", - " toolgroup_id: builtin::memory\n", + " provider_id: rag-runtime\n", + " toolgroup_id: builtin::rag\n", "- args: null\n", " mcp_endpoint: null\n", " provider_id: code-interpreter\n", @@ -1804,9 +1804,9 @@ "data": { "text/html": [ "
ToolGroup(\n", - "│ identifier='builtin::memory',\n", - "│ provider_id='memory-runtime',\n", - "│ provider_resource_id='builtin::memory',\n", + "│ identifier='builtin::rag',\n", + "│ provider_id='rag-runtime',\n", + "│ provider_resource_id='builtin::rag',\n", "│ type='tool_group',\n", "│ args=None,\n", "│ mcp_endpoint=None\n", @@ -1815,9 +1815,9 @@ ], "text/plain": [ "\u001b[1;35mToolGroup\u001b[0m\u001b[1m(\u001b[0m\n", - "\u001b[2;32m│ \u001b[0m\u001b[33midentifier\u001b[0m=\u001b[32m'builtin::memory'\u001b[0m,\n", - "\u001b[2;32m│ \u001b[0m\u001b[33mprovider_id\u001b[0m=\u001b[32m'memory-runtime'\u001b[0m,\n", - "\u001b[2;32m│ \u001b[0m\u001b[33mprovider_resource_id\u001b[0m=\u001b[32m'builtin::memory'\u001b[0m,\n", + "\u001b[2;32m│ \u001b[0m\u001b[33midentifier\u001b[0m=\u001b[32m'builtin::rag'\u001b[0m,\n", + "\u001b[2;32m│ \u001b[0m\u001b[33mprovider_id\u001b[0m=\u001b[32m'rag-runtime'\u001b[0m,\n", + "\u001b[2;32m│ \u001b[0m\u001b[33mprovider_resource_id\u001b[0m=\u001b[32m'builtin::rag'\u001b[0m,\n", "\u001b[2;32m│ \u001b[0m\u001b[33mtype\u001b[0m=\u001b[32m'tool_group'\u001b[0m,\n", "\u001b[2;32m│ \u001b[0m\u001b[33margs\u001b[0m=\u001b[3;35mNone\u001b[0m,\n", "\u001b[2;32m│ \u001b[0m\u001b[33mmcp_endpoint\u001b[0m=\u001b[3;35mNone\u001b[0m\n", @@ -2118,7 +2118,7 @@ " enable_session_persistence=False,\n", " toolgroups = [\n", " {\n", - " \"name\": \"builtin::memory\",\n", + " \"name\": \"builtin::rag\",\n", " \"args\" : {\n", " \"vector_db_ids\": [vector_db_id],\n", " }\n", diff --git a/docs/source/distributions/remote_hosted_distro/nvidia.md b/docs/source/distributions/remote_hosted_distro/nvidia.md index e4c3a155f..61b41b1d9 100644 --- a/docs/source/distributions/remote_hosted_distro/nvidia.md +++ b/docs/source/distributions/remote_hosted_distro/nvidia.md @@ -11,7 +11,7 @@ The `llamastack/distribution-nvidia` distribution consists of the following prov | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss` | diff --git a/docs/source/distributions/self_hosted_distro/bedrock.md b/docs/source/distributions/self_hosted_distro/bedrock.md index a66325560..f9a9f29cd 100644 --- a/docs/source/distributions/self_hosted_distro/bedrock.md +++ b/docs/source/distributions/self_hosted_distro/bedrock.md @@ -18,7 +18,7 @@ The `llamastack/distribution-bedrock` distribution consists of the following pro | safety | `remote::bedrock` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/cerebras.md b/docs/source/distributions/self_hosted_distro/cerebras.md index 211082b7a..a44e6287a 100644 --- a/docs/source/distributions/self_hosted_distro/cerebras.md +++ b/docs/source/distributions/self_hosted_distro/cerebras.md @@ -11,7 +11,7 @@ The `llamastack/distribution-cerebras` distribution consists of the following pr | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/fireworks.md b/docs/source/distributions/self_hosted_distro/fireworks.md index 39043b1c1..453cd746d 100644 --- a/docs/source/distributions/self_hosted_distro/fireworks.md +++ b/docs/source/distributions/self_hosted_distro/fireworks.md @@ -21,7 +21,7 @@ The `llamastack/distribution-fireworks` distribution consists of the following p | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md b/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md index 8475aab3a..a371011fe 100644 --- a/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md +++ b/docs/source/distributions/self_hosted_distro/meta-reference-gpu.md @@ -21,7 +21,7 @@ The `llamastack/distribution-meta-reference-gpu` distribution consists of the fo | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md b/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md index 6f1adb5a9..a32ccb65e 100644 --- a/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md +++ b/docs/source/distributions/self_hosted_distro/meta-reference-quantized-gpu.md @@ -21,7 +21,7 @@ The `llamastack/distribution-meta-reference-quantized-gpu` distribution consists | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/ollama.md b/docs/source/distributions/self_hosted_distro/ollama.md index f5ba31feb..b03a5ee16 100644 --- a/docs/source/distributions/self_hosted_distro/ollama.md +++ b/docs/source/distributions/self_hosted_distro/ollama.md @@ -21,7 +21,7 @@ The `llamastack/distribution-ollama` distribution consists of the following prov | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/remote-vllm.md b/docs/source/distributions/self_hosted_distro/remote-vllm.md index c2b3544d3..95dd392c1 100644 --- a/docs/source/distributions/self_hosted_distro/remote-vllm.md +++ b/docs/source/distributions/self_hosted_distro/remote-vllm.md @@ -20,7 +20,7 @@ The `llamastack/distribution-remote-vllm` distribution consists of the following | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/tgi.md b/docs/source/distributions/self_hosted_distro/tgi.md index c21a6a586..1883b926c 100644 --- a/docs/source/distributions/self_hosted_distro/tgi.md +++ b/docs/source/distributions/self_hosted_distro/tgi.md @@ -22,7 +22,7 @@ The `llamastack/distribution-tgi` distribution consists of the following provide | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/distributions/self_hosted_distro/together.md b/docs/source/distributions/self_hosted_distro/together.md index 65a711522..2d5c8fc77 100644 --- a/docs/source/distributions/self_hosted_distro/together.md +++ b/docs/source/distributions/self_hosted_distro/together.md @@ -21,7 +21,7 @@ The `llamastack/distribution-together` distribution consists of the following pr | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::memory-runtime`, `remote::model-context-protocol` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime`, `remote::model-context-protocol` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/llama_stack/distribution/ui/page/playground/rag.py b/llama_stack/distribution/ui/page/playground/rag.py index 465e11560..49991dc54 100644 --- a/llama_stack/distribution/ui/page/playground/rag.py +++ b/llama_stack/distribution/ui/page/playground/rag.py @@ -135,7 +135,7 @@ def rag_chat_page(): }, toolgroups=[ dict( - name="builtin::memory", + name="builtin::rag", args={ "vector_db_ids": [ vector_db_id for vector_db_id in selected_vector_dbs diff --git a/llama_stack/providers/inline/agents/meta_reference/agent_instance.py b/llama_stack/providers/inline/agents/meta_reference/agent_instance.py index a57b989a0..32801e514 100644 --- a/llama_stack/providers/inline/agents/meta_reference/agent_instance.py +++ b/llama_stack/providers/inline/agents/meta_reference/agent_instance.py @@ -81,7 +81,7 @@ def make_random_string(length: int = 8): TOOLS_ATTACHMENT_KEY_REGEX = re.compile(r"__tools_attachment__=(\{.*?\})") MEMORY_QUERY_TOOL = "query_from_memory" WEB_SEARCH_TOOL = "web_search" -MEMORY_GROUP = "builtin::memory" +RAG_TOOL_GROUP = "builtin::rag" class ChatAgent(ShieldRunnerMixin): @@ -391,7 +391,7 @@ class ChatAgent(ShieldRunnerMixin): session_id, documents, input_messages, tool_defs ) - if MEMORY_GROUP in toolgroups and len(input_messages) > 0: + if RAG_TOOL_GROUP in toolgroups and len(input_messages) > 0: with tracing.span(MEMORY_QUERY_TOOL) as span: step_id = str(uuid.uuid4()) yield AgentTurnResponseStreamChunk( @@ -403,7 +403,7 @@ class ChatAgent(ShieldRunnerMixin): ) ) - args = toolgroup_args.get(MEMORY_GROUP, {}) + args = toolgroup_args.get(RAG_TOOL_GROUP, {}) vector_db_ids = args.get("vector_db_ids", []) query_config = args.get("query_config") if query_config: @@ -509,7 +509,7 @@ class ChatAgent(ShieldRunnerMixin): tools=[ tool for tool in tool_defs.values() - if tool_to_group.get(tool.tool_name, None) != MEMORY_GROUP + if tool_to_group.get(tool.tool_name, None) != RAG_TOOL_GROUP ], tool_prompt_format=self.agent_config.tool_prompt_format, stream=True, @@ -756,7 +756,7 @@ class ChatAgent(ShieldRunnerMixin): for tool_def in tools.data: if ( toolgroup_name.startswith("builtin") - and toolgroup_name != MEMORY_GROUP + and toolgroup_name != RAG_TOOL_GROUP ): tool_name = tool_def.identifier built_in_type = BuiltinTool.brave_search diff --git a/llama_stack/providers/inline/agents/meta_reference/tests/test_chat_agent.py b/llama_stack/providers/inline/agents/meta_reference/tests/test_chat_agent.py index 205868279..09fccd3c6 100644 --- a/llama_stack/providers/inline/agents/meta_reference/tests/test_chat_agent.py +++ b/llama_stack/providers/inline/agents/meta_reference/tests/test_chat_agent.py @@ -152,7 +152,7 @@ class MockToolGroupsAPI: toolgroup_id=MEMORY_TOOLGROUP, tool_host=ToolHost.client, description="Mock tool", - provider_id="builtin::memory", + provider_id="builtin::rag", parameters=[], ) ] @@ -260,7 +260,7 @@ async def get_chat_agent(get_agents_impl): return await impl.get_agent(response.agent_id) -MEMORY_TOOLGROUP = "builtin::memory" +MEMORY_TOOLGROUP = "builtin::rag" CODE_INTERPRETER_TOOLGROUP = "builtin::code_interpreter" diff --git a/llama_stack/providers/registry/tool_runtime.py b/llama_stack/providers/registry/tool_runtime.py index 426fe22f2..927ca1886 100644 --- a/llama_stack/providers/registry/tool_runtime.py +++ b/llama_stack/providers/registry/tool_runtime.py @@ -19,7 +19,7 @@ def available_providers() -> List[ProviderSpec]: return [ InlineProviderSpec( api=Api.tool_runtime, - provider_type="inline::memory-runtime", + provider_type="inline::rag-runtime", pip_packages=[], module="llama_stack.providers.inline.tool_runtime.memory", config_class="llama_stack.providers.inline.tool_runtime.memory.config.MemoryToolRuntimeConfig", diff --git a/llama_stack/providers/tests/agents/test_agents.py b/llama_stack/providers/tests/agents/test_agents.py index f11aef3ec..68ee9133c 100644 --- a/llama_stack/providers/tests/agents/test_agents.py +++ b/llama_stack/providers/tests/agents/test_agents.py @@ -184,7 +184,7 @@ class TestAgents: agent_config = AgentConfig( **{ **common_params, - "toolgroups": ["builtin::memory"], + "toolgroups": ["builtin::rag"], "tool_choice": ToolChoice.auto, } ) diff --git a/llama_stack/providers/tests/tools/fixtures.py b/llama_stack/providers/tests/tools/fixtures.py index 03752881a..a2dd4239a 100644 --- a/llama_stack/providers/tests/tools/fixtures.py +++ b/llama_stack/providers/tests/tools/fixtures.py @@ -22,8 +22,8 @@ def tool_runtime_memory_and_search() -> ProviderFixture: return ProviderFixture( providers=[ Provider( - provider_id="memory-runtime", - provider_type="inline::memory-runtime", + provider_id="rag-runtime", + provider_type="inline::rag-runtime", config={}, ), Provider( @@ -47,8 +47,8 @@ def tool_runtime_memory_and_search() -> ProviderFixture: @pytest.fixture(scope="session") def tool_group_input_memory() -> ToolGroupInput: return ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ) diff --git a/llama_stack/templates/bedrock/bedrock.py b/llama_stack/templates/bedrock/bedrock.py index 20f670891..6b83e9536 100644 --- a/llama_stack/templates/bedrock/bedrock.py +++ b/llama_stack/templates/bedrock/bedrock.py @@ -29,7 +29,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -58,8 +58,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/bedrock/build.yaml b/llama_stack/templates/bedrock/build.yaml index 9ae11e9bb..6c07b0478 100644 --- a/llama_stack/templates/bedrock/build.yaml +++ b/llama_stack/templates/bedrock/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/bedrock/run.yaml b/llama_stack/templates/bedrock/run.yaml index 577263bbf..39408c1bd 100644 --- a/llama_stack/templates/bedrock/run.yaml +++ b/llama_stack/templates/bedrock/run.yaml @@ -78,8 +78,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -111,7 +111,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/cerebras/build.yaml b/llama_stack/templates/cerebras/build.yaml index 6d43ed0ca..9d5ab1a52 100644 --- a/llama_stack/templates/cerebras/build.yaml +++ b/llama_stack/templates/cerebras/build.yaml @@ -27,5 +27,5 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime image_type: conda diff --git a/llama_stack/templates/cerebras/cerebras.py b/llama_stack/templates/cerebras/cerebras.py index be51e635d..50a878645 100644 --- a/llama_stack/templates/cerebras/cerebras.py +++ b/llama_stack/templates/cerebras/cerebras.py @@ -33,7 +33,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", ], } @@ -79,8 +79,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/cerebras/run.yaml b/llama_stack/templates/cerebras/run.yaml index 0553f0749..5a70890a8 100644 --- a/llama_stack/templates/cerebras/run.yaml +++ b/llama_stack/templates/cerebras/run.yaml @@ -83,8 +83,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} metadata_store: type: sqlite @@ -113,7 +113,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/fireworks/build.yaml b/llama_stack/templates/fireworks/build.yaml index 7e19cd5e6..cdd60ec2a 100644 --- a/llama_stack/templates/fireworks/build.yaml +++ b/llama_stack/templates/fireworks/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/fireworks/fireworks.py b/llama_stack/templates/fireworks/fireworks.py index 5f1b9e8a0..546a8b82a 100644 --- a/llama_stack/templates/fireworks/fireworks.py +++ b/llama_stack/templates/fireworks/fireworks.py @@ -38,7 +38,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -86,8 +86,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/fireworks/run-with-safety.yaml b/llama_stack/templates/fireworks/run-with-safety.yaml index 659ec5191..a4b425436 100644 --- a/llama_stack/templates/fireworks/run-with-safety.yaml +++ b/llama_stack/templates/fireworks/run-with-safety.yaml @@ -89,8 +89,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -168,7 +168,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/fireworks/run.yaml b/llama_stack/templates/fireworks/run.yaml index 9fb61f842..a497317bd 100644 --- a/llama_stack/templates/fireworks/run.yaml +++ b/llama_stack/templates/fireworks/run.yaml @@ -83,8 +83,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -157,7 +157,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/hf-endpoint/build.yaml b/llama_stack/templates/hf-endpoint/build.yaml index 82a460bd9..c2eaaa05b 100644 --- a/llama_stack/templates/hf-endpoint/build.yaml +++ b/llama_stack/templates/hf-endpoint/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/hf-endpoint/hf_endpoint.py b/llama_stack/templates/hf-endpoint/hf_endpoint.py index f9bfe85f9..4533fd95b 100644 --- a/llama_stack/templates/hf-endpoint/hf_endpoint.py +++ b/llama_stack/templates/hf-endpoint/hf_endpoint.py @@ -33,7 +33,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -76,8 +76,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/hf-endpoint/run-with-safety.yaml b/llama_stack/templates/hf-endpoint/run-with-safety.yaml index dfa094fe6..0329f580b 100644 --- a/llama_stack/templates/hf-endpoint/run-with-safety.yaml +++ b/llama_stack/templates/hf-endpoint/run-with-safety.yaml @@ -88,8 +88,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -120,7 +120,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/hf-endpoint/run.yaml b/llama_stack/templates/hf-endpoint/run.yaml index fb5d7fa31..8163fe28e 100644 --- a/llama_stack/templates/hf-endpoint/run.yaml +++ b/llama_stack/templates/hf-endpoint/run.yaml @@ -83,8 +83,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -110,7 +110,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/hf-serverless/build.yaml b/llama_stack/templates/hf-serverless/build.yaml index 0eb4e0509..f9303cfab 100644 --- a/llama_stack/templates/hf-serverless/build.yaml +++ b/llama_stack/templates/hf-serverless/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/hf-serverless/hf_serverless.py b/llama_stack/templates/hf-serverless/hf_serverless.py index 4f3c29404..8438de7a5 100644 --- a/llama_stack/templates/hf-serverless/hf_serverless.py +++ b/llama_stack/templates/hf-serverless/hf_serverless.py @@ -33,7 +33,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -77,8 +77,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/hf-serverless/run-with-safety.yaml b/llama_stack/templates/hf-serverless/run-with-safety.yaml index 0575efaef..9cee920a5 100644 --- a/llama_stack/templates/hf-serverless/run-with-safety.yaml +++ b/llama_stack/templates/hf-serverless/run-with-safety.yaml @@ -88,8 +88,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -120,7 +120,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/hf-serverless/run.yaml b/llama_stack/templates/hf-serverless/run.yaml index b87edd744..c8ad0d38d 100644 --- a/llama_stack/templates/hf-serverless/run.yaml +++ b/llama_stack/templates/hf-serverless/run.yaml @@ -83,8 +83,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -110,7 +110,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/meta-reference-gpu/build.yaml b/llama_stack/templates/meta-reference-gpu/build.yaml index f5371f0d6..b9130fc7d 100644 --- a/llama_stack/templates/meta-reference-gpu/build.yaml +++ b/llama_stack/templates/meta-reference-gpu/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/meta-reference-gpu/meta_reference.py b/llama_stack/templates/meta-reference-gpu/meta_reference.py index dae4f0218..a3f82b0c8 100644 --- a/llama_stack/templates/meta-reference-gpu/meta_reference.py +++ b/llama_stack/templates/meta-reference-gpu/meta_reference.py @@ -37,7 +37,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -83,8 +83,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml b/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml index 54ddef155..0faaabb15 100644 --- a/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml +++ b/llama_stack/templates/meta-reference-gpu/run-with-safety.yaml @@ -90,8 +90,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -122,7 +122,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/meta-reference-gpu/run.yaml b/llama_stack/templates/meta-reference-gpu/run.yaml index cde581d19..6ffe1fa36 100644 --- a/llama_stack/templates/meta-reference-gpu/run.yaml +++ b/llama_stack/templates/meta-reference-gpu/run.yaml @@ -84,8 +84,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -111,7 +111,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/meta-reference-quantized-gpu/build.yaml b/llama_stack/templates/meta-reference-quantized-gpu/build.yaml index aa23ad313..7bbcfe5f2 100644 --- a/llama_stack/templates/meta-reference-quantized-gpu/build.yaml +++ b/llama_stack/templates/meta-reference-quantized-gpu/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py b/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py index 4e9cbf1fe..8c2a6ec9f 100644 --- a/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py +++ b/llama_stack/templates/meta-reference-quantized-gpu/meta_reference.py @@ -32,7 +32,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -42,8 +42,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/meta-reference-quantized-gpu/run.yaml b/llama_stack/templates/meta-reference-quantized-gpu/run.yaml index cc5793f8f..5ff87a901 100644 --- a/llama_stack/templates/meta-reference-quantized-gpu/run.yaml +++ b/llama_stack/templates/meta-reference-quantized-gpu/run.yaml @@ -86,8 +86,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -113,7 +113,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/nvidia/build.yaml b/llama_stack/templates/nvidia/build.yaml index d6a510e2e..e9748721a 100644 --- a/llama_stack/templates/nvidia/build.yaml +++ b/llama_stack/templates/nvidia/build.yaml @@ -25,6 +25,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/nvidia/nvidia.py b/llama_stack/templates/nvidia/nvidia.py index 5693ba12d..19eb4bd5d 100644 --- a/llama_stack/templates/nvidia/nvidia.py +++ b/llama_stack/templates/nvidia/nvidia.py @@ -28,7 +28,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -56,8 +56,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/nvidia/run.yaml b/llama_stack/templates/nvidia/run.yaml index 317aa1031..c57ca2b9a 100644 --- a/llama_stack/templates/nvidia/run.yaml +++ b/llama_stack/templates/nvidia/run.yaml @@ -80,8 +80,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -143,7 +143,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/ollama/build.yaml b/llama_stack/templates/ollama/build.yaml index c3ed88fb8..0fee6808c 100644 --- a/llama_stack/templates/ollama/build.yaml +++ b/llama_stack/templates/ollama/build.yaml @@ -27,5 +27,5 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime image_type: conda diff --git a/llama_stack/templates/ollama/ollama.py b/llama_stack/templates/ollama/ollama.py index bdbd1e142..d14cb3aad 100644 --- a/llama_stack/templates/ollama/ollama.py +++ b/llama_stack/templates/ollama/ollama.py @@ -35,7 +35,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", ], } name = "ollama" @@ -77,8 +77,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/ollama/run-with-safety.yaml b/llama_stack/templates/ollama/run-with-safety.yaml index afb0b1938..5b5c9c253 100644 --- a/llama_stack/templates/ollama/run-with-safety.yaml +++ b/llama_stack/templates/ollama/run-with-safety.yaml @@ -85,8 +85,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} metadata_store: type: sqlite @@ -117,7 +117,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/ollama/run.yaml b/llama_stack/templates/ollama/run.yaml index 976068670..3cc1cb2ac 100644 --- a/llama_stack/templates/ollama/run.yaml +++ b/llama_stack/templates/ollama/run.yaml @@ -82,8 +82,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} metadata_store: type: sqlite @@ -106,7 +106,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/remote-vllm/build.yaml b/llama_stack/templates/remote-vllm/build.yaml index 409b2ba10..74d9f32d9 100644 --- a/llama_stack/templates/remote-vllm/build.yaml +++ b/llama_stack/templates/remote-vllm/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/remote-vllm/run-with-safety.yaml b/llama_stack/templates/remote-vllm/run-with-safety.yaml index e26d0f99f..4a0fa9a85 100644 --- a/llama_stack/templates/remote-vllm/run-with-safety.yaml +++ b/llama_stack/templates/remote-vllm/run-with-safety.yaml @@ -90,8 +90,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -122,7 +122,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/remote-vllm/run.yaml b/llama_stack/templates/remote-vllm/run.yaml index dc54d216d..9631f94a2 100644 --- a/llama_stack/templates/remote-vllm/run.yaml +++ b/llama_stack/templates/remote-vllm/run.yaml @@ -84,8 +84,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -111,7 +111,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/remote-vllm/vllm.py b/llama_stack/templates/remote-vllm/vllm.py index f91ad24a7..6c835ef86 100644 --- a/llama_stack/templates/remote-vllm/vllm.py +++ b/llama_stack/templates/remote-vllm/vllm.py @@ -35,7 +35,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -80,8 +80,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/tgi/build.yaml b/llama_stack/templates/tgi/build.yaml index bc31ef7e7..8bc628158 100644 --- a/llama_stack/templates/tgi/build.yaml +++ b/llama_stack/templates/tgi/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/tgi/run-with-safety.yaml b/llama_stack/templates/tgi/run-with-safety.yaml index ea8057137..503505c32 100644 --- a/llama_stack/templates/tgi/run-with-safety.yaml +++ b/llama_stack/templates/tgi/run-with-safety.yaml @@ -83,8 +83,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -110,7 +110,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/tgi/run.yaml b/llama_stack/templates/tgi/run.yaml index d537d0fce..f1953c513 100644 --- a/llama_stack/templates/tgi/run.yaml +++ b/llama_stack/templates/tgi/run.yaml @@ -82,8 +82,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -109,7 +109,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/tgi/tgi.py b/llama_stack/templates/tgi/tgi.py index 230fcac2a..e49c98d72 100644 --- a/llama_stack/templates/tgi/tgi.py +++ b/llama_stack/templates/tgi/tgi.py @@ -35,7 +35,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -80,8 +80,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/together/build.yaml b/llama_stack/templates/together/build.yaml index 2160adb8e..90ee5bcee 100644 --- a/llama_stack/templates/together/build.yaml +++ b/llama_stack/templates/together/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/together/run-with-safety.yaml b/llama_stack/templates/together/run-with-safety.yaml index 54b918eea..ec351108e 100644 --- a/llama_stack/templates/together/run-with-safety.yaml +++ b/llama_stack/templates/together/run-with-safety.yaml @@ -89,8 +89,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -163,7 +163,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/together/run.yaml b/llama_stack/templates/together/run.yaml index 2c0475796..c2afd98e9 100644 --- a/llama_stack/templates/together/run.yaml +++ b/llama_stack/templates/together/run.yaml @@ -83,8 +83,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -152,7 +152,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/together/together.py b/llama_stack/templates/together/together.py index ec64527d2..5e9520433 100644 --- a/llama_stack/templates/together/together.py +++ b/llama_stack/templates/together/together.py @@ -38,7 +38,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -76,8 +76,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/llama_stack/templates/vllm-gpu/build.yaml b/llama_stack/templates/vllm-gpu/build.yaml index 45f543071..d24046613 100644 --- a/llama_stack/templates/vllm-gpu/build.yaml +++ b/llama_stack/templates/vllm-gpu/build.yaml @@ -27,6 +27,6 @@ distribution_spec: - remote::brave-search - remote::tavily-search - inline::code-interpreter - - inline::memory-runtime + - inline::rag-runtime - remote::model-context-protocol image_type: conda diff --git a/llama_stack/templates/vllm-gpu/run.yaml b/llama_stack/templates/vllm-gpu/run.yaml index 2d9ec6a3f..165e4d51d 100644 --- a/llama_stack/templates/vllm-gpu/run.yaml +++ b/llama_stack/templates/vllm-gpu/run.yaml @@ -86,8 +86,8 @@ providers: - provider_id: code-interpreter provider_type: inline::code-interpreter config: {} - - provider_id: memory-runtime - provider_type: inline::memory-runtime + - provider_id: rag-runtime + provider_type: inline::rag-runtime config: {} - provider_id: model-context-protocol provider_type: remote::model-context-protocol @@ -113,7 +113,7 @@ eval_tasks: [] tool_groups: - toolgroup_id: builtin::websearch provider_id: tavily-search -- toolgroup_id: builtin::memory - provider_id: memory-runtime +- toolgroup_id: builtin::rag + provider_id: rag-runtime - toolgroup_id: builtin::code_interpreter provider_id: code-interpreter diff --git a/llama_stack/templates/vllm-gpu/vllm.py b/llama_stack/templates/vllm-gpu/vllm.py index a8f13ce40..54ebd2d41 100644 --- a/llama_stack/templates/vllm-gpu/vllm.py +++ b/llama_stack/templates/vllm-gpu/vllm.py @@ -32,7 +32,7 @@ def get_distribution_template() -> DistributionTemplate: "remote::brave-search", "remote::tavily-search", "inline::code-interpreter", - "inline::memory-runtime", + "inline::rag-runtime", "remote::model-context-protocol", ], } @@ -72,8 +72,8 @@ def get_distribution_template() -> DistributionTemplate: provider_id="tavily-search", ), ToolGroupInput( - toolgroup_id="builtin::memory", - provider_id="memory-runtime", + toolgroup_id="builtin::rag", + provider_id="rag-runtime", ), ToolGroupInput( toolgroup_id="builtin::code_interpreter", diff --git a/tests/client-sdk/agents/test_agents.py b/tests/client-sdk/agents/test_agents.py index 7c13f5768..c6be91232 100644 --- a/tests/client-sdk/agents/test_agents.py +++ b/tests/client-sdk/agents/test_agents.py @@ -292,7 +292,7 @@ def test_rag_agent(llama_stack_client, agent_config): **agent_config, "toolgroups": [ dict( - name="builtin::memory", + name="builtin::rag", args={ "vector_db_ids": [vector_db_id], },