diff --git a/docs/getting_started.ipynb b/docs/getting_started.ipynb index b764d4d34..cdaf074b8 100644 --- a/docs/getting_started.ipynb +++ b/docs/getting_started.ipynb @@ -1050,8 +1050,6 @@ "text/html": [ "
ToolGroup(\n", "│ identifier='builtin::code_interpreter',\n", - "│ provider_id='code-interpreter',\n", - "│ provider_resource_id='builtin::code_interpreter',\n", "│ type='tool_group',\n", "│ args=None,\n", "│ mcp_endpoint=None\n", @@ -1061,7 +1059,6 @@ "text/plain": [ "\u001b[1;35mToolGroup\u001b[0m\u001b[1m(\u001b[0m\n", "\u001b[2;32m│ \u001b[0m\u001b[33midentifier\u001b[0m=\u001b[32m'builtin::code_interpreter'\u001b[0m,\n", - "\u001b[2;32m│ \u001b[0m\u001b[33mprovider_id\u001b[0m=\u001b[32m'code-interpreter'\u001b[0m,\n", "\u001b[2;32m│ \u001b[0m\u001b[33mprovider_resource_id\u001b[0m=\u001b[32m'builtin::code_interpreter'\u001b[0m,\n", "\u001b[2;32m│ \u001b[0m\u001b[33mtype\u001b[0m=\u001b[32m'tool_group'\u001b[0m,\n", "\u001b[2;32m│ \u001b[0m\u001b[33margs\u001b[0m=\u001b[3;35mNone\u001b[0m,\n", diff --git a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb index 5de7f715e..413b693d1 100644 --- a/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb +++ b/docs/notebooks/Llama_Stack_Benchmark_Evals.ipynb @@ -337,9 +337,6 @@ " provider_id: tavily-search\n", " provider_type: remote::tavily-search\n", " - config: {}\n", - " provider_id: code-interpreter\n", - " provider_type: inline::code-interpreter\n", - " - config: {}\n", " provider_id: rag-runtime\n", " provider_type: inline::rag-runtime\n", " - config: {}\n", @@ -378,10 +375,6 @@ " toolgroup_id: builtin::rag\n", "- args: null\n", " mcp_endpoint: null\n", - " provider_id: code-interpreter\n", - " toolgroup_id: builtin::code_interpreter\n", - "- args: null\n", - " mcp_endpoint: null\n", " provider_id: wolfram-alpha\n", " toolgroup_id: builtin::wolfram_alpha\n", "vector_dbs: []\n", @@ -617,9 +610,6 @@ " provider_id: tavily-search\n", " provider_type: remote::tavily-search\n", " - config: \u001b[1m{\u001b[0m\u001b[1m}\u001b[0m\n", - " provider_id: code-interpreter\n", - " provider_type: inlin\u001b[1;92me::c\u001b[0mode-interpreter\n", - " - config: \u001b[1m{\u001b[0m\u001b[1m}\u001b[0m\n", " provider_id: rag-runtime\n", " provider_type: inline::rag-runtime\n", " - config: \u001b[1m{\u001b[0m\u001b[1m}\u001b[0m\n", @@ -658,10 +648,6 @@ " toolgroup_id: builtin::rag\n", "- args: null\n", " mcp_endpoint: null\n", - " provider_id: code-interpreter\n", - " toolgroup_id: builtin::code_interpreter\n", - "- args: null\n", - " mcp_endpoint: null\n", " provider_id: wolfram-alpha\n", " toolgroup_id: builtin::wolfram_alpha\n", "vector_dbs: \u001b[1m[\u001b[0m\u001b[1m]\u001b[0m\n", diff --git a/docs/source/building_applications/tools.md b/docs/source/building_applications/tools.md index 95c69ffa3..c7af17bfa 100644 --- a/docs/source/building_applications/tools.md +++ b/docs/source/building_applications/tools.md @@ -165,34 +165,6 @@ all_tools = client.tools.list_tools() group_tools = client.tools.list_tools(toolgroup_id="search_tools") ``` -## Simple Example: Using an Agent with the Code-Interpreter Tool - -```python -from llama_stack_client import Agent - -# Instantiate the AI agent with the given configuration -agent = Agent( - client, - name="code-interpreter", - description="A code interpreter agent for executing Python code snippets", - instructions=""" - You are a highly reliable, concise, and precise assistant. - Always show the generated code, never generate your own code, and never anticipate results. - """, - model="meta-llama/Llama-3.2-3B-Instruct", - tools=["builtin::code_interpreter"], - max_infer_iters=5, -) - -# Start a session -session_id = agent.create_session("tool_session") - -# Send a query to the AI agent for code execution -response = agent.create_turn( - messages=[{"role": "user", "content": "Run this code: print(3 ** 4 - 5 * 2)"}], - session_id=session_id, -) -``` ## Simple Example 2: Using an Agent with the Web Search Tool 1. Start by registering a Tavily API key at [Tavily](https://tavily.com/). 2. [Optional] Provide the API key directly to the Llama Stack server diff --git a/docs/source/distributions/self_hosted_distro/dell.md b/docs/source/distributions/self_hosted_distro/dell.md index 96b0ef478..2e987985c 100644 --- a/docs/source/distributions/self_hosted_distro/dell.md +++ b/docs/source/distributions/self_hosted_distro/dell.md @@ -23,7 +23,7 @@ The `llamastack/distribution-dell` distribution consists of the following provid | safety | `inline::llama-guard` | | scoring | `inline::basic`, `inline::llm-as-judge`, `inline::braintrust` | | telemetry | `inline::meta-reference` | -| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::code-interpreter`, `inline::rag-runtime` | +| tool_runtime | `remote::brave-search`, `remote::tavily-search`, `inline::rag-runtime` | | vector_io | `inline::faiss`, `remote::chromadb`, `remote::pgvector` | diff --git a/docs/source/references/llama_stack_client_cli_reference.md b/docs/source/references/llama_stack_client_cli_reference.md index 0b84027f0..cd4dd4cd7 100644 --- a/docs/source/references/llama_stack_client_cli_reference.md +++ b/docs/source/references/llama_stack_client_cli_reference.md @@ -253,8 +253,6 @@ llama-stack-client toolgroups list +---------------------------+------------------+------+---------------+ | identifier | provider_id | args | mcp_endpoint | +===========================+==================+======+===============+ -| builtin::code_interpreter | code-interpreter | None | None | -+---------------------------+------------------+------+---------------+ | builtin::rag | rag-runtime | None | None | +---------------------------+------------------+------+---------------+ | builtin::websearch | tavily-search | None | None | diff --git a/tests/external-provider/llama-stack-provider-ollama/run.yaml b/tests/external-provider/llama-stack-provider-ollama/run.yaml index 5afeb1448..666189f03 100644 --- a/tests/external-provider/llama-stack-provider-ollama/run.yaml +++ b/tests/external-provider/llama-stack-provider-ollama/run.yaml @@ -53,9 +53,6 @@ providers: config: api_key: ${env.TAVILY_SEARCH_API_KEY:} max_results: 3 - - provider_id: code-interpreter - provider_type: inline::code-interpreter - config: {} - provider_id: rag-runtime provider_type: inline::rag-runtime config: {} @@ -90,8 +87,6 @@ tool_groups: provider_id: tavily-search - toolgroup_id: builtin::rag provider_id: rag-runtime -- toolgroup_id: builtin::code_interpreter - provider_id: code-interpreter - toolgroup_id: builtin::wolfram_alpha provider_id: wolfram-alpha server: diff --git a/tests/verifications/openai-api-verification-run.yaml b/tests/verifications/openai-api-verification-run.yaml index d80aa3c75..4c322af28 100644 --- a/tests/verifications/openai-api-verification-run.yaml +++ b/tests/verifications/openai-api-verification-run.yaml @@ -74,9 +74,6 @@ providers: config: api_key: ${env.TAVILY_SEARCH_API_KEY:} max_results: 3 - - provider_id: code-interpreter - provider_type: inline::code-interpreter - config: {} - provider_id: rag-runtime provider_type: inline::rag-runtime config: {} @@ -156,8 +153,6 @@ tool_groups: provider_id: tavily-search - toolgroup_id: builtin::rag provider_id: rag-runtime -- toolgroup_id: builtin::code_interpreter - provider_id: code-interpreter - toolgroup_id: builtin::wolfram_alpha provider_id: wolfram-alpha server: