From 77c10352253340bcd2c49a202144c056fb8564c7 Mon Sep 17 00:00:00 2001 From: Swapna Lekkala Date: Tue, 9 Sep 2025 11:32:39 -0700 Subject: [PATCH] fix docs --- .../langchain/Llama_Stack_LangChain.ipynb | 110 +++--------------- 1 file changed, 17 insertions(+), 93 deletions(-) diff --git a/docs/notebooks/langchain/Llama_Stack_LangChain.ipynb b/docs/notebooks/langchain/Llama_Stack_LangChain.ipynb index fc3ffc037..d44ac6994 100644 --- a/docs/notebooks/langchain/Llama_Stack_LangChain.ipynb +++ b/docs/notebooks/langchain/Llama_Stack_LangChain.ipynb @@ -55,7 +55,7 @@ "text": [ "Requirement already satisfied: uv in /Users/swapna942/miniconda3/lib/python3.12/site-packages (0.7.20)\n", "\u001b[2mUsing Python 3.12.11 environment at: /Users/swapna942/miniconda3\u001b[0m\n", - "\u001b[2mAudited \u001b[1m7 packages\u001b[0m \u001b[2min 94ms\u001b[0m\u001b[0m\n" + "\u001b[2mAudited \u001b[1m7 packages\u001b[0m \u001b[2min 42ms\u001b[0m\u001b[0m\n" ] } ], @@ -154,8 +154,8 @@ "name": "stdout", "output_type": "stream", "text": [ - "Building and starting Llama Stack server with PID: 81294\n", - "Waiting for server to start......\n", + "Building and starting Llama Stack server with PID: 19747\n", + "Waiting for server to start....\n", "Server is ready!\n" ] } @@ -188,7 +188,7 @@ "output_type": "stream", "text": [ "\u001b[2mUsing Python 3.12.11 environment at: /Users/swapna942/miniconda3\u001b[0m\n", - "\u001b[2mAudited \u001b[1m1 package\u001b[0m \u001b[2min 26ms\u001b[0m\u001b[0m\n" + "\u001b[2mAudited \u001b[1m1 package\u001b[0m \u001b[2min 27ms\u001b[0m\u001b[0m\n" ] } ], @@ -254,27 +254,7 @@ "name": "stdout", "output_type": "stream", "text": [ - "Available models:\n", - "- all-minilm\n", - "- nvidia/meta/llama-3.1-405b-instruct\n", - "- nvidia/meta/llama-3.1-70b-instruct\n", - "- nvidia/meta/llama-3.1-8b-instruct\n", - "- nvidia/meta/llama-3.2-11b-vision-instruct\n", - "- nvidia/meta/llama-3.2-1b-instruct\n", - "- nvidia/meta/llama-3.2-3b-instruct\n", - "- nvidia/meta/llama-3.2-90b-vision-instruct\n", - "- nvidia/meta/llama-3.3-70b-instruct\n", - "- nvidia/meta/llama3-70b-instruct\n", - "- nvidia/meta/llama3-8b-instruct\n", - "- nvidia/nvidia/llama-3.2-nv-embedqa-1b-v2\n", - "- nvidia/nvidia/nv-embedqa-e5-v5\n", - "- nvidia/nvidia/nv-embedqa-mistral-7b-v2\n", - "- nvidia/snowflake/arctic-embed-l\n", - "- ollama/all-minilm:l6-v2\n", - "- ollama/llama-guard3:1b\n", - "- ollama/llama-guard3:8b\n", - "- ollama/llama3.2:3b-instruct-fp16\n", - "- ollama/nomic-embed-text\n", + "Available Fireworks models:\n", "- fireworks/accounts/fireworks/models/llama-v3p1-8b-instruct\n", "- fireworks/accounts/fireworks/models/llama-v3p1-70b-instruct\n", "- fireworks/accounts/fireworks/models/llama-v3p1-405b-instruct\n", @@ -287,63 +267,6 @@ "- fireworks/nomic-ai/nomic-embed-text-v1.5\n", "- fireworks/accounts/fireworks/models/llama-guard-3-8b\n", "- fireworks/accounts/fireworks/models/llama-guard-3-11b-vision\n", - "- together/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo\n", - "- together/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo\n", - "- together/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo\n", - "- together/meta-llama/Llama-3.2-3B-Instruct-Turbo\n", - "- together/meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo\n", - "- together/meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo\n", - "- together/meta-llama/Llama-3.3-70B-Instruct-Turbo\n", - "- together/togethercomputer/m2-bert-80M-8k-retrieval\n", - "- together/togethercomputer/m2-bert-80M-32k-retrieval\n", - "- together/meta-llama/Llama-4-Scout-17B-16E-Instruct\n", - "- together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8\n", - "- together/meta-llama/Llama-Guard-3-8B\n", - "- together/meta-llama/Llama-Guard-3-11B-Vision-Turbo\n", - "- bedrock/meta.llama3-1-8b-instruct-v1:0\n", - "- bedrock/meta.llama3-1-70b-instruct-v1:0\n", - "- bedrock/meta.llama3-1-405b-instruct-v1:0\n", - "- openai/gpt-3.5-turbo-0125\n", - "- openai/gpt-3.5-turbo\n", - "- openai/gpt-3.5-turbo-instruct\n", - "- openai/gpt-4\n", - "- openai/gpt-4-turbo\n", - "- openai/gpt-4o\n", - "- openai/gpt-4o-2024-08-06\n", - "- openai/gpt-4o-mini\n", - "- openai/gpt-4o-audio-preview\n", - "- openai/chatgpt-4o-latest\n", - "- openai/o1\n", - "- openai/o1-mini\n", - "- openai/o3-mini\n", - "- openai/o4-mini\n", - "- openai/text-embedding-3-small\n", - "- openai/text-embedding-3-large\n", - "- anthropic/claude-3-5-sonnet-latest\n", - "- anthropic/claude-3-7-sonnet-latest\n", - "- anthropic/claude-3-5-haiku-latest\n", - "- anthropic/voyage-3\n", - "- anthropic/voyage-3-lite\n", - "- anthropic/voyage-code-3\n", - "- gemini/gemini-1.5-flash\n", - "- gemini/gemini-1.5-pro\n", - "- gemini/gemini-2.0-flash\n", - "- gemini/gemini-2.0-flash-lite\n", - "- gemini/gemini-2.5-flash\n", - "- gemini/gemini-2.5-flash-lite\n", - "- gemini/gemini-2.5-pro\n", - "- gemini/text-embedding-004\n", - "- groq/llama3-8b-8192\n", - "- groq/llama-3.1-8b-instant\n", - "- groq/llama3-70b-8192\n", - "- groq/llama-3.3-70b-versatile\n", - "- groq/llama-3.2-3b-preview\n", - "- groq/meta-llama/llama-4-scout-17b-16e-instruct\n", - "- groq/meta-llama/llama-4-maverick-17b-128e-instruct\n", - "- sambanova/Meta-Llama-3.1-8B-Instruct\n", - "- sambanova/Meta-Llama-3.3-70B-Instruct\n", - "- sambanova/Llama-4-Maverick-17B-128E-Instruct\n", - "- sentence-transformers/all-MiniLM-L6-v2\n", "----\n", "Available shields (safety models):\n", "code-scanner\n", @@ -354,9 +277,10 @@ } ], "source": [ - "print(\"Available models:\")\n", + "print(\"Available Fireworks models:\")\n", "for m in client.models.list():\n", - " print(f\"- {m.identifier}\")\n", + " if m.identifier.startswith(\"fireworks/\"):\n", + " print(f\"- {m.identifier}\")\n", "\n", "print(\"----\")\n", "print(\"Available shields (safety models):\")\n", @@ -401,9 +325,9 @@ "name": "stdout", "output_type": "stream", "text": [ - "File(id='file-33012cc7c18a4430ad1a11b3397c3e29', bytes=41, created_at=1757441277, expires_at=1788977277, filename='shipping_policy.txt', object='file', purpose='assistants')\n", - "File(id='file-e7692c2f044e44859ea02b0630195b4e', bytes=48, created_at=1757441277, expires_at=1788977277, filename='returns_policy.txt', object='file', purpose='assistants')\n", - "File(id='file-9ff178cbb72b4e79a7513ce5a95a6fe1', bytes=45, created_at=1757441277, expires_at=1788977277, filename='support.txt', object='file', purpose='assistants')\n" + "File(id='file-54652c95c56c4c34918a97d7ff8a4320', bytes=41, created_at=1757442621, expires_at=1788978621, filename='shipping_policy.txt', object='file', purpose='assistants')\n", + "File(id='file-fb1227c1d1854da1bd774d21e5b7e41c', bytes=48, created_at=1757442621, expires_at=1788978621, filename='returns_policy.txt', object='file', purpose='assistants')\n", + "File(id='file-673f874852fe42798675a13d06a256e2', bytes=45, created_at=1757442621, expires_at=1788978621, filename='support.txt', object='file', purpose='assistants')\n" ] }, { @@ -461,7 +385,7 @@ "name": "stderr", "output_type": "stream", "text": [ - "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_d9fa50e8-3ccc-4db3-9102-1a71cd62ab64/search \"HTTP/1.1 200 OK\"\n" + "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_708c060b-45da-423e-8354-68529b4fd1a6/search \"HTTP/1.1 200 OK\"\n" ] }, { @@ -548,7 +472,7 @@ { "data": { "text/plain": [ - "AIMessage(content='Silent llama, soft and slow,\\n Majestic creature, with a gentle glow.', additional_kwargs={'refusal': None}, response_metadata={'token_usage': None, 'model_name': 'fireworks/accounts/fireworks/models/llama-v3p1-8b-instruct', 'system_fingerprint': None, 'id': 'chatcmpl-8957d071-9f1a-4a7f-9c0e-d516e02194d1', 'service_tier': None, 'finish_reason': 'stop', 'logprobs': None}, id='run--f4331df3-e787-4973-b2e8-7ca90d4e19df-0')" + "AIMessage(content=\"A llama's gentle eyes shine bright,\\nIn the Andes, it roams through morning light.\", additional_kwargs={'refusal': None}, response_metadata={'token_usage': None, 'model_name': 'fireworks/accounts/fireworks/models/llama-v3p1-8b-instruct', 'system_fingerprint': None, 'id': 'chatcmpl-602b5967-82a3-476b-9cd2-7d3b29b76ee8', 'service_tier': None, 'finish_reason': 'stop', 'logprobs': None}, id='run--0933c465-ff4d-4a7b-b7fb-fd97dd8244f3-0')" ] }, "execution_count": 10, @@ -646,7 +570,7 @@ "name": "stderr", "output_type": "stream", "text": [ - "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_d9fa50e8-3ccc-4db3-9102-1a71cd62ab64/search \"HTTP/1.1 200 OK\"\n", + "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_708c060b-45da-423e-8354-68529b4fd1a6/search \"HTTP/1.1 200 OK\"\n", "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/chat/completions \"HTTP/1.1 200 OK\"\n" ] }, @@ -655,7 +579,7 @@ "output_type": "stream", "text": [ "❓ How long does shipping take?\n", - "💡 According to the provided context, Acme ships globally in 3-5 business days.\n" + "💡 Acme ships globally in 3-5 business days. This means that shipping typically takes between 3 to 5 working days from the date of dispatch or order fulfillment.\n" ] } ], @@ -684,7 +608,7 @@ "name": "stderr", "output_type": "stream", "text": [ - "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_d9fa50e8-3ccc-4db3-9102-1a71cd62ab64/search \"HTTP/1.1 200 OK\"\n", + "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_708c060b-45da-423e-8354-68529b4fd1a6/search \"HTTP/1.1 200 OK\"\n", "INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/chat/completions \"HTTP/1.1 200 OK\"\n" ] }, @@ -693,7 +617,7 @@ "output_type": "stream", "text": [ "❓ Can I return a product after 40 days?\n", - "💡 Based on the given context, returns are accepted within 30 days of purchase. Therefore, after 40 days, returning a product is not accepted according to the given policy.\n" + "💡 Based on the provided context, you cannot return a product after 40 days. The return window is limited to 30 days from the date of purchase.\n" ] } ],