mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-05 12:21:52 +00:00
fix docs
This commit is contained in:
parent
01281277af
commit
77c1035225
1 changed files with 17 additions and 93 deletions
|
@ -55,7 +55,7 @@
|
|||
"text": [
|
||||
"Requirement already satisfied: uv in /Users/swapna942/miniconda3/lib/python3.12/site-packages (0.7.20)\n",
|
||||
"\u001b[2mUsing Python 3.12.11 environment at: /Users/swapna942/miniconda3\u001b[0m\n",
|
||||
"\u001b[2mAudited \u001b[1m7 packages\u001b[0m \u001b[2min 94ms\u001b[0m\u001b[0m\n"
|
||||
"\u001b[2mAudited \u001b[1m7 packages\u001b[0m \u001b[2min 42ms\u001b[0m\u001b[0m\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
|
@ -154,8 +154,8 @@
|
|||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"Building and starting Llama Stack server with PID: 81294\n",
|
||||
"Waiting for server to start......\n",
|
||||
"Building and starting Llama Stack server with PID: 19747\n",
|
||||
"Waiting for server to start....\n",
|
||||
"Server is ready!\n"
|
||||
]
|
||||
}
|
||||
|
@ -188,7 +188,7 @@
|
|||
"output_type": "stream",
|
||||
"text": [
|
||||
"\u001b[2mUsing Python 3.12.11 environment at: /Users/swapna942/miniconda3\u001b[0m\n",
|
||||
"\u001b[2mAudited \u001b[1m1 package\u001b[0m \u001b[2min 26ms\u001b[0m\u001b[0m\n"
|
||||
"\u001b[2mAudited \u001b[1m1 package\u001b[0m \u001b[2min 27ms\u001b[0m\u001b[0m\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
|
@ -254,27 +254,7 @@
|
|||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"Available models:\n",
|
||||
"- all-minilm\n",
|
||||
"- nvidia/meta/llama-3.1-405b-instruct\n",
|
||||
"- nvidia/meta/llama-3.1-70b-instruct\n",
|
||||
"- nvidia/meta/llama-3.1-8b-instruct\n",
|
||||
"- nvidia/meta/llama-3.2-11b-vision-instruct\n",
|
||||
"- nvidia/meta/llama-3.2-1b-instruct\n",
|
||||
"- nvidia/meta/llama-3.2-3b-instruct\n",
|
||||
"- nvidia/meta/llama-3.2-90b-vision-instruct\n",
|
||||
"- nvidia/meta/llama-3.3-70b-instruct\n",
|
||||
"- nvidia/meta/llama3-70b-instruct\n",
|
||||
"- nvidia/meta/llama3-8b-instruct\n",
|
||||
"- nvidia/nvidia/llama-3.2-nv-embedqa-1b-v2\n",
|
||||
"- nvidia/nvidia/nv-embedqa-e5-v5\n",
|
||||
"- nvidia/nvidia/nv-embedqa-mistral-7b-v2\n",
|
||||
"- nvidia/snowflake/arctic-embed-l\n",
|
||||
"- ollama/all-minilm:l6-v2\n",
|
||||
"- ollama/llama-guard3:1b\n",
|
||||
"- ollama/llama-guard3:8b\n",
|
||||
"- ollama/llama3.2:3b-instruct-fp16\n",
|
||||
"- ollama/nomic-embed-text\n",
|
||||
"Available Fireworks models:\n",
|
||||
"- fireworks/accounts/fireworks/models/llama-v3p1-8b-instruct\n",
|
||||
"- fireworks/accounts/fireworks/models/llama-v3p1-70b-instruct\n",
|
||||
"- fireworks/accounts/fireworks/models/llama-v3p1-405b-instruct\n",
|
||||
|
@ -287,63 +267,6 @@
|
|||
"- fireworks/nomic-ai/nomic-embed-text-v1.5\n",
|
||||
"- fireworks/accounts/fireworks/models/llama-guard-3-8b\n",
|
||||
"- fireworks/accounts/fireworks/models/llama-guard-3-11b-vision\n",
|
||||
"- together/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo\n",
|
||||
"- together/meta-llama/Meta-Llama-3.1-70B-Instruct-Turbo\n",
|
||||
"- together/meta-llama/Meta-Llama-3.1-405B-Instruct-Turbo\n",
|
||||
"- together/meta-llama/Llama-3.2-3B-Instruct-Turbo\n",
|
||||
"- together/meta-llama/Llama-3.2-11B-Vision-Instruct-Turbo\n",
|
||||
"- together/meta-llama/Llama-3.2-90B-Vision-Instruct-Turbo\n",
|
||||
"- together/meta-llama/Llama-3.3-70B-Instruct-Turbo\n",
|
||||
"- together/togethercomputer/m2-bert-80M-8k-retrieval\n",
|
||||
"- together/togethercomputer/m2-bert-80M-32k-retrieval\n",
|
||||
"- together/meta-llama/Llama-4-Scout-17B-16E-Instruct\n",
|
||||
"- together/meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8\n",
|
||||
"- together/meta-llama/Llama-Guard-3-8B\n",
|
||||
"- together/meta-llama/Llama-Guard-3-11B-Vision-Turbo\n",
|
||||
"- bedrock/meta.llama3-1-8b-instruct-v1:0\n",
|
||||
"- bedrock/meta.llama3-1-70b-instruct-v1:0\n",
|
||||
"- bedrock/meta.llama3-1-405b-instruct-v1:0\n",
|
||||
"- openai/gpt-3.5-turbo-0125\n",
|
||||
"- openai/gpt-3.5-turbo\n",
|
||||
"- openai/gpt-3.5-turbo-instruct\n",
|
||||
"- openai/gpt-4\n",
|
||||
"- openai/gpt-4-turbo\n",
|
||||
"- openai/gpt-4o\n",
|
||||
"- openai/gpt-4o-2024-08-06\n",
|
||||
"- openai/gpt-4o-mini\n",
|
||||
"- openai/gpt-4o-audio-preview\n",
|
||||
"- openai/chatgpt-4o-latest\n",
|
||||
"- openai/o1\n",
|
||||
"- openai/o1-mini\n",
|
||||
"- openai/o3-mini\n",
|
||||
"- openai/o4-mini\n",
|
||||
"- openai/text-embedding-3-small\n",
|
||||
"- openai/text-embedding-3-large\n",
|
||||
"- anthropic/claude-3-5-sonnet-latest\n",
|
||||
"- anthropic/claude-3-7-sonnet-latest\n",
|
||||
"- anthropic/claude-3-5-haiku-latest\n",
|
||||
"- anthropic/voyage-3\n",
|
||||
"- anthropic/voyage-3-lite\n",
|
||||
"- anthropic/voyage-code-3\n",
|
||||
"- gemini/gemini-1.5-flash\n",
|
||||
"- gemini/gemini-1.5-pro\n",
|
||||
"- gemini/gemini-2.0-flash\n",
|
||||
"- gemini/gemini-2.0-flash-lite\n",
|
||||
"- gemini/gemini-2.5-flash\n",
|
||||
"- gemini/gemini-2.5-flash-lite\n",
|
||||
"- gemini/gemini-2.5-pro\n",
|
||||
"- gemini/text-embedding-004\n",
|
||||
"- groq/llama3-8b-8192\n",
|
||||
"- groq/llama-3.1-8b-instant\n",
|
||||
"- groq/llama3-70b-8192\n",
|
||||
"- groq/llama-3.3-70b-versatile\n",
|
||||
"- groq/llama-3.2-3b-preview\n",
|
||||
"- groq/meta-llama/llama-4-scout-17b-16e-instruct\n",
|
||||
"- groq/meta-llama/llama-4-maverick-17b-128e-instruct\n",
|
||||
"- sambanova/Meta-Llama-3.1-8B-Instruct\n",
|
||||
"- sambanova/Meta-Llama-3.3-70B-Instruct\n",
|
||||
"- sambanova/Llama-4-Maverick-17B-128E-Instruct\n",
|
||||
"- sentence-transformers/all-MiniLM-L6-v2\n",
|
||||
"----\n",
|
||||
"Available shields (safety models):\n",
|
||||
"code-scanner\n",
|
||||
|
@ -354,9 +277,10 @@
|
|||
}
|
||||
],
|
||||
"source": [
|
||||
"print(\"Available models:\")\n",
|
||||
"print(\"Available Fireworks models:\")\n",
|
||||
"for m in client.models.list():\n",
|
||||
" print(f\"- {m.identifier}\")\n",
|
||||
" if m.identifier.startswith(\"fireworks/\"):\n",
|
||||
" print(f\"- {m.identifier}\")\n",
|
||||
"\n",
|
||||
"print(\"----\")\n",
|
||||
"print(\"Available shields (safety models):\")\n",
|
||||
|
@ -401,9 +325,9 @@
|
|||
"name": "stdout",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"File(id='file-33012cc7c18a4430ad1a11b3397c3e29', bytes=41, created_at=1757441277, expires_at=1788977277, filename='shipping_policy.txt', object='file', purpose='assistants')\n",
|
||||
"File(id='file-e7692c2f044e44859ea02b0630195b4e', bytes=48, created_at=1757441277, expires_at=1788977277, filename='returns_policy.txt', object='file', purpose='assistants')\n",
|
||||
"File(id='file-9ff178cbb72b4e79a7513ce5a95a6fe1', bytes=45, created_at=1757441277, expires_at=1788977277, filename='support.txt', object='file', purpose='assistants')\n"
|
||||
"File(id='file-54652c95c56c4c34918a97d7ff8a4320', bytes=41, created_at=1757442621, expires_at=1788978621, filename='shipping_policy.txt', object='file', purpose='assistants')\n",
|
||||
"File(id='file-fb1227c1d1854da1bd774d21e5b7e41c', bytes=48, created_at=1757442621, expires_at=1788978621, filename='returns_policy.txt', object='file', purpose='assistants')\n",
|
||||
"File(id='file-673f874852fe42798675a13d06a256e2', bytes=45, created_at=1757442621, expires_at=1788978621, filename='support.txt', object='file', purpose='assistants')\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
|
@ -461,7 +385,7 @@
|
|||
"name": "stderr",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_d9fa50e8-3ccc-4db3-9102-1a71cd62ab64/search \"HTTP/1.1 200 OK\"\n"
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_708c060b-45da-423e-8354-68529b4fd1a6/search \"HTTP/1.1 200 OK\"\n"
|
||||
]
|
||||
},
|
||||
{
|
||||
|
@ -548,7 +472,7 @@
|
|||
{
|
||||
"data": {
|
||||
"text/plain": [
|
||||
"AIMessage(content='Silent llama, soft and slow,\\n Majestic creature, with a gentle glow.', additional_kwargs={'refusal': None}, response_metadata={'token_usage': None, 'model_name': 'fireworks/accounts/fireworks/models/llama-v3p1-8b-instruct', 'system_fingerprint': None, 'id': 'chatcmpl-8957d071-9f1a-4a7f-9c0e-d516e02194d1', 'service_tier': None, 'finish_reason': 'stop', 'logprobs': None}, id='run--f4331df3-e787-4973-b2e8-7ca90d4e19df-0')"
|
||||
"AIMessage(content=\"A llama's gentle eyes shine bright,\\nIn the Andes, it roams through morning light.\", additional_kwargs={'refusal': None}, response_metadata={'token_usage': None, 'model_name': 'fireworks/accounts/fireworks/models/llama-v3p1-8b-instruct', 'system_fingerprint': None, 'id': 'chatcmpl-602b5967-82a3-476b-9cd2-7d3b29b76ee8', 'service_tier': None, 'finish_reason': 'stop', 'logprobs': None}, id='run--0933c465-ff4d-4a7b-b7fb-fd97dd8244f3-0')"
|
||||
]
|
||||
},
|
||||
"execution_count": 10,
|
||||
|
@ -646,7 +570,7 @@
|
|||
"name": "stderr",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_d9fa50e8-3ccc-4db3-9102-1a71cd62ab64/search \"HTTP/1.1 200 OK\"\n",
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_708c060b-45da-423e-8354-68529b4fd1a6/search \"HTTP/1.1 200 OK\"\n",
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/chat/completions \"HTTP/1.1 200 OK\"\n"
|
||||
]
|
||||
},
|
||||
|
@ -655,7 +579,7 @@
|
|||
"output_type": "stream",
|
||||
"text": [
|
||||
"❓ How long does shipping take?\n",
|
||||
"💡 According to the provided context, Acme ships globally in 3-5 business days.\n"
|
||||
"💡 Acme ships globally in 3-5 business days. This means that shipping typically takes between 3 to 5 working days from the date of dispatch or order fulfillment.\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
|
@ -684,7 +608,7 @@
|
|||
"name": "stderr",
|
||||
"output_type": "stream",
|
||||
"text": [
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_d9fa50e8-3ccc-4db3-9102-1a71cd62ab64/search \"HTTP/1.1 200 OK\"\n",
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/vector_stores/vs_708c060b-45da-423e-8354-68529b4fd1a6/search \"HTTP/1.1 200 OK\"\n",
|
||||
"INFO:httpx:HTTP Request: POST http://0.0.0.0:8321/v1/openai/v1/chat/completions \"HTTP/1.1 200 OK\"\n"
|
||||
]
|
||||
},
|
||||
|
@ -693,7 +617,7 @@
|
|||
"output_type": "stream",
|
||||
"text": [
|
||||
"❓ Can I return a product after 40 days?\n",
|
||||
"💡 Based on the given context, returns are accepted within 30 days of purchase. Therefore, after 40 days, returning a product is not accepted according to the given policy.\n"
|
||||
"💡 Based on the provided context, you cannot return a product after 40 days. The return window is limited to 30 days from the date of purchase.\n"
|
||||
]
|
||||
}
|
||||
],
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue