From 1f2df59ecee2070e49053173d57b1ee44a5f049e Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?S=C3=A9bastien=20Han?= Date: Thu, 10 Apr 2025 18:37:48 +0200 Subject: [PATCH] docs: fix model name (#1926) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit # What does this PR do? Use llama3.2:3b for consistency. Signed-off-by: Sébastien Han --- docs/source/getting_started/index.md | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index e9ad51961..82329e60e 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -9,10 +9,10 @@ In this guide, we'll walk through how to build a RAG agent locally using Llama S ### 1. Download a Llama model with Ollama ```bash -ollama pull llama3.2:3b-instruct-fp16 +ollama pull llama3.2:3b ``` -This will instruct the Ollama service to download the Llama 3.2 3B Instruct model, which we'll use in the rest of this guide. +This will instruct the Ollama service to download the Llama 3.2 3B model, which we'll use in the rest of this guide. ```{admonition} Note :class: tip @@ -176,7 +176,7 @@ python inference.py ``` Sample output: ``` -Model: llama3.2:3b-instruct-fp16 +Model: llama3.2:3b Here is a haiku about coding: Lines of code unfold