mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-31 16:01:46 +00:00
additional instructions in quickstart
This commit is contained in:
parent
7a4fa9e30d
commit
fc6c39b598
1 changed files with 7 additions and 2 deletions
|
@ -114,7 +114,12 @@ The `llama-stack-client` library offers a robust and efficient python methods fo
|
||||||
pip install llama-stack-client
|
pip install llama-stack-client
|
||||||
```
|
```
|
||||||
|
|
||||||
### 2. Create a Python Script (`test_llama_stack.py`)
|
### 3. Create Python Script (`test_llama_stack.py`)
|
||||||
|
```bash
|
||||||
|
touch test_llama_stack.py
|
||||||
|
```
|
||||||
|
|
||||||
|
### 4. Create a Chat Completion Request in Python
|
||||||
|
|
||||||
```python
|
```python
|
||||||
from llama_stack_client import LlamaStackClient
|
from llama_stack_client import LlamaStackClient
|
||||||
|
@ -136,7 +141,7 @@ response = client.inference.chat_completion(
|
||||||
print(response.completion_message.content)
|
print(response.completion_message.content)
|
||||||
```
|
```
|
||||||
|
|
||||||
### 3. Run the Python Script
|
### 5. Run the Python Script
|
||||||
|
|
||||||
```bash
|
```bash
|
||||||
python test_llama_stack.py
|
python test_llama_stack.py
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue