mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-02 08:44:44 +00:00
refine
This commit is contained in:
parent
92a367340c
commit
7b0deee899
3 changed files with 2 additions and 3 deletions
|
@ -96,5 +96,5 @@ If you are using Llama Stack Safety / Shield APIs, use:
|
|||
llama stack run distributions/meta-reference-gpu/run-with-safety.yaml \
|
||||
--port 5001 \
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-Llama-Guard-3-1B
|
||||
--env SAFETY_MODEL=Llama-Guard-3-1B
|
||||
```
|
||||
|
|
|
@ -76,7 +76,7 @@ docker run \
|
|||
llamastack/distribution-meta-reference-quantized-gpu \
|
||||
--port $LLAMA_STACK_PORT \
|
||||
--env INFERENCE_MODEL=Llama3.2-3B-Instruct \
|
||||
--env SAFETY_MODEL=meta-Llama-Guard-3-1B
|
||||
--env SAFETY_MODEL=Llama-Guard-3-1B
|
||||
```
|
||||
|
||||
### Via Conda
|
||||
|
|
|
@ -17,5 +17,4 @@ async def get_provider_impl(
|
|||
|
||||
impl = MetaReferenceInferenceImpl(config)
|
||||
await impl.initialize()
|
||||
|
||||
return impl
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue