forked from phoenix-oss/llama-stack-mirror
Use GPUs 0 and 1
This commit is contained in:
parent
1245a625ce
commit
e4f14eafe2
1 changed files with 2 additions and 2 deletions
|
@ -22,7 +22,7 @@ services:
|
||||||
devices:
|
devices:
|
||||||
- nvidia.com/gpu=all
|
- nvidia.com/gpu=all
|
||||||
environment:
|
environment:
|
||||||
- CUDA_VISIBLE_DEVICES=4
|
- CUDA_VISIBLE_DEVICES=0
|
||||||
- HUGGING_FACE_HUB_TOKEN=$HF_TOKEN
|
- HUGGING_FACE_HUB_TOKEN=$HF_TOKEN
|
||||||
command: >
|
command: >
|
||||||
--gpu-memory-utilization 0.75
|
--gpu-memory-utilization 0.75
|
||||||
|
@ -48,7 +48,7 @@ services:
|
||||||
devices:
|
devices:
|
||||||
- nvidia.com/gpu=all
|
- nvidia.com/gpu=all
|
||||||
environment:
|
environment:
|
||||||
- CUDA_VISIBLE_DEVICES=5
|
- CUDA_VISIBLE_DEVICES=1
|
||||||
- HUGGING_FACE_HUB_TOKEN=$HF_TOKEN
|
- HUGGING_FACE_HUB_TOKEN=$HF_TOKEN
|
||||||
command: >
|
command: >
|
||||||
--gpu-memory-utilization 0.75
|
--gpu-memory-utilization 0.75
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue