From 6598d817ec54dba28a368a65032b6c64dbc099fd Mon Sep 17 00:00:00 2001 From: Ishaan Jaff Date: Fri, 17 Jan 2025 10:02:22 -0800 Subject: [PATCH] doc batch settings --- docs/my-website/docs/batches.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/my-website/docs/batches.md b/docs/my-website/docs/batches.md index 9dae7fda25..c415ab0c39 100644 --- a/docs/my-website/docs/batches.md +++ b/docs/my-website/docs/batches.md @@ -179,7 +179,7 @@ First, set up where you want to store the batch files. You can use S3, GCS, or A ```yaml model_list: - - model_name: vllm-models + - model_name: vllm-model litellm_params: model: openai/facebook/opt-125m # the `openai/` prefix tells litellm it's openai compatible api_base: http://0.0.0.0:4000/v1 @@ -187,7 +187,7 @@ model_list: batch_settings: # Configure S3 for batch file storage - model: vllm-models + model: vllm-model batch_storage_params: s3_bucket_name: my-batch-bucket # AWS Bucket Name for S3 s3_region_name: us-west-2 # AWS Region Name for S3