mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-26 11:14:04 +00:00
* use folder for caching * fix importing caching * fix clickhouse pyright * fix linting * fix correctly pass kwargs and args * fix test case for embedding * fix linting * fix embedding caching logic * fix refactor handle utils.py * fix test_embedding_caching_azure_individual_items_reordered |
||
---|---|---|
.. | ||
main.py | ||
Readme.md |
Implementation of litellm.batch_completion
, litellm.batch_completion_models
, litellm.batch_completion_models_all_responses
Doc: https://docs.litellm.ai/docs/completion/batching
LiteLLM Python SDK allows you to:
litellm.batch_completion
Batch litellm.completion function for a given model.litellm.batch_completion_models
Send a request to multiple language models concurrently and return the response as soon as one of the models responds.litellm.batch_completion_models_all_responses
Send a request to multiple language models concurrently and return a list of responses from all models that respond.