forked from phoenix/litellm-mirror
.. | ||
tests | ||
.env.template | ||
__init__.py | ||
Dockerfile | ||
main.py | ||
openapi.json | ||
README.md | ||
requirements.txt | ||
utils.py |
litellm-server
A simple, fast, and lightweight OpenAI-compatible server to call 100+ LLM APIs.
Usage
docker run -e PORT=8000 -e OPENAI_API_KEY=<your-openai-key> -p 8000:8000 ghcr.io/berriai/litellm:latest
# UVICORN: OpenAI Proxy running on http://0.0.0.0:8000
curl http://0.0.0.0:8000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-3.5-turbo",
"messages": [{"role": "user", "content": "Say this is a test!"}],
"temperature": 0.7
}'
See how to call Huggingface,Bedrock,TogetherAI,Anthropic, etc.
Endpoints:
/chat/completions
- chat completions endpoint to call 100+ LLMs/models
- available models on server
Save Model-specific params (API Base, API Keys, Temperature, etc.)
Use the router_config_template.yaml to save model-specific information like api_base, api_key, temperature, max_tokens, etc.
- Create a
config.yaml
file
model_list:
- model_name: gpt-3.5-turbo
litellm_params: # params for litellm.completion() - https://docs.litellm.ai/docs/completion/input#input---request-body
model: azure/chatgpt-v-2 # azure/<your-deployment-name>
api_key: your_azure_api_key
api_version: your_azure_api_version
api_base: your_azure_api_base
- model_name: mistral-7b
litellm_params:
model: ollama/mistral
api_base: your_ollama_api_base
- Start the server
docker run --name litellm_server_1 -e PORT=8000 -p 8000:8000 -v $(pwd)/config.yaml:/app/config.yaml litellm_server
Running Locally
$ git clone https://github.com/BerriAI/litellm.git
$ cd ./litellm/litellm_server
$ uvicorn main:app --host 0.0.0.0 --port 8000
Custom Config
- Create + Modify router_config.yaml (save your azure/openai/etc. deployment info)
cp ./router_config_template.yaml ./router_config.yaml
- Build Docker Image
docker build -t litellm_server . --build-arg CONFIG_FILE=./router_config.yaml
- Run Docker Image
docker run --name litellm_server -e PORT=8000 -p 8000:8000 litellm_server