Generation/Completion/Chat Completion Models
OpenAI Chat Completion Models
Model Name |
Function Call |
Required OS Variables |
gpt-3.5-turbo |
completion('gpt-3.5-turbo', messages) |
os.environ['OPENAI_API_KEY'] |
gpt-3.5-turbo-16k |
completion('gpt-3.5-turbo-16k', messages) |
os.environ['OPENAI_API_KEY'] |
gpt-3.5-turbo-16k-0613 |
completion('gpt-3.5-turbo-16k-0613', messages) |
os.environ['OPENAI_API_KEY'] |
gpt-4 |
completion('gpt-4', messages) |
os.environ['OPENAI_API_KEY'] |
Azure OpenAI Chat Completion Models
Model Name |
Function Call |
Required OS Variables |
gpt-3.5-turbo |
completion('gpt-3.5-turbo', messages, azure=True) |
os.environ['AZURE_API_KEY'] ,
os.environ['AZURE_API_BASE'] ,
os.environ['AZURE_API_VERSION'] |
gpt-4 |
completion('gpt-4', messages, azure=True) |
os.environ['AZURE_API_KEY'] ,
os.environ['AZURE_API_BASE'] ,
os.environ['AZURE_API_VERSION'] |
OpenAI Text Completion Models
Model Name |
Function Call |
Required OS Variables |
text-davinci-003 |
completion('text-davinci-003', messages) |
os.environ['OPENAI_API_KEY'] |
Cohere Models
Model Name |
Function Call |
Required OS Variables |
command-nightly |
completion('command-nightly', messages) |
os.environ['COHERE_API_KEY'] |
Anthropic Models
Model Name |
Function Call |
Required OS Variables |
claude-instant-1 |
completion('claude-instant-1', messages) |
os.environ['ANTHROPIC_API_KEY'] |
claude-2 |
completion('claude-2', messages) |
os.environ['ANTHROPIC_API_KEY'] |
Hugging Face Inference API
All text2text-generation
and text-generation
models are supported by liteLLM. You can use any text model from Hugging Face with the following steps:
- Copy the
model repo
URL from Hugging Face and set it as the model
parameter in the completion call.
- Set
hugging_face
parameter to True
.
- Make sure to set the hugging face API key
Here are some examples of supported models:
Note that the models mentioned in the table are examples, and you can use any text model available on Hugging Face by following the steps above.
Model Name |
Function Call |
Required OS Variables |
stabilityai/stablecode-completion-alpha-3b-4k |
completion(model="stabilityai/stablecode-completion-alpha-3b-4k", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
bigcode/starcoder |
completion(model="bigcode/starcoder", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
google/flan-t5-xxl |
completion(model="google/flan-t5-xxl", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |
google/flan-t5-large |
completion(model="google/flan-t5-large", messages=messages, hugging_face=True) |
os.environ['HF_TOKEN'] |