(fix) proxy + docs: use openai.chat.completions.create instead of openai.ChatCompletions

This commit is contained in:
ishaan-jaff 2023-11-13 08:24:24 -08:00
parent 60770e1eec
commit cf0ab7155e
10 changed files with 20 additions and 21 deletions

View file

@ -113,7 +113,7 @@ import openai
openai.api_base = "http://0.0.0.0:8000"
print(openai.ChatCompletion.create(model="test", messages=[{"role":"user", "content":"Hey!"}]))
print(openai.chat.completions.create(model="test", messages=[{"role":"user", "content":"Hey!"}]))
```
## Supported Provider ([Docs](https://docs.litellm.ai/docs/providers))

View file

@ -109,7 +109,7 @@ model_list = [{ # list of model deployments
router = Router(model_list=model_list)
# openai.ChatCompletion.create replacement
# openai.chat.completions.create replacement
response = router.completion(model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hey, how's it going?"}]

View file

@ -115,7 +115,7 @@ print(openai.api_base)
print(f'LiteLLM: response from proxy with streaming')
response = openai.ChatCompletion.create(
response = openai.chat.completions.create(
model="ollama/llama2",
messages = [
{

View file

@ -33,7 +33,7 @@ import openai
openai.api_base = "http://0.0.0.0:8000"
print(openai.ChatCompletion.create(model="test", messages=[{"role":"user", "content":"Hey!"}]))
print(openai.chat.completions.create(model="test", messages=[{"role":"user", "content":"Hey!"}]))
```
### Supported LLMs
@ -541,7 +541,7 @@ If you're repo let's you set model name, you can call the specific model by just
import openai
openai.api_base = "http://0.0.0.0:8000"
completion = openai.ChatCompletion.create(model="zephyr-alpha", messages=[{"role": "user", "content": "Hello world"}])
completion = openai.chat.completions.create(model="zephyr-alpha", messages=[{"role": "user", "content": "Hello world"}])
print(completion.choices[0].message.content)
```
@ -551,7 +551,7 @@ If you're repo only let's you specify api base, then you can add the model name
import openai
openai.api_base = "http://0.0.0.0:8000/openai/deployments/zephyr-alpha/chat/completions" # zephyr-alpha will be used
completion = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages=[{"role": "user", "content": "Hello world"}])
completion = openai.chat.completions.create(model="gpt-3.5-turbo", messages=[{"role": "user", "content": "Hello world"}])
print(completion.choices[0].message.content)
```

View file

@ -19,7 +19,7 @@ import openai
openai.api_base = "http://0.0.0.0:8000"
print(openai.ChatCompletion.create(model="test", messages=[{"role":"user", "content":"Hey!"}]))
print(openai.chat.completions.create(model="test", messages=[{"role":"user", "content":"Hey!"}]))
```
[**See how to call Huggingface,Bedrock,TogetherAI,Anthropic, etc.**](https://docs.litellm.ai/docs/simple_proxy)

View file

@ -125,11 +125,10 @@ def run_server(host, port, api_base, api_version, model, alias, add_key, headers
openai.api_base = api_base
openai.api_key = "temp-key"
print(openai.api_base)
response = openai.Completion.create(model="gpt-3.5-turbo", prompt='this is a test request, write a short poem')
response = openai.completions.create(model="gpt-3.5-turbo", prompt='this is a test request, write a short poem')
print(response)
response = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages = [
response = openai.chat.completions.create(model="gpt-3.5-turbo", messages = [
{
"role": "user",
"content": "this is a test request, write a short poem"
@ -138,7 +137,7 @@ def run_server(host, port, api_base, api_version, model, alias, add_key, headers
click.echo(f'LiteLLM: response from proxy {response}')
click.echo(f'LiteLLM: response from proxy with streaming {response}')
response = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages = [
response = openai.chat.completions.create(model="gpt-3.5-turbo", messages = [
{
"role": "user",
"content": "this is a test request, write a short poem"

View file

@ -4,7 +4,7 @@ print("making request")
openai.api_key = "anything" # this gets passed as a header
response = openai.ChatCompletion.create(
response = openai.chat.completions.create(
model = "bedrock/anthropic.claude-instant-v1",
messages = [
{
@ -22,7 +22,7 @@ response = openai.ChatCompletion.create(
print(response)
# response = openai.ChatCompletion.create(
# response = openai.chat.completions.create(
# model = "gpt-3.5-turbo",
# messages = [
# {

View file

@ -4,7 +4,7 @@ dotenv.load_dotenv()
openai.api_key = os.getenv("ANTHROPIC_API_KEY") # this gets passed as a header
response1 = openai.ChatCompletion.create(
response1 = openai.chat.completions.create(
model = "claude-instant-1",
messages = [
{
@ -21,7 +21,7 @@ except:
time.sleep(1) # allow time for request to be stored
response2 = openai.ChatCompletion.create(
response2 = openai.chat.completions.create(
model = "claude-instant-1",
messages = [
{
@ -39,7 +39,7 @@ except:
openai.api_key = os.getenv("OPENAI_API_KEY")
try:
response3 = openai.ChatCompletion.create(
response3 = openai.chat.completions.create(
model = "gpt-3.5-turbo",
messages = [
{
@ -58,7 +58,7 @@ except:
openai.api_key = os.getenv("ANTHROPIC_API_KEY") # this gets passed as a header
# switch caching off using cache flag
response4 = openai.ChatCompletion.create(
response4 = openai.chat.completions.create(
model = "claude-instant-1",
messages = [
{

View file

@ -5,7 +5,7 @@ openai.api_key = "this can be anything"
print("making request")
api_key = ""
response = openai.ChatCompletion.create(
response = openai.chat.completions.create(
model = "gpt-3.5-turbo",
messages = [
{
@ -21,7 +21,7 @@ response = openai.ChatCompletion.create(
print(response)
response = openai.ChatCompletion.create(
response = openai.chat.completions.create(
model = "gpt-3.5-turbo",
messages = [
{

View file

@ -4,7 +4,7 @@ openai.api_key = "this can be anything"
print("making request")
api_key = ""
response = openai.ChatCompletion.create(
response = openai.chat.completions.create(
model = "openrouter/google/palm-2-chat-bison",
messages = [
{
@ -20,7 +20,7 @@ response = openai.ChatCompletion.create(
print(response)
response = openai.ChatCompletion.create(
response = openai.chat.completions.create(
model = "openrouter/google/palm-2-chat-bison",
messages = [
{