(fix) model_prices_and_context_window.json error

This commit is contained in:
ishaan-jaff 2024-02-07 15:42:37 -08:00
parent 0300512743
commit 68926c6524
3 changed files with 10 additions and 10 deletions

View file

@ -8,6 +8,15 @@
"litellm_provider": "openai", "litellm_provider": "openai",
"mode": "chat" "mode": "chat"
}, },
"gpt-4-turbo-preview": {
"max_tokens": 8192,
"max_input_tokens": 8192,
"max_output_tokens": 4096,
"input_cost_per_token": 0.00001,
"output_cost_per_token": 0.00003,
"litellm_provider": "openai",
"mode": "chat"
},
"gpt-4-0314": { "gpt-4-0314": {
"max_tokens": 8192, "max_tokens": 8192,
"max_input_tokens": 8192, "max_input_tokens": 8192,

View file

@ -1743,7 +1743,7 @@ def test_azure_cloudflare_api():
def test_completion_anyscale_2(): def test_completion_anyscale_2():
try: try:
# litellm.set_verbose=True # litellm.set_verbose= True
messages = [ messages = [
{"role": "system", "content": "You're a good bot"}, {"role": "system", "content": "You're a good bot"},
{ {

View file

@ -8,15 +8,6 @@
"litellm_provider": "openai", "litellm_provider": "openai",
"mode": "chat" "mode": "chat"
}, },
"gpt-4-turbo-preview": {
"max_tokens": 128000,
"max_input_tokens": 128000,
"max_output_tokens": 4096,
"input_cost_per_token": 0.00001,
"output_cost_per_token": 0.00003,
"litellm_provider": "openai",
"mode": "chat"
},
"gpt-4-turbo-preview": { "gpt-4-turbo-preview": {
"max_tokens": 8192, "max_tokens": 8192,
"max_input_tokens": 8192, "max_input_tokens": 8192,