From 945267a511d8434e9ee4ff6104deeeac500bc894 Mon Sep 17 00:00:00 2001 From: Krrish Dholakia Date: Tue, 8 Oct 2024 22:10:06 -0700 Subject: [PATCH] build: bump version --- litellm/model_prices_and_context_window_backup.json | 12 ++++++------ pyproject.toml | 4 ++-- 2 files changed, 8 insertions(+), 8 deletions(-) diff --git a/litellm/model_prices_and_context_window_backup.json b/litellm/model_prices_and_context_window_backup.json index 7d060dd83..0f1ad77a6 100644 --- a/litellm/model_prices_and_context_window_backup.json +++ b/litellm/model_prices_and_context_window_backup.json @@ -1491,8 +1491,8 @@ "max_tokens": 8192, "max_input_tokens": 8192, "max_output_tokens": 8192, - "input_cost_per_token": 0.00000059, - "output_cost_per_token": 0.00000079, + "input_cost_per_token": 0.00000005, + "output_cost_per_token": 0.00000008, "litellm_provider": "groq", "mode": "chat", "supports_function_calling": true @@ -5303,7 +5303,7 @@ "input_cost_per_token": 0.0, "output_cost_per_token": 0.0, "litellm_provider": "ollama", - "mode": "completion" + "mode": "chat" }, "ollama/llama2:7b": { "max_tokens": 4096, @@ -5312,7 +5312,7 @@ "input_cost_per_token": 0.0, "output_cost_per_token": 0.0, "litellm_provider": "ollama", - "mode": "completion" + "mode": "chat" }, "ollama/llama2:13b": { "max_tokens": 4096, @@ -5321,7 +5321,7 @@ "input_cost_per_token": 0.0, "output_cost_per_token": 0.0, "litellm_provider": "ollama", - "mode": "completion" + "mode": "chat" }, "ollama/llama2:70b": { "max_tokens": 4096, @@ -5330,7 +5330,7 @@ "input_cost_per_token": 0.0, "output_cost_per_token": 0.0, "litellm_provider": "ollama", - "mode": "completion" + "mode": "chat" }, "ollama/llama2-uncensored": { "max_tokens": 4096, diff --git a/pyproject.toml b/pyproject.toml index c8a4796d4..555968557 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [tool.poetry] name = "litellm" -version = "1.48.19" +version = "1.48.20" description = "Library to easily interface with LLM API providers" authors = ["BerriAI"] license = "MIT" @@ -91,7 +91,7 @@ requires = ["poetry-core", "wheel"] build-backend = "poetry.core.masonry.api" [tool.commitizen] -version = "1.48.19" +version = "1.48.20" version_files = [ "pyproject.toml:^version" ]