forked from phoenix/litellm-mirror
bump: version 1.34.4 → 1.34.5
This commit is contained in:
parent
bec093675c
commit
4d22fe0947
3 changed files with 15 additions and 9 deletions
|
@ -14,20 +14,26 @@ spec:
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- name: litellm-container
|
- name: litellm-container
|
||||||
image: ghcr.io/berriai/litellm:main-latest
|
image: ghcr.io/berriai/litellm-database:main-v1.34.4.dev2
|
||||||
imagePullPolicy: Always
|
|
||||||
env:
|
env:
|
||||||
- name: AZURE_API_KEY
|
- name: AZURE_API_KEY
|
||||||
value: "d6f****"
|
value: "d6f82361954b450188295b448e2091ca"
|
||||||
- name: AZURE_API_BASE
|
- name: AZURE_API_BASE
|
||||||
value: "https://openai"
|
value: "https://openai-gpt-4-test-v-1.openai.azure.com/"
|
||||||
- name: LITELLM_MASTER_KEY
|
- name: LITELLM_MASTER_KEY
|
||||||
value: "sk-1234"
|
value: "sk-1234"
|
||||||
- name: DATABASE_URL
|
- name: DATABASE_URL
|
||||||
value: "postgresql://ishaan:*********"
|
value: "postgresql://ishaan:HogdJCvK89Xe@ep-bitter-fog-a573fgz9.us-east-2.aws.neon.tech/local?sslmode=require"
|
||||||
|
- name: SLACK_WEBHOOK_URL
|
||||||
|
value: "https://hooks.slack.com/services/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"
|
||||||
|
- name: ishu
|
||||||
|
value: "ishaanGMGhj"
|
||||||
args:
|
args:
|
||||||
- "--config"
|
- "--config"
|
||||||
- "/app/proxy_config.yaml" # Update the path to mount the config file
|
- "/app/proxy_config.yaml" # Update the path to mount the config file
|
||||||
|
- "--run_gunicorn"
|
||||||
|
- "--num_workers"
|
||||||
|
- "4"
|
||||||
volumeMounts: # Define volume mount for proxy_config.yaml
|
volumeMounts: # Define volume mount for proxy_config.yaml
|
||||||
- name: config-volume
|
- name: config-volume
|
||||||
mountPath: /app
|
mountPath: /app
|
||||||
|
|
|
@ -7,8 +7,8 @@ model_list:
|
||||||
api_key: os.environ/AZURE_API_KEY # The `os.environ/` prefix tells litellm to read this from the env. See https://docs.litellm.ai/docs/simple_proxy#load-api-keys-from-vault
|
api_key: os.environ/AZURE_API_KEY # The `os.environ/` prefix tells litellm to read this from the env. See https://docs.litellm.ai/docs/simple_proxy#load-api-keys-from-vault
|
||||||
- model_name: gpt-3.5-turbo-large
|
- model_name: gpt-3.5-turbo-large
|
||||||
litellm_params:
|
litellm_params:
|
||||||
"model": "gpt-3.5-turbo-1106"
|
model: "gpt-3.5-turbo-1106"
|
||||||
"api_key": os.environ/OPENAI_API_KEY
|
api_key: os.environ/OPENAI_API_KEY
|
||||||
- model_name: gpt-4
|
- model_name: gpt-4
|
||||||
litellm_params:
|
litellm_params:
|
||||||
model: azure/chatgpt-v-2
|
model: azure/chatgpt-v-2
|
||||||
|
|
|
@ -1,6 +1,6 @@
|
||||||
[tool.poetry]
|
[tool.poetry]
|
||||||
name = "litellm"
|
name = "litellm"
|
||||||
version = "1.34.4"
|
version = "1.34.5"
|
||||||
description = "Library to easily interface with LLM API providers"
|
description = "Library to easily interface with LLM API providers"
|
||||||
authors = ["BerriAI"]
|
authors = ["BerriAI"]
|
||||||
license = "MIT"
|
license = "MIT"
|
||||||
|
@ -80,7 +80,7 @@ requires = ["poetry-core", "wheel"]
|
||||||
build-backend = "poetry.core.masonry.api"
|
build-backend = "poetry.core.masonry.api"
|
||||||
|
|
||||||
[tool.commitizen]
|
[tool.commitizen]
|
||||||
version = "1.34.4"
|
version = "1.34.5"
|
||||||
version_files = [
|
version_files = [
|
||||||
"pyproject.toml:^version"
|
"pyproject.toml:^version"
|
||||||
]
|
]
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue