forked from phoenix/litellm-mirror
script to add new models
This commit is contained in:
parent
340f7317e8
commit
166eb24c54
1 changed files with 72 additions and 0 deletions
72
cookbook/misc/add_new_models.py
Normal file
72
cookbook/misc/add_new_models.py
Normal file
|
@ -0,0 +1,72 @@
|
|||
import requests
|
||||
import json
|
||||
|
||||
|
||||
def get_initial_config():
|
||||
proxy_base_url = input("Enter your proxy base URL (e.g., http://localhost:4000): ")
|
||||
master_key = input("Enter your LITELLM_MASTER_KEY ")
|
||||
return proxy_base_url, master_key
|
||||
|
||||
|
||||
def get_user_input():
|
||||
model_name = input(
|
||||
"Enter model_name (this is the 'model' passed in /chat/completions requests):"
|
||||
)
|
||||
model = input("litellm_params: Enter model eg. 'azure/<your-deployment-name>': ")
|
||||
tpm = int(input("litellm_params: Enter tpm (tokens per minute): "))
|
||||
rpm = int(input("litellm_params: Enter rpm (requests per minute): "))
|
||||
api_key = input("litellm_params: Enter api_key: ")
|
||||
api_base = input("litellm_params: Enter api_base: ")
|
||||
api_version = input("litellm_params: Enter api_version: ")
|
||||
timeout = int(input("litellm_params: Enter timeout (0 for default): "))
|
||||
stream_timeout = int(
|
||||
input("litellm_params: Enter stream_timeout (0 for default): ")
|
||||
)
|
||||
max_retries = int(input("litellm_params: Enter max_retries (0 for default): "))
|
||||
|
||||
return {
|
||||
"model_name": model_name,
|
||||
"litellm_params": {
|
||||
"model": model,
|
||||
"tpm": tpm,
|
||||
"rpm": rpm,
|
||||
"api_key": api_key,
|
||||
"api_base": api_base,
|
||||
"api_version": api_version,
|
||||
"timeout": timeout,
|
||||
"stream_timeout": stream_timeout,
|
||||
"max_retries": max_retries,
|
||||
},
|
||||
}
|
||||
|
||||
|
||||
def make_request(proxy_base_url, master_key, data):
|
||||
url = f"{proxy_base_url}/model/new"
|
||||
headers = {
|
||||
"Content-Type": "application/json",
|
||||
"Authorization": f"Bearer {master_key}",
|
||||
}
|
||||
|
||||
response = requests.post(url, headers=headers, json=data)
|
||||
|
||||
print(f"Status Code: {response.status_code}")
|
||||
print(f"Response from adding model: {response.text}")
|
||||
|
||||
|
||||
def main():
|
||||
proxy_base_url, master_key = get_initial_config()
|
||||
|
||||
while True:
|
||||
print("Adding new Model to your proxy server...")
|
||||
data = get_user_input()
|
||||
make_request(proxy_base_url, master_key, data)
|
||||
|
||||
add_another = input("Do you want to add another model? (yes/no): ").lower()
|
||||
if add_another != "yes":
|
||||
break
|
||||
|
||||
print("Script finished.")
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
Loading…
Add table
Add a link
Reference in a new issue