forked from phoenix/litellm-mirror
script to add new models
This commit is contained in:
parent
340f7317e8
commit
166eb24c54
1 changed files with 72 additions and 0 deletions
72
cookbook/misc/add_new_models.py
Normal file
72
cookbook/misc/add_new_models.py
Normal file
|
@ -0,0 +1,72 @@
|
||||||
|
import requests
|
||||||
|
import json
|
||||||
|
|
||||||
|
|
||||||
|
def get_initial_config():
|
||||||
|
proxy_base_url = input("Enter your proxy base URL (e.g., http://localhost:4000): ")
|
||||||
|
master_key = input("Enter your LITELLM_MASTER_KEY ")
|
||||||
|
return proxy_base_url, master_key
|
||||||
|
|
||||||
|
|
||||||
|
def get_user_input():
|
||||||
|
model_name = input(
|
||||||
|
"Enter model_name (this is the 'model' passed in /chat/completions requests):"
|
||||||
|
)
|
||||||
|
model = input("litellm_params: Enter model eg. 'azure/<your-deployment-name>': ")
|
||||||
|
tpm = int(input("litellm_params: Enter tpm (tokens per minute): "))
|
||||||
|
rpm = int(input("litellm_params: Enter rpm (requests per minute): "))
|
||||||
|
api_key = input("litellm_params: Enter api_key: ")
|
||||||
|
api_base = input("litellm_params: Enter api_base: ")
|
||||||
|
api_version = input("litellm_params: Enter api_version: ")
|
||||||
|
timeout = int(input("litellm_params: Enter timeout (0 for default): "))
|
||||||
|
stream_timeout = int(
|
||||||
|
input("litellm_params: Enter stream_timeout (0 for default): ")
|
||||||
|
)
|
||||||
|
max_retries = int(input("litellm_params: Enter max_retries (0 for default): "))
|
||||||
|
|
||||||
|
return {
|
||||||
|
"model_name": model_name,
|
||||||
|
"litellm_params": {
|
||||||
|
"model": model,
|
||||||
|
"tpm": tpm,
|
||||||
|
"rpm": rpm,
|
||||||
|
"api_key": api_key,
|
||||||
|
"api_base": api_base,
|
||||||
|
"api_version": api_version,
|
||||||
|
"timeout": timeout,
|
||||||
|
"stream_timeout": stream_timeout,
|
||||||
|
"max_retries": max_retries,
|
||||||
|
},
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
def make_request(proxy_base_url, master_key, data):
|
||||||
|
url = f"{proxy_base_url}/model/new"
|
||||||
|
headers = {
|
||||||
|
"Content-Type": "application/json",
|
||||||
|
"Authorization": f"Bearer {master_key}",
|
||||||
|
}
|
||||||
|
|
||||||
|
response = requests.post(url, headers=headers, json=data)
|
||||||
|
|
||||||
|
print(f"Status Code: {response.status_code}")
|
||||||
|
print(f"Response from adding model: {response.text}")
|
||||||
|
|
||||||
|
|
||||||
|
def main():
|
||||||
|
proxy_base_url, master_key = get_initial_config()
|
||||||
|
|
||||||
|
while True:
|
||||||
|
print("Adding new Model to your proxy server...")
|
||||||
|
data = get_user_input()
|
||||||
|
make_request(proxy_base_url, master_key, data)
|
||||||
|
|
||||||
|
add_another = input("Do you want to add another model? (yes/no): ").lower()
|
||||||
|
if add_another != "yes":
|
||||||
|
break
|
||||||
|
|
||||||
|
print("Script finished.")
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == "__main__":
|
||||||
|
main()
|
Loading…
Add table
Add a link
Reference in a new issue