forked from phoenix/litellm-mirror
read ft config
This commit is contained in:
parent
9d90f174a7
commit
bd7b485d09
2 changed files with 29 additions and 0 deletions
|
@ -25,6 +25,27 @@ model_list:
|
||||||
api_key: "os.environ/OPENAI_API_KEY"
|
api_key: "os.environ/OPENAI_API_KEY"
|
||||||
model_info:
|
model_info:
|
||||||
mode: audio_speech
|
mode: audio_speech
|
||||||
|
|
||||||
|
# For /fine_tuning/jobs endpoints
|
||||||
|
finetune_settings:
|
||||||
|
- custom_llm_provider: azure
|
||||||
|
api_base: https://exampleopenaiendpoint-production.up.railway.app/
|
||||||
|
api_key: fake-key
|
||||||
|
api_version: "2023-03-15-preview"
|
||||||
|
- custom_llm_provider: openai
|
||||||
|
api_key: fake-key
|
||||||
|
|
||||||
|
# for /files endpoints
|
||||||
|
files_settings:
|
||||||
|
- custom_llm_provider: azure
|
||||||
|
api_base: https://exampleopenaiendpoint-production.up.railway.app/
|
||||||
|
api_key: fake-key
|
||||||
|
api_version: "2023-03-15-preview"
|
||||||
|
- custom_llm_provider: openai
|
||||||
|
api_key: fake-key
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
general_settings:
|
general_settings:
|
||||||
master_key: sk-1234
|
master_key: sk-1234
|
||||||
|
|
||||||
|
|
|
@ -154,6 +154,7 @@ from litellm.proxy.common_utils.openai_endpoint_utils import (
|
||||||
remove_sensitive_info_from_deployment,
|
remove_sensitive_info_from_deployment,
|
||||||
)
|
)
|
||||||
from litellm.proxy.fine_tuning_endpoints.endpoints import router as fine_tuning_router
|
from litellm.proxy.fine_tuning_endpoints.endpoints import router as fine_tuning_router
|
||||||
|
from litellm.proxy.fine_tuning_endpoints.endpoints import set_fine_tuning_config
|
||||||
from litellm.proxy.guardrails.init_guardrails import initialize_guardrails
|
from litellm.proxy.guardrails.init_guardrails import initialize_guardrails
|
||||||
from litellm.proxy.health_check import perform_health_check
|
from litellm.proxy.health_check import perform_health_check
|
||||||
from litellm.proxy.health_endpoints._health_endpoints import router as health_router
|
from litellm.proxy.health_endpoints._health_endpoints import router as health_router
|
||||||
|
@ -1808,6 +1809,13 @@ class ProxyConfig:
|
||||||
assistant_settings["litellm_params"][k] = v
|
assistant_settings["litellm_params"][k] = v
|
||||||
assistants_config = AssistantsTypedDict(**assistant_settings) # type: ignore
|
assistants_config = AssistantsTypedDict(**assistant_settings) # type: ignore
|
||||||
|
|
||||||
|
## /fine_tuning/jobs endpoints config
|
||||||
|
finetuning_config = config.get("finetune_settings", None)
|
||||||
|
set_fine_tuning_config(config=finetuning_config)
|
||||||
|
|
||||||
|
## /files endpoint config
|
||||||
|
files_config = config.get("files_settings", None)
|
||||||
|
|
||||||
## ROUTER SETTINGS (e.g. routing_strategy, ...)
|
## ROUTER SETTINGS (e.g. routing_strategy, ...)
|
||||||
router_settings = config.get("router_settings", None)
|
router_settings = config.get("router_settings", None)
|
||||||
if router_settings and isinstance(router_settings, dict):
|
if router_settings and isinstance(router_settings, dict):
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue