mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
fix get_chat_completion_prompt
This commit is contained in:
parent
903f792777
commit
7fabee3eff
1 changed files with 3 additions and 3 deletions
|
@ -164,6 +164,9 @@ class HumanloopLogger(CustomLogger):
|
|||
"humanloop_api_key"
|
||||
) or get_secret_str("HUMANLOOP_API_KEY")
|
||||
|
||||
if prompt_id is None:
|
||||
raise ValueError("prompt_id is required for Humanloop integration")
|
||||
|
||||
if humanloop_api_key is None:
|
||||
return super().get_chat_completion_prompt(
|
||||
model=model,
|
||||
|
@ -174,9 +177,6 @@ class HumanloopLogger(CustomLogger):
|
|||
dynamic_callback_params=dynamic_callback_params,
|
||||
)
|
||||
|
||||
if prompt_id is None:
|
||||
raise ValueError("prompt_id is required for Humanloop integration")
|
||||
|
||||
prompt_template = prompt_manager._get_prompt_from_id(
|
||||
humanloop_prompt_id=prompt_id, humanloop_api_key=humanloop_api_key
|
||||
)
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue