forked from phoenix/litellm-mirror
Merge pull request #2473 from BerriAI/litellm_fix_compatible_provider_model_name
fix(openai.py): return model name with custom llm provider for openai-compatible endpoints (e.g. mistral, together ai, etc.)
This commit is contained in:
commit
0d18f3c0ca
4 changed files with 5 additions and 1 deletions
|
@ -6427,7 +6427,7 @@ def convert_to_model_response_object(
|
|||
"system_fingerprint"
|
||||
]
|
||||
|
||||
if "model" in response_object:
|
||||
if "model" in response_object and model_response_object.model is None:
|
||||
model_response_object.model = response_object["model"]
|
||||
|
||||
if start_time is not None and end_time is not None:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue