forked from phoenix/litellm-mirror
Merge pull request #2473 from BerriAI/litellm_fix_compatible_provider_model_name
fix(openai.py): return model name with custom llm provider for openai-compatible endpoints (e.g. mistral, together ai, etc.)
This commit is contained in:
commit
0d18f3c0ca
4 changed files with 5 additions and 1 deletions
|
@ -289,6 +289,7 @@ def test_completion_mistral_api():
|
|||
cost = litellm.completion_cost(completion_response=response)
|
||||
print("cost to make mistral completion=", cost)
|
||||
assert cost > 0.0
|
||||
assert response.model == "mistral/mistral-tiny"
|
||||
except Exception as e:
|
||||
pytest.fail(f"Error occurred: {e}")
|
||||
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue