mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-25 02:34:29 +00:00
huggingface/mistralai/Mistral-7B-Instruct-v0.3
This commit is contained in:
parent
3fe1f3b3b2
commit
970e9c7507
2 changed files with 3 additions and 3 deletions
|
@ -102,7 +102,7 @@ async def test_huggingface_text_completion_logprobs():
|
|||
client = AsyncHTTPHandler()
|
||||
with patch.object(client, "post", return_value=return_val) as mock_post:
|
||||
response = await litellm.atext_completion(
|
||||
model="huggingface/mistralai/Mistral-7B-v0.1",
|
||||
model="huggingface/mistralai/Mistral-7B-Instruct-v0.3",
|
||||
prompt="good morning",
|
||||
client=client,
|
||||
)
|
||||
|
|
|
@ -3940,7 +3940,7 @@ def test_completion_hf_prompt_array():
|
|||
litellm.set_verbose = True
|
||||
print("\n testing hf mistral\n")
|
||||
response = text_completion(
|
||||
model="huggingface/mistralai/Mistral-7B-v0.1",
|
||||
model="huggingface/mistralai/Mistral-7B-Instruct-v0.3",
|
||||
prompt=token_prompt, # token prompt is a 2d list,
|
||||
max_tokens=0,
|
||||
temperature=0.0,
|
||||
|
@ -3971,7 +3971,7 @@ def test_text_completion_stream():
|
|||
try:
|
||||
for _ in range(2): # check if closed client used
|
||||
response = text_completion(
|
||||
model="huggingface/mistralai/Mistral-7B-v0.1",
|
||||
model="huggingface/mistralai/Mistral-7B-Instruct-v0.3",
|
||||
prompt="good morning",
|
||||
stream=True,
|
||||
max_tokens=10,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue