mirror of
https://github.com/BerriAI/litellm.git
synced 2025-04-27 11:43:54 +00:00
(fix) linting errors
This commit is contained in:
parent
c05f8bf554
commit
b99b137f10
1 changed files with 3 additions and 3 deletions
|
@ -525,7 +525,7 @@ def completion(
|
||||||
# https://platform.openai.com/docs/api-reference/completions/create
|
# https://platform.openai.com/docs/api-reference/completions/create
|
||||||
prompt = messages[0]["content"]
|
prompt = messages[0]["content"]
|
||||||
else:
|
else:
|
||||||
prompt = " ".join([message["content"] for message in messages])
|
prompt = " ".join([message["content"] for message in messages]) # type: ignore
|
||||||
## LOGGING
|
## LOGGING
|
||||||
logging.pre_call(
|
logging.pre_call(
|
||||||
input=prompt,
|
input=prompt,
|
||||||
|
@ -1188,7 +1188,7 @@ def completion(
|
||||||
model_response["choices"][0]["message"]["content"] = response_string
|
model_response["choices"][0]["message"]["content"] = response_string
|
||||||
model_response["created"] = time.time()
|
model_response["created"] = time.time()
|
||||||
model_response["model"] = "ollama/" + model
|
model_response["model"] = "ollama/" + model
|
||||||
prompt_tokens = len(encoding.encode(prompt))
|
prompt_tokens = len(encoding.encode(prompt)) # type: ignore
|
||||||
completion_tokens = len(encoding.encode(response_string))
|
completion_tokens = len(encoding.encode(response_string))
|
||||||
model_response["usage"] = {
|
model_response["usage"] = {
|
||||||
"prompt_tokens": prompt_tokens,
|
"prompt_tokens": prompt_tokens,
|
||||||
|
@ -1284,7 +1284,7 @@ def completion(
|
||||||
)
|
)
|
||||||
|
|
||||||
"""
|
"""
|
||||||
prompt = " ".join([message["content"] for message in messages])
|
prompt = " ".join([message["content"] for message in messages]) # type: ignore
|
||||||
resp = requests.post(url, json={
|
resp = requests.post(url, json={
|
||||||
'model': model,
|
'model': model,
|
||||||
'params': {
|
'params': {
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue