forked from phoenix/litellm-mirror
fix(ollama_chat.py): map tool call to assistant for ollama calls
This commit is contained in:
parent
60238f866b
commit
12bb705f31
1 changed files with 8 additions and 3 deletions
|
@ -68,9 +68,9 @@ class OllamaConfig:
|
||||||
repeat_last_n: Optional[int] = None
|
repeat_last_n: Optional[int] = None
|
||||||
repeat_penalty: Optional[float] = None
|
repeat_penalty: Optional[float] = None
|
||||||
temperature: Optional[float] = None
|
temperature: Optional[float] = None
|
||||||
stop: Optional[
|
stop: Optional[list] = (
|
||||||
list
|
None # stop is a list based on this - https://github.com/jmorganca/ollama/pull/442
|
||||||
] = None # stop is a list based on this - https://github.com/jmorganca/ollama/pull/442
|
)
|
||||||
tfs_z: Optional[float] = None
|
tfs_z: Optional[float] = None
|
||||||
num_predict: Optional[int] = None
|
num_predict: Optional[int] = None
|
||||||
top_k: Optional[int] = None
|
top_k: Optional[int] = None
|
||||||
|
@ -147,6 +147,11 @@ def get_ollama_response(
|
||||||
|
|
||||||
stream = optional_params.pop("stream", False)
|
stream = optional_params.pop("stream", False)
|
||||||
format = optional_params.pop("format", None)
|
format = optional_params.pop("format", None)
|
||||||
|
|
||||||
|
for m in messages:
|
||||||
|
if "role" in m and m["role"] == "tool":
|
||||||
|
m["role"] = "assistant"
|
||||||
|
|
||||||
data = {
|
data = {
|
||||||
"model": model,
|
"model": model,
|
||||||
"messages": messages,
|
"messages": messages,
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue