forked from phoenix-oss/llama-stack-mirror
add completion() for ollama (#280)
This commit is contained in:
parent
e2a5a2e10d
commit
1d241bf3fe
5 changed files with 138 additions and 15 deletions
|
@ -4,6 +4,10 @@ providers:
|
|||
config:
|
||||
host: localhost
|
||||
port: 11434
|
||||
- provider_id: meta-reference
|
||||
provider_type: meta-reference
|
||||
config:
|
||||
model: Llama3.2-1B-Instruct
|
||||
- provider_id: test-tgi
|
||||
provider_type: remote::tgi
|
||||
config:
|
||||
|
|
|
@ -132,7 +132,10 @@ async def test_completion(inference_settings):
|
|||
params = inference_settings["common_params"]
|
||||
|
||||
provider = inference_impl.routing_table.get_provider_impl(params["model"])
|
||||
if provider.__provider_id__ != "meta-reference":
|
||||
if provider.__provider_spec__.provider_type not in (
|
||||
"meta-reference",
|
||||
"remote::ollama",
|
||||
):
|
||||
pytest.skip("Other inference providers don't support completion() yet")
|
||||
|
||||
response = await inference_impl.completion(
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue