forked from phoenix-oss/llama-stack-mirror
Add function for stopping inference (#224)
This commit is contained in:
parent
6b094b72d3
commit
0f66ae0f61
1 changed files with 4 additions and 0 deletions
|
@ -34,6 +34,10 @@ public class LocalInference: Inference {
|
|||
}
|
||||
}
|
||||
|
||||
public func stop() {
|
||||
runnerHolder.runner?.stop()
|
||||
}
|
||||
|
||||
public func chatCompletion(request: Components.Schemas.ChatCompletionRequest) -> AsyncStream<Components.Schemas.ChatCompletionResponseStreamChunk> {
|
||||
return AsyncStream { continuation in
|
||||
runnerQueue.async {
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue