mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 12:07:34 +00:00
getting closer to a distro definition, distro install + configure works
This commit is contained in:
parent
dac2b5a1ed
commit
041cafbee3
11 changed files with 471 additions and 130 deletions
|
@ -16,8 +16,8 @@ from .api.datatypes import (
|
|||
ToolCallParseStatus,
|
||||
)
|
||||
from .api.endpoints import (
|
||||
ChatCompletionResponse,
|
||||
ChatCompletionRequest,
|
||||
ChatCompletionResponse,
|
||||
ChatCompletionResponseStreamChunk,
|
||||
CompletionRequest,
|
||||
Inference,
|
||||
|
@ -25,6 +25,13 @@ from .api.endpoints import (
|
|||
from .model_parallel import LlamaModelParallelGenerator
|
||||
|
||||
|
||||
def get_adapter_impl(config: InlineImplConfig) -> Inference:
|
||||
assert isinstance(
|
||||
config, InlineImplConfig
|
||||
), f"Unexpected config type: {type(config)}"
|
||||
return InferenceImpl(config)
|
||||
|
||||
|
||||
class InferenceImpl(Inference):
|
||||
|
||||
def __init__(self, config: InlineImplConfig) -> None:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue