mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 20:14:13 +00:00
split batch_inference from inference
This commit is contained in:
parent
986a865e62
commit
dc433f6c90
5 changed files with 75 additions and 12 deletions
|
@ -9,6 +9,7 @@ from llama_toolchain.agentic_system.api import * # noqa: F403
|
|||
from llama_toolchain.dataset.api import * # noqa: F403
|
||||
from llama_toolchain.evaluations.api import * # noqa: F403
|
||||
from llama_toolchain.inference.api import * # noqa: F403
|
||||
from llama_toolchain.batch_inference.api import * # noqa: F403
|
||||
from llama_toolchain.memory.api import * # noqa: F403
|
||||
from llama_toolchain.observability.api import * # noqa: F403
|
||||
from llama_toolchain.post_training.api import * # noqa: F403
|
||||
|
@ -18,6 +19,7 @@ from llama_toolchain.synthetic_data_generation.api import * # noqa: F403
|
|||
|
||||
class LlamaStack(
|
||||
Inference,
|
||||
BatchInference,
|
||||
AgenticSystem,
|
||||
RewardScoring,
|
||||
SyntheticDataGeneration,
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue