mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-12 04:00:42 +00:00
update deps, mypy
This commit is contained in:
parent
16ec064cc8
commit
1fa6fe6101
2 changed files with 3 additions and 2 deletions
|
|
@ -6,7 +6,7 @@
|
||||||
|
|
||||||
from pathlib import Path
|
from pathlib import Path
|
||||||
|
|
||||||
from llama_stack.core.datatypes import BuildProvider, ModelInput, Provider, ShieldInput
|
from llama_stack.core.datatypes import BuildProvider, ModelInput, Provider, ShieldInput, ToolGroupInput
|
||||||
from llama_stack.distributions.template import DistributionTemplate, RunConfigSettings
|
from llama_stack.distributions.template import DistributionTemplate, RunConfigSettings
|
||||||
from llama_stack.providers.inline.files.localfs.config import LocalfsFilesImplConfig
|
from llama_stack.providers.inline.files.localfs.config import LocalfsFilesImplConfig
|
||||||
from llama_stack.providers.remote.datasetio.nvidia import NvidiaDatasetIOConfig
|
from llama_stack.providers.remote.datasetio.nvidia import NvidiaDatasetIOConfig
|
||||||
|
|
@ -66,7 +66,7 @@ def get_distribution_template(name: str = "nvidia") -> DistributionTemplate:
|
||||||
provider_id="nvidia",
|
provider_id="nvidia",
|
||||||
)
|
)
|
||||||
|
|
||||||
default_tool_groups = []
|
default_tool_groups: list[ToolGroupInput] = []
|
||||||
|
|
||||||
return DistributionTemplate(
|
return DistributionTemplate(
|
||||||
name=name,
|
name=name,
|
||||||
|
|
|
||||||
|
|
@ -42,6 +42,7 @@ def available_providers() -> list[ProviderSpec]:
|
||||||
# CrossEncoder depends on torchao.quantization
|
# CrossEncoder depends on torchao.quantization
|
||||||
pip_packages=[
|
pip_packages=[
|
||||||
"torch torchvision torchao>=0.12.0 --extra-index-url https://download.pytorch.org/whl/cpu",
|
"torch torchvision torchao>=0.12.0 --extra-index-url https://download.pytorch.org/whl/cpu",
|
||||||
|
"numpy tqdm transformers",
|
||||||
"sentence-transformers --no-deps",
|
"sentence-transformers --no-deps",
|
||||||
# required by some SentenceTransformers architectures for tensor rearrange/merge ops
|
# required by some SentenceTransformers architectures for tensor rearrange/merge ops
|
||||||
"einops",
|
"einops",
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue