mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-28 06:50:25 +00:00
Merge branch 'main' into add-watsonx-inference-adapter
This commit is contained in:
commit
8cf8bd35f8
141 changed files with 206076 additions and 6917 deletions
|
|
@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
|
|||
|
||||
[project]
|
||||
name = "llama_stack"
|
||||
version = "0.1.8"
|
||||
version = "0.2.1"
|
||||
authors = [{ name = "Meta Llama", email = "llama-oss@meta.com" }]
|
||||
description = "Llama Stack"
|
||||
readme = "README.md"
|
||||
|
|
@ -27,7 +27,7 @@ dependencies = [
|
|||
"huggingface-hub",
|
||||
"jinja2>=3.1.6",
|
||||
"jsonschema",
|
||||
"llama-stack-client>=0.1.8",
|
||||
"llama-stack-client>=0.1.9",
|
||||
"prompt-toolkit",
|
||||
"python-dotenv",
|
||||
"pydantic>=2",
|
||||
|
|
@ -79,6 +79,7 @@ docs = [
|
|||
"sphinx-autobuild",
|
||||
"myst-parser",
|
||||
"sphinx-rtd-theme",
|
||||
"sphinx_rtd_dark_mode",
|
||||
"sphinx-copybutton",
|
||||
"sphinx-tabs",
|
||||
"sphinx-design",
|
||||
|
|
@ -200,6 +201,7 @@ exclude = [
|
|||
"^llama_stack/distribution/routers/",
|
||||
"^llama_stack/distribution/server/endpoints\\.py$",
|
||||
"^llama_stack/distribution/server/server\\.py$",
|
||||
"^llama_stack/distribution/server/websocket_server\\.py$",
|
||||
"^llama_stack/distribution/stack\\.py$",
|
||||
"^llama_stack/distribution/store/registry\\.py$",
|
||||
"^llama_stack/distribution/ui/page/playground/chat\\.py$",
|
||||
|
|
@ -212,8 +214,8 @@ exclude = [
|
|||
"^llama_stack/models/llama/llama3/tokenizer\\.py$",
|
||||
"^llama_stack/models/llama/llama3/tool_utils\\.py$",
|
||||
"^llama_stack/models/llama/llama3_3/prompts\\.py$",
|
||||
"^llama_stack/models/llama/llama4/",
|
||||
"^llama_stack/models/llama/sku_list\\.py$",
|
||||
"^llama_stack/providers/datatypes\\.py$",
|
||||
"^llama_stack/providers/inline/agents/meta_reference/",
|
||||
"^llama_stack/providers/inline/agents/meta_reference/agent_instance\\.py$",
|
||||
"^llama_stack/providers/inline/agents/meta_reference/agents\\.py$",
|
||||
|
|
@ -224,6 +226,7 @@ exclude = [
|
|||
"^llama_stack/providers/inline/inference/meta_reference/inference\\.py$",
|
||||
"^llama_stack/providers/inline/inference/meta_reference/llama3/generation\\.py$",
|
||||
"^llama_stack/providers/inline/inference/meta_reference/llama3/multimodal/model\\.py$",
|
||||
"^llama_stack/providers/inline/inference/meta_reference/llama4/",
|
||||
"^llama_stack/providers/inline/inference/meta_reference/parallel_utils\\.py$",
|
||||
"^llama_stack/providers/inline/inference/meta_reference/quantization/fp8_impls\\.py$",
|
||||
"^llama_stack/providers/inline/inference/meta_reference/quantization/loader\\.py$",
|
||||
|
|
@ -247,7 +250,6 @@ exclude = [
|
|||
"^llama_stack/providers/remote/inference/gemini/",
|
||||
"^llama_stack/providers/remote/inference/groq/",
|
||||
"^llama_stack/providers/remote/inference/nvidia/",
|
||||
"^llama_stack/providers/remote/inference/ollama/",
|
||||
"^llama_stack/providers/remote/inference/openai/",
|
||||
"^llama_stack/providers/remote/inference/passthrough/",
|
||||
"^llama_stack/providers/remote/inference/runpod/",
|
||||
|
|
@ -292,6 +294,7 @@ exclude = [
|
|||
"^llama_stack/providers/utils/telemetry/dataset_mixin\\.py$",
|
||||
"^llama_stack/providers/utils/telemetry/trace_protocol\\.py$",
|
||||
"^llama_stack/providers/utils/telemetry/tracing\\.py$",
|
||||
"^llama_stack/scripts/",
|
||||
"^llama_stack/strong_typing/auxiliary\\.py$",
|
||||
"^llama_stack/strong_typing/deserializer\\.py$",
|
||||
"^llama_stack/strong_typing/inspection\\.py$",
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue