From a0a00f13455963240379291bb0394ae9ae6e1e4e Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 21 Nov 2024 15:17:37 -0800 Subject: [PATCH 1/3] Update telemetry to have TEXT be the default log format --- docs/source/distributions/index.md | 4 ++-- docs/source/getting_started/index.md | 3 +-- .../providers/inline/meta_reference/telemetry/config.py | 2 +- 3 files changed, 4 insertions(+), 5 deletions(-) diff --git a/docs/source/distributions/index.md b/docs/source/distributions/index.md index bedc9706e..4b66a5fc8 100644 --- a/docs/source/distributions/index.md +++ b/docs/source/distributions/index.md @@ -1,5 +1,4 @@ -# Llama Stack Distributions - +# Building Llama Stacks ```{toctree} :maxdepth: 2 @@ -12,6 +11,7 @@ ondevice_distro/index ## Introduction Llama Stack Distributions are pre-built Docker containers/Conda environments that assemble APIs and Providers to provide a consistent whole to the end application developer. + These distributions allow you to mix-and-match providers - some could be backed by local code and some could be remote. This flexibility enables you to choose the optimal setup for your use case, such as serving a small model locally while using a cloud provider for larger models, all while maintaining a consistent API interface for your application. diff --git a/docs/source/getting_started/index.md b/docs/source/getting_started/index.md index 5875f2776..72f651fd5 100644 --- a/docs/source/getting_started/index.md +++ b/docs/source/getting_started/index.md @@ -149,7 +149,6 @@ if __name__ == "__main__": ## Next Steps -- You can mix and match different providers for inference, memory, agents, evals etc. See [Building custom distributions](../distributions/index.md) -- [Developer Cookbook](developer_cookbook.md) +You can mix and match different providers for inference, memory, agents, evals etc. See [Building Llama Stacks](../distributions/index.md) For example applications and more detailed tutorials, visit our [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repository. diff --git a/llama_stack/providers/inline/meta_reference/telemetry/config.py b/llama_stack/providers/inline/meta_reference/telemetry/config.py index 34d5bc08e..a1db1d4d8 100644 --- a/llama_stack/providers/inline/meta_reference/telemetry/config.py +++ b/llama_stack/providers/inline/meta_reference/telemetry/config.py @@ -18,4 +18,4 @@ class LogFormat(Enum): @json_schema_type class ConsoleConfig(BaseModel): - log_format: LogFormat = LogFormat.JSON + log_format: LogFormat = LogFormat.TEXT From c1025ebfdb767786570891950d6ee062e1605baa Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 21 Nov 2024 15:20:06 -0800 Subject: [PATCH 2/3] Delete some dead code --- .../inline/agents/meta_reference/agent_instance.py | 6 ------ 1 file changed, 6 deletions(-) diff --git a/llama_stack/providers/inline/agents/meta_reference/agent_instance.py b/llama_stack/providers/inline/agents/meta_reference/agent_instance.py index b9737054a..e1713c0e3 100644 --- a/llama_stack/providers/inline/agents/meta_reference/agent_instance.py +++ b/llama_stack/providers/inline/agents/meta_reference/agent_instance.py @@ -396,12 +396,6 @@ class ChatAgent(ShieldRunnerMixin): n_iter = 0 while True: msg = input_messages[-1] - if msg.role == Role.user.value: - color = "blue" - elif msg.role == Role.ipython.value: - color = "yellow" - else: - color = None if len(str(msg)) > 1000: msg_str = f"{str(msg)[:500]}......{str(msg)[-500:]}" else: From 2137b0af40741aae8d8b2d1c4274066200590382 Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Thu, 21 Nov 2024 16:28:30 -0800 Subject: [PATCH 3/3] Bump version to 0.0.54 --- requirements.txt | 4 ++-- setup.py | 2 +- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/requirements.txt b/requirements.txt index fddf51880..9aa8ebc76 100644 --- a/requirements.txt +++ b/requirements.txt @@ -2,8 +2,8 @@ blobfile fire httpx huggingface-hub -llama-models>=0.0.53 -llama-stack-client>=0.0.53 +llama-models>=0.0.54 +llama-stack-client>=0.0.54 prompt-toolkit python-dotenv pydantic>=2 diff --git a/setup.py b/setup.py index 13f389a11..bf013b77a 100644 --- a/setup.py +++ b/setup.py @@ -16,7 +16,7 @@ def read_requirements(): setup( name="llama_stack", - version="0.0.53", + version="0.0.54", author="Meta Llama", author_email="llama-oss@meta.com", description="Llama Stack",