mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-06-28 02:53:30 +00:00
Update telemetry to have TEXT be the default log format
This commit is contained in:
parent
945db5dac2
commit
a0a00f1345
3 changed files with 4 additions and 5 deletions
|
@ -1,5 +1,4 @@
|
|||
# Llama Stack Distributions
|
||||
|
||||
# Building Llama Stacks
|
||||
|
||||
```{toctree}
|
||||
:maxdepth: 2
|
||||
|
@ -12,6 +11,7 @@ ondevice_distro/index
|
|||
## Introduction
|
||||
|
||||
Llama Stack Distributions are pre-built Docker containers/Conda environments that assemble APIs and Providers to provide a consistent whole to the end application developer.
|
||||
|
||||
These distributions allow you to mix-and-match providers - some could be backed by local code and some could be remote. This flexibility enables you to choose the optimal setup for your use case, such as serving a small model locally while using a cloud provider for larger models, all while maintaining a consistent API interface for your application.
|
||||
|
||||
|
||||
|
|
|
@ -149,7 +149,6 @@ if __name__ == "__main__":
|
|||
|
||||
## Next Steps
|
||||
|
||||
- You can mix and match different providers for inference, memory, agents, evals etc. See [Building custom distributions](../distributions/index.md)
|
||||
- [Developer Cookbook](developer_cookbook.md)
|
||||
You can mix and match different providers for inference, memory, agents, evals etc. See [Building Llama Stacks](../distributions/index.md)
|
||||
|
||||
For example applications and more detailed tutorials, visit our [llama-stack-apps](https://github.com/meta-llama/llama-stack-apps/tree/main/examples) repository.
|
||||
|
|
|
@ -18,4 +18,4 @@ class LogFormat(Enum):
|
|||
|
||||
@json_schema_type
|
||||
class ConsoleConfig(BaseModel):
|
||||
log_format: LogFormat = LogFormat.JSON
|
||||
log_format: LogFormat = LogFormat.TEXT
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue