forked from phoenix-oss/llama-stack-mirror
Compare commits
1 commit
kvant
...
create-pul
Author | SHA1 | Date | |
---|---|---|---|
|
075c5401f5 |
1 changed files with 327 additions and 305 deletions
632
CHANGELOG.md
632
CHANGELOG.md
|
@ -1,5 +1,26 @@
|
||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
# v0.2.7
|
||||||
|
Published on: 2025-05-16T20:38:10Z
|
||||||
|
|
||||||
|
## Highlights
|
||||||
|
|
||||||
|
This is a small update. But a couple highlights:
|
||||||
|
|
||||||
|
* feat: function tools in OpenAI Responses by @bbrowning in https://github.com/meta-llama/llama-stack/pull/2094, getting closer to ready. Streaming is the next missing piece.
|
||||||
|
* feat: Adding support for customizing chunk context in RAG insertion and querying by @franciscojavierarceo in https://github.com/meta-llama/llama-stack/pull/2134
|
||||||
|
* feat: scaffolding for Llama Stack UI by @ehhuang in https://github.com/meta-llama/llama-stack/pull/2149, more to come in the coming releases.
|
||||||
|
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
# v0.2.6
|
||||||
|
Published on: 2025-05-12T18:06:52Z
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
# v0.2.5
|
# v0.2.5
|
||||||
Published on: 2025-05-04T20:16:49Z
|
Published on: 2025-05-04T20:16:49Z
|
||||||
|
|
||||||
|
@ -10,42 +31,42 @@ Published on: 2025-05-04T20:16:49Z
|
||||||
# v0.2.4
|
# v0.2.4
|
||||||
Published on: 2025-04-29T17:26:01Z
|
Published on: 2025-04-29T17:26:01Z
|
||||||
|
|
||||||
## Highlights
|
## Highlights
|
||||||
|
|
||||||
* One-liner to install and run Llama Stack yay! by @reluctantfuturist in https://github.com/meta-llama/llama-stack/pull/1383
|
* One-liner to install and run Llama Stack yay! by @reluctantfuturist in https://github.com/meta-llama/llama-stack/pull/1383
|
||||||
* support for NVIDIA NeMo datastore by @raspawar in https://github.com/meta-llama/llama-stack/pull/1852
|
* support for NVIDIA NeMo datastore by @raspawar in https://github.com/meta-llama/llama-stack/pull/1852
|
||||||
* (yuge!) Kubernetes authentication by @leseb in https://github.com/meta-llama/llama-stack/pull/1778
|
* (yuge!) Kubernetes authentication by @leseb in https://github.com/meta-llama/llama-stack/pull/1778
|
||||||
* (yuge!) OpenAI Responses API by @bbrowning in https://github.com/meta-llama/llama-stack/pull/1989
|
* (yuge!) OpenAI Responses API by @bbrowning in https://github.com/meta-llama/llama-stack/pull/1989
|
||||||
* add api.llama provider, llama-guard-4 model by @ashwinb in https://github.com/meta-llama/llama-stack/pull/2058
|
* add api.llama provider, llama-guard-4 model by @ashwinb in https://github.com/meta-llama/llama-stack/pull/2058
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.2.3
|
# v0.2.3
|
||||||
Published on: 2025-04-25T22:46:21Z
|
Published on: 2025-04-25T22:46:21Z
|
||||||
|
|
||||||
## Highlights
|
## Highlights
|
||||||
|
|
||||||
* OpenAI compatible inference endpoints and client-SDK support. `client.chat.completions.create()` now works.
|
* OpenAI compatible inference endpoints and client-SDK support. `client.chat.completions.create()` now works.
|
||||||
* significant improvements and functionality added to the nVIDIA distribution
|
* significant improvements and functionality added to the nVIDIA distribution
|
||||||
* many improvements to the test verification suite.
|
* many improvements to the test verification suite.
|
||||||
* new inference providers: Ramalama, IBM WatsonX
|
* new inference providers: Ramalama, IBM WatsonX
|
||||||
* many improvements to the Playground UI
|
* many improvements to the Playground UI
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.2.2
|
# v0.2.2
|
||||||
Published on: 2025-04-13T01:19:49Z
|
Published on: 2025-04-13T01:19:49Z
|
||||||
|
|
||||||
## Main changes
|
## Main changes
|
||||||
|
|
||||||
- Bring Your Own Provider (@leseb) - use out-of-tree provider code to execute the distribution server
|
- Bring Your Own Provider (@leseb) - use out-of-tree provider code to execute the distribution server
|
||||||
- OpenAI compatible inference API in progress (@bbrowning)
|
- OpenAI compatible inference API in progress (@bbrowning)
|
||||||
- Provider verifications (@ehhuang)
|
- Provider verifications (@ehhuang)
|
||||||
- Many updates and fixes to playground
|
- Many updates and fixes to playground
|
||||||
- Several llama4 related fixes
|
- Several llama4 related fixes
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
@ -59,10 +80,10 @@ Published on: 2025-04-05T23:13:00Z
|
||||||
# v0.2.0
|
# v0.2.0
|
||||||
Published on: 2025-04-05T19:04:29Z
|
Published on: 2025-04-05T19:04:29Z
|
||||||
|
|
||||||
## Llama 4 Support
|
## Llama 4 Support
|
||||||
|
|
||||||
Checkout more at https://www.llama.com
|
Checkout more at https://www.llama.com
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
@ -70,58 +91,58 @@ Checkout more at https://www.llama.com
|
||||||
# v0.1.9
|
# v0.1.9
|
||||||
Published on: 2025-03-29T00:52:23Z
|
Published on: 2025-03-29T00:52:23Z
|
||||||
|
|
||||||
### Build and Test Agents
|
### Build and Test Agents
|
||||||
* Agents: Entire document context with attachments
|
* Agents: Entire document context with attachments
|
||||||
* RAG: Documentation with sqlite-vec faiss comparison
|
* RAG: Documentation with sqlite-vec faiss comparison
|
||||||
* Getting started: Fixes to getting started notebook.
|
* Getting started: Fixes to getting started notebook.
|
||||||
|
|
||||||
### Agent Evals and Model Customization
|
### Agent Evals and Model Customization
|
||||||
* (**New**) Post-training: Add nemo customizer
|
* (**New**) Post-training: Add nemo customizer
|
||||||
|
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
* Moved sqlite-vec to non-blocking calls
|
* Moved sqlite-vec to non-blocking calls
|
||||||
* Don't return a payload on file delete
|
* Don't return a payload on file delete
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.8
|
# v0.1.8
|
||||||
Published on: 2025-03-24T01:28:50Z
|
Published on: 2025-03-24T01:28:50Z
|
||||||
|
|
||||||
# v0.1.8 Release Notes
|
# v0.1.8 Release Notes
|
||||||
|
|
||||||
### Build and Test Agents
|
### Build and Test Agents
|
||||||
* Safety: Integrated NVIDIA as a safety provider.
|
* Safety: Integrated NVIDIA as a safety provider.
|
||||||
* VectorDB: Added Qdrant as an inline provider.
|
* VectorDB: Added Qdrant as an inline provider.
|
||||||
* Agents: Added support for multiple tool groups in agents.
|
* Agents: Added support for multiple tool groups in agents.
|
||||||
* Agents: Simplified imports for Agents in client package
|
* Agents: Simplified imports for Agents in client package
|
||||||
|
|
||||||
|
|
||||||
### Agent Evals and Model Customization
|
### Agent Evals and Model Customization
|
||||||
* Introduced DocVQA and IfEval benchmarks.
|
* Introduced DocVQA and IfEval benchmarks.
|
||||||
|
|
||||||
### Deploying and Monitoring Agents
|
### Deploying and Monitoring Agents
|
||||||
* Introduced a Containerfile and image workflow for the Playground.
|
* Introduced a Containerfile and image workflow for the Playground.
|
||||||
* Implemented support for Bearer (API Key) authentication.
|
* Implemented support for Bearer (API Key) authentication.
|
||||||
* Added attribute-based access control for resources.
|
* Added attribute-based access control for resources.
|
||||||
* Fixes on docker deployments: use --pull always and standardized the default port to 8321
|
* Fixes on docker deployments: use --pull always and standardized the default port to 8321
|
||||||
* Deprecated: /v1/inspect/providers use /v1/providers/ instead
|
* Deprecated: /v1/inspect/providers use /v1/providers/ instead
|
||||||
|
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
* Consolidated scripts under the ./scripts directory.
|
* Consolidated scripts under the ./scripts directory.
|
||||||
* Addressed mypy violations in various modules.
|
* Addressed mypy violations in various modules.
|
||||||
* Added Dependabot scans for Python dependencies.
|
* Added Dependabot scans for Python dependencies.
|
||||||
* Implemented a scheduled workflow to update the changelog automatically.
|
* Implemented a scheduled workflow to update the changelog automatically.
|
||||||
* Enforced concurrency to reduce CI loads.
|
* Enforced concurrency to reduce CI loads.
|
||||||
|
|
||||||
|
|
||||||
### New Contributors
|
### New Contributors
|
||||||
* @cmodi-meta made their first contribution in https://github.com/meta-llama/llama-stack/pull/1650
|
* @cmodi-meta made their first contribution in https://github.com/meta-llama/llama-stack/pull/1650
|
||||||
* @jeffmaury made their first contribution in https://github.com/meta-llama/llama-stack/pull/1671
|
* @jeffmaury made their first contribution in https://github.com/meta-llama/llama-stack/pull/1671
|
||||||
* @derekhiggins made their first contribution in https://github.com/meta-llama/llama-stack/pull/1698
|
* @derekhiggins made their first contribution in https://github.com/meta-llama/llama-stack/pull/1698
|
||||||
* @Bobbins228 made their first contribution in https://github.com/meta-llama/llama-stack/pull/1745
|
* @Bobbins228 made their first contribution in https://github.com/meta-llama/llama-stack/pull/1745
|
||||||
|
|
||||||
**Full Changelog**: https://github.com/meta-llama/llama-stack/compare/v0.1.7...v0.1.8
|
**Full Changelog**: https://github.com/meta-llama/llama-stack/compare/v0.1.7...v0.1.8
|
||||||
|
|
||||||
---
|
---
|
||||||
|
@ -129,73 +150,73 @@ Published on: 2025-03-24T01:28:50Z
|
||||||
# v0.1.7
|
# v0.1.7
|
||||||
Published on: 2025-03-14T22:30:51Z
|
Published on: 2025-03-14T22:30:51Z
|
||||||
|
|
||||||
## 0.1.7 Release Notes
|
## 0.1.7 Release Notes
|
||||||
|
|
||||||
### Build and Test Agents
|
### Build and Test Agents
|
||||||
* Inference: ImageType is now refactored to LlamaStackImageType
|
* Inference: ImageType is now refactored to LlamaStackImageType
|
||||||
* Inference: Added tests to measure TTFT
|
* Inference: Added tests to measure TTFT
|
||||||
* Inference: Bring back usage metrics
|
* Inference: Bring back usage metrics
|
||||||
* Agents: Added endpoint for get agent, list agents and list sessions
|
* Agents: Added endpoint for get agent, list agents and list sessions
|
||||||
* Agents: Automated conversion of type hints in client tool for lite llm format
|
* Agents: Automated conversion of type hints in client tool for lite llm format
|
||||||
* Agents: Deprecated ToolResponseMessage in agent.resume API
|
* Agents: Deprecated ToolResponseMessage in agent.resume API
|
||||||
* Added Provider API for listing and inspecting provider info
|
* Added Provider API for listing and inspecting provider info
|
||||||
|
|
||||||
### Agent Evals and Model Customization
|
### Agent Evals and Model Customization
|
||||||
* Eval: Added new eval benchmarks Math 500 and BFCL v3
|
* Eval: Added new eval benchmarks Math 500 and BFCL v3
|
||||||
* Deploy and Monitoring of Agents
|
* Deploy and Monitoring of Agents
|
||||||
* Telemetry: Fix tracing to work across coroutines
|
* Telemetry: Fix tracing to work across coroutines
|
||||||
|
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
* Display code coverage for unit tests
|
* Display code coverage for unit tests
|
||||||
* Updated call sites (inference, tool calls, agents) to move to async non blocking calls
|
* Updated call sites (inference, tool calls, agents) to move to async non blocking calls
|
||||||
* Unit tests also run on Python 3.11, 3.12, and 3.13
|
* Unit tests also run on Python 3.11, 3.12, and 3.13
|
||||||
* Added ollama inference to Integration tests CI
|
* Added ollama inference to Integration tests CI
|
||||||
* Improved documentation across examples, testing, CLI, updated providers table )
|
* Improved documentation across examples, testing, CLI, updated providers table )
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.6
|
# v0.1.6
|
||||||
Published on: 2025-03-08T04:35:08Z
|
Published on: 2025-03-08T04:35:08Z
|
||||||
|
|
||||||
## 0.1.6 Release Notes
|
## 0.1.6 Release Notes
|
||||||
|
|
||||||
### Build and Test Agents
|
### Build and Test Agents
|
||||||
* Inference: Fixed support for inline vllm provider
|
* Inference: Fixed support for inline vllm provider
|
||||||
* (**New**) Agent: Build & Monitor Agent Workflows with Llama Stack + Anthropic's Best Practice [Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Agent_Workflows.ipynb)
|
* (**New**) Agent: Build & Monitor Agent Workflows with Llama Stack + Anthropic's Best Practice [Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_Agent_Workflows.ipynb)
|
||||||
* (**New**) Agent: Revamped agent [documentation](https://llama-stack.readthedocs.io/en/latest/building_applications/agent.html) with more details and examples
|
* (**New**) Agent: Revamped agent [documentation](https://llama-stack.readthedocs.io/en/latest/building_applications/agent.html) with more details and examples
|
||||||
* Agent: Unify tools and Python SDK Agents API
|
* Agent: Unify tools and Python SDK Agents API
|
||||||
* Agent: AsyncAgent Python SDK wrapper supporting async client tool calls
|
* Agent: AsyncAgent Python SDK wrapper supporting async client tool calls
|
||||||
* Agent: Support python functions without @client_tool decorator as client tools
|
* Agent: Support python functions without @client_tool decorator as client tools
|
||||||
* Agent: deprecation for allow_resume_turn flag, and remove need to specify tool_prompt_format
|
* Agent: deprecation for allow_resume_turn flag, and remove need to specify tool_prompt_format
|
||||||
* VectorIO: MilvusDB support added
|
* VectorIO: MilvusDB support added
|
||||||
|
|
||||||
### Agent Evals and Model Customization
|
### Agent Evals and Model Customization
|
||||||
* (**New**) Agent: Llama Stack RAG Lifecycle [Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_RAG_Lifecycle.ipynb)
|
* (**New**) Agent: Llama Stack RAG Lifecycle [Notebook](https://github.com/meta-llama/llama-stack/blob/main/docs/notebooks/Llama_Stack_RAG_Lifecycle.ipynb)
|
||||||
* Eval: Documentation for eval, scoring, adding new benchmarks
|
* Eval: Documentation for eval, scoring, adding new benchmarks
|
||||||
* Eval: Distribution template to run benchmarks on llama & non-llama models
|
* Eval: Distribution template to run benchmarks on llama & non-llama models
|
||||||
* Eval: Ability to register new custom LLM-as-judge scoring functions
|
* Eval: Ability to register new custom LLM-as-judge scoring functions
|
||||||
* (**New**) Looking for contributors for open benchmarks. See [documentation](https://llama-stack.readthedocs.io/en/latest/references/evals_reference/index.html#open-benchmark-contributing-guide) for details.
|
* (**New**) Looking for contributors for open benchmarks. See [documentation](https://llama-stack.readthedocs.io/en/latest/references/evals_reference/index.html#open-benchmark-contributing-guide) for details.
|
||||||
|
|
||||||
### Deploy and Monitoring of Agents
|
### Deploy and Monitoring of Agents
|
||||||
* Better support for different log levels across all components for better monitoring
|
* Better support for different log levels across all components for better monitoring
|
||||||
|
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
* Enhance OpenAPI spec to include Error types across all APIs
|
* Enhance OpenAPI spec to include Error types across all APIs
|
||||||
* Moved all tests to /tests and created unit tests to run on each PR
|
* Moved all tests to /tests and created unit tests to run on each PR
|
||||||
* Removed all dependencies on llama-models repo
|
* Removed all dependencies on llama-models repo
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.5.1
|
# v0.1.5.1
|
||||||
Published on: 2025-02-28T22:37:44Z
|
Published on: 2025-02-28T22:37:44Z
|
||||||
|
|
||||||
## 0.1.5.1 Release Notes
|
## 0.1.5.1 Release Notes
|
||||||
* Fixes for security risk in https://github.com/meta-llama/llama-stack/pull/1327 and https://github.com/meta-llama/llama-stack/pull/1328
|
* Fixes for security risk in https://github.com/meta-llama/llama-stack/pull/1327 and https://github.com/meta-llama/llama-stack/pull/1328
|
||||||
|
|
||||||
**Full Changelog**: https://github.com/meta-llama/llama-stack/compare/v0.1.5...v0.1.5.1
|
**Full Changelog**: https://github.com/meta-llama/llama-stack/compare/v0.1.5...v0.1.5.1
|
||||||
|
|
||||||
---
|
---
|
||||||
|
@ -203,176 +224,176 @@ Published on: 2025-02-28T22:37:44Z
|
||||||
# v0.1.5
|
# v0.1.5
|
||||||
Published on: 2025-02-28T18:14:01Z
|
Published on: 2025-02-28T18:14:01Z
|
||||||
|
|
||||||
## 0.1.5 Release Notes
|
## 0.1.5 Release Notes
|
||||||
### Build Agents
|
### Build Agents
|
||||||
* Inference: Support more non-llama models (openai, anthropic, gemini)
|
* Inference: Support more non-llama models (openai, anthropic, gemini)
|
||||||
* Inference: Can use the provider's model name in addition to the HF alias
|
* Inference: Can use the provider's model name in addition to the HF alias
|
||||||
* Inference: Fixed issues with calling tools that weren't specified in the prompt
|
* Inference: Fixed issues with calling tools that weren't specified in the prompt
|
||||||
* RAG: Improved system prompt for RAG and no more need for hard-coded rag-tool calling
|
* RAG: Improved system prompt for RAG and no more need for hard-coded rag-tool calling
|
||||||
* Embeddings: Added support for Nemo retriever embedding models
|
* Embeddings: Added support for Nemo retriever embedding models
|
||||||
* Tools: Added support for MCP tools in Ollama Distribution
|
* Tools: Added support for MCP tools in Ollama Distribution
|
||||||
* Distributions: Added new Groq distribution
|
* Distributions: Added new Groq distribution
|
||||||
|
|
||||||
### Customize Models
|
### Customize Models
|
||||||
* Save post-trained checkpoint in SafeTensor format to allow Ollama inference provider to use the post-trained model
|
* Save post-trained checkpoint in SafeTensor format to allow Ollama inference provider to use the post-trained model
|
||||||
|
|
||||||
### Monitor agents
|
### Monitor agents
|
||||||
* More comprehensive logging of agent steps including client tools
|
* More comprehensive logging of agent steps including client tools
|
||||||
* Telemetry inputs/outputs are now structured and queryable
|
* Telemetry inputs/outputs are now structured and queryable
|
||||||
* Ability to retrieve agents session, turn, step by ids
|
* Ability to retrieve agents session, turn, step by ids
|
||||||
|
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
* Moved executorch Swift code out of this repo into the llama-stack-client-swift repo, similar to kotlin
|
* Moved executorch Swift code out of this repo into the llama-stack-client-swift repo, similar to kotlin
|
||||||
* Move most logging to use logger instead of prints
|
* Move most logging to use logger instead of prints
|
||||||
* Completed text /chat-completion and /completion tests
|
* Completed text /chat-completion and /completion tests
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.4
|
# v0.1.4
|
||||||
Published on: 2025-02-25T00:02:43Z
|
Published on: 2025-02-25T00:02:43Z
|
||||||
|
|
||||||
## v0.1.4 Release Notes
|
## v0.1.4 Release Notes
|
||||||
Here are the key changes coming as part of this release:
|
Here are the key changes coming as part of this release:
|
||||||
|
|
||||||
### Build and Test Agents
|
### Build and Test Agents
|
||||||
* Inference: Added support for non-llama models
|
* Inference: Added support for non-llama models
|
||||||
* Inference: Added option to list all downloaded models and remove models
|
* Inference: Added option to list all downloaded models and remove models
|
||||||
* Agent: Introduce new api agents.resume_turn to include client side tool execution in the same turn
|
* Agent: Introduce new api agents.resume_turn to include client side tool execution in the same turn
|
||||||
* Agent: AgentConfig introduces new variable “tool_config” that allows for better tool configuration and system prompt overrides
|
* Agent: AgentConfig introduces new variable “tool_config” that allows for better tool configuration and system prompt overrides
|
||||||
* Agent: Added logging for agent step start and completion times
|
* Agent: Added logging for agent step start and completion times
|
||||||
* Agent: Added support for logging for tool execution metadata
|
* Agent: Added support for logging for tool execution metadata
|
||||||
* Embedding: Updated /inference/embeddings to support asymmetric models, truncation and variable sized outputs
|
* Embedding: Updated /inference/embeddings to support asymmetric models, truncation and variable sized outputs
|
||||||
* Embedding: Updated embedding models for Ollama, Together, and Fireworks with available defaults
|
* Embedding: Updated embedding models for Ollama, Together, and Fireworks with available defaults
|
||||||
* VectorIO: Improved performance of sqlite-vec using chunked writes
|
* VectorIO: Improved performance of sqlite-vec using chunked writes
|
||||||
### Agent Evals and Model Customization
|
### Agent Evals and Model Customization
|
||||||
* Deprecated api /eval-tasks. Use /eval/benchmark instead
|
* Deprecated api /eval-tasks. Use /eval/benchmark instead
|
||||||
* Added CPU training support for TorchTune
|
* Added CPU training support for TorchTune
|
||||||
### Deploy and Monitoring of Agents
|
### Deploy and Monitoring of Agents
|
||||||
* Consistent view of client and server tool calls in telemetry
|
* Consistent view of client and server tool calls in telemetry
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
* Made tests more data-driven for consistent evaluation
|
* Made tests more data-driven for consistent evaluation
|
||||||
* Fixed documentation links and improved API reference generation
|
* Fixed documentation links and improved API reference generation
|
||||||
* Various small fixes for build scripts and system reliability
|
* Various small fixes for build scripts and system reliability
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.3
|
# v0.1.3
|
||||||
Published on: 2025-02-14T20:24:32Z
|
Published on: 2025-02-14T20:24:32Z
|
||||||
|
|
||||||
## v0.1.3 Release
|
## v0.1.3 Release
|
||||||
|
|
||||||
Here are some key changes that are coming as part of this release.
|
Here are some key changes that are coming as part of this release.
|
||||||
|
|
||||||
### Build and Test Agents
|
### Build and Test Agents
|
||||||
Streamlined the initial development experience
|
Streamlined the initial development experience
|
||||||
- Added support for llama stack run --image-type venv
|
- Added support for llama stack run --image-type venv
|
||||||
- Enhanced vector store options with new sqlite-vec provider and improved Qdrant integration
|
- Enhanced vector store options with new sqlite-vec provider and improved Qdrant integration
|
||||||
- vLLM improvements for tool calling and logprobs
|
- vLLM improvements for tool calling and logprobs
|
||||||
- Better handling of sporadic code_interpreter tool calls
|
- Better handling of sporadic code_interpreter tool calls
|
||||||
|
|
||||||
### Agent Evals
|
### Agent Evals
|
||||||
Better benchmarking and Agent performance assessment
|
Better benchmarking and Agent performance assessment
|
||||||
- Renamed eval API /eval-task to /benchmarks
|
- Renamed eval API /eval-task to /benchmarks
|
||||||
- Improved documentation and notebooks for RAG and evals
|
- Improved documentation and notebooks for RAG and evals
|
||||||
|
|
||||||
### Deploy and Monitoring of Agents
|
### Deploy and Monitoring of Agents
|
||||||
Improved production readiness
|
Improved production readiness
|
||||||
- Added usage metrics collection for chat completions
|
- Added usage metrics collection for chat completions
|
||||||
- CLI improvements for provider information
|
- CLI improvements for provider information
|
||||||
- Improved error handling and system reliability
|
- Improved error handling and system reliability
|
||||||
- Better model endpoint handling and accessibility
|
- Better model endpoint handling and accessibility
|
||||||
- Improved signal handling on distro server
|
- Improved signal handling on distro server
|
||||||
|
|
||||||
### Better Engineering
|
### Better Engineering
|
||||||
Infrastructure and code quality improvements
|
Infrastructure and code quality improvements
|
||||||
- Faster text-based chat completion tests
|
- Faster text-based chat completion tests
|
||||||
- Improved testing for non-streaming agent apis
|
- Improved testing for non-streaming agent apis
|
||||||
- Standardized import formatting with ruff linter
|
- Standardized import formatting with ruff linter
|
||||||
- Added conventional commits standard
|
- Added conventional commits standard
|
||||||
- Fixed documentation parsing issues
|
- Fixed documentation parsing issues
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.2
|
# v0.1.2
|
||||||
Published on: 2025-02-07T22:06:49Z
|
Published on: 2025-02-07T22:06:49Z
|
||||||
|
|
||||||
# TL;DR
|
# TL;DR
|
||||||
- Several stabilizations to development flows after the switch to `uv`
|
- Several stabilizations to development flows after the switch to `uv`
|
||||||
- Migrated CI workflows to new OSS repo - [llama-stack-ops](https://github.com/meta-llama/llama-stack-ops)
|
- Migrated CI workflows to new OSS repo - [llama-stack-ops](https://github.com/meta-llama/llama-stack-ops)
|
||||||
- Added automated rebuilds for ReadTheDocs
|
- Added automated rebuilds for ReadTheDocs
|
||||||
- Llama Stack server supports HTTPS
|
- Llama Stack server supports HTTPS
|
||||||
- Added system prompt overrides support
|
- Added system prompt overrides support
|
||||||
- Several bug fixes and improvements to documentation (check out Kubernetes deployment guide by @terrytangyuan )
|
- Several bug fixes and improvements to documentation (check out Kubernetes deployment guide by @terrytangyuan )
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.1
|
# v0.1.1
|
||||||
Published on: 2025-02-02T02:29:24Z
|
Published on: 2025-02-02T02:29:24Z
|
||||||
|
|
||||||
A bunch of small / big improvements everywhere including support for Windows, switching to `uv` and many provider improvements.
|
A bunch of small / big improvements everywhere including support for Windows, switching to `uv` and many provider improvements.
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
# v0.1.0
|
# v0.1.0
|
||||||
Published on: 2025-01-24T17:47:47Z
|
Published on: 2025-01-24T17:47:47Z
|
||||||
|
|
||||||
We are excited to announce a stable API release of Llama Stack, which enables developers to build RAG applications and Agents using tools and safety shields, monitor and those agents with telemetry, and evaluate the agent with scoring functions.
|
We are excited to announce a stable API release of Llama Stack, which enables developers to build RAG applications and Agents using tools and safety shields, monitor and those agents with telemetry, and evaluate the agent with scoring functions.
|
||||||
|
|
||||||
## Context
|
## Context
|
||||||
GenAI application developers need more than just an LLM - they need to integrate tools, connect with their data sources, establish guardrails, and ground the LLM responses effectively. Currently, developers must piece together various tools and APIs, complicating the development lifecycle and increasing costs. The result is that developers are spending more time on these integrations rather than focusing on the application logic itself. The bespoke coupling of components also makes it challenging to adopt state-of-the-art solutions in the rapidly evolving GenAI space. This is particularly difficult for open models like Llama, as best practices are not widely established in the open.
|
GenAI application developers need more than just an LLM - they need to integrate tools, connect with their data sources, establish guardrails, and ground the LLM responses effectively. Currently, developers must piece together various tools and APIs, complicating the development lifecycle and increasing costs. The result is that developers are spending more time on these integrations rather than focusing on the application logic itself. The bespoke coupling of components also makes it challenging to adopt state-of-the-art solutions in the rapidly evolving GenAI space. This is particularly difficult for open models like Llama, as best practices are not widely established in the open.
|
||||||
|
|
||||||
Llama Stack was created to provide developers with a comprehensive and coherent interface that simplifies AI application development and codifies best practices across the Llama ecosystem. Since our launch in September 2024, we have seen a huge uptick in interest in Llama Stack APIs by both AI developers and from partners building AI services with Llama models. Partners like Nvidia, Fireworks, and Ollama have collaborated with us to develop implementations across various APIs, including inference, memory, and safety.
|
Llama Stack was created to provide developers with a comprehensive and coherent interface that simplifies AI application development and codifies best practices across the Llama ecosystem. Since our launch in September 2024, we have seen a huge uptick in interest in Llama Stack APIs by both AI developers and from partners building AI services with Llama models. Partners like Nvidia, Fireworks, and Ollama have collaborated with us to develop implementations across various APIs, including inference, memory, and safety.
|
||||||
|
|
||||||
With Llama Stack, you can easily build a RAG agent which can also search the web, do complex math, and custom tool calling. You can use telemetry to inspect those traces, and convert telemetry into evals datasets. And with Llama Stack’s plugin architecture and prepackage distributions, you choose to run your agent anywhere - in the cloud with our partners, deploy your own environment using virtualenv, conda, or Docker, operate locally with Ollama, or even run on mobile devices with our SDKs. Llama Stack offers unprecedented flexibility while also simplifying the developer experience.
|
With Llama Stack, you can easily build a RAG agent which can also search the web, do complex math, and custom tool calling. You can use telemetry to inspect those traces, and convert telemetry into evals datasets. And with Llama Stack’s plugin architecture and prepackage distributions, you choose to run your agent anywhere - in the cloud with our partners, deploy your own environment using virtualenv, conda, or Docker, operate locally with Ollama, or even run on mobile devices with our SDKs. Llama Stack offers unprecedented flexibility while also simplifying the developer experience.
|
||||||
|
|
||||||
## Release
|
## Release
|
||||||
After iterating on the APIs for the last 3 months, today we’re launching a stable release (V1) of the Llama Stack APIs and the corresponding llama-stack server and client packages(v0.1.0). We now have automated tests for providers. These tests make sure that all provider implementations are verified. Developers can now easily and reliably select distributions or providers based on their specific requirements.
|
After iterating on the APIs for the last 3 months, today we’re launching a stable release (V1) of the Llama Stack APIs and the corresponding llama-stack server and client packages(v0.1.0). We now have automated tests for providers. These tests make sure that all provider implementations are verified. Developers can now easily and reliably select distributions or providers based on their specific requirements.
|
||||||
|
|
||||||
There are example standalone apps in llama-stack-apps.
|
There are example standalone apps in llama-stack-apps.
|
||||||
|
|
||||||
|
|
||||||
## Key Features of this release
|
## Key Features of this release
|
||||||
|
|
||||||
- **Unified API Layer**
|
- **Unified API Layer**
|
||||||
- Inference: Run LLM models
|
- Inference: Run LLM models
|
||||||
- RAG: Store and retrieve knowledge for RAG
|
- RAG: Store and retrieve knowledge for RAG
|
||||||
- Agents: Build multi-step agentic workflows
|
- Agents: Build multi-step agentic workflows
|
||||||
- Tools: Register tools that can be called by the agent
|
- Tools: Register tools that can be called by the agent
|
||||||
- Safety: Apply content filtering and safety policies
|
- Safety: Apply content filtering and safety policies
|
||||||
- Evaluation: Test model and agent quality
|
- Evaluation: Test model and agent quality
|
||||||
- Telemetry: Collect and analyze usage data and complex agentic traces
|
- Telemetry: Collect and analyze usage data and complex agentic traces
|
||||||
- Post Training ( Coming Soon ): Fine tune models for specific use cases
|
- Post Training ( Coming Soon ): Fine tune models for specific use cases
|
||||||
|
|
||||||
- **Rich Provider Ecosystem**
|
- **Rich Provider Ecosystem**
|
||||||
- Local Development: Meta's Reference, Ollama
|
- Local Development: Meta's Reference, Ollama
|
||||||
- Cloud: Fireworks, Together, Nvidia, AWS Bedrock, Groq, Cerebras
|
- Cloud: Fireworks, Together, Nvidia, AWS Bedrock, Groq, Cerebras
|
||||||
- On-premises: Nvidia NIM, vLLM, TGI, Dell-TGI
|
- On-premises: Nvidia NIM, vLLM, TGI, Dell-TGI
|
||||||
- On-device: iOS and Android support
|
- On-device: iOS and Android support
|
||||||
|
|
||||||
- **Built for Production**
|
- **Built for Production**
|
||||||
- Pre-packaged distributions for common deployment scenarios
|
- Pre-packaged distributions for common deployment scenarios
|
||||||
- Backwards compatibility across model versions
|
- Backwards compatibility across model versions
|
||||||
- Comprehensive evaluation capabilities
|
- Comprehensive evaluation capabilities
|
||||||
- Full observability and monitoring
|
- Full observability and monitoring
|
||||||
|
|
||||||
- **Multiple developer interfaces**
|
- **Multiple developer interfaces**
|
||||||
- CLI: Command line interface
|
- CLI: Command line interface
|
||||||
- Python SDK
|
- Python SDK
|
||||||
- Swift iOS SDK
|
- Swift iOS SDK
|
||||||
- Kotlin Android SDK
|
- Kotlin Android SDK
|
||||||
|
|
||||||
- **Sample llama stack applications**
|
- **Sample llama stack applications**
|
||||||
- Python
|
- Python
|
||||||
- iOS
|
- iOS
|
||||||
- Android
|
- Android
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
@ -386,8 +407,8 @@ Published on: 2025-01-22T22:24:01Z
|
||||||
# v0.0.63
|
# v0.0.63
|
||||||
Published on: 2024-12-18T07:17:43Z
|
Published on: 2024-12-18T07:17:43Z
|
||||||
|
|
||||||
A small but important bug-fix release to update the URL datatype for the client-SDKs. The issue affected multimodal agentic turns especially.
|
A small but important bug-fix release to update the URL datatype for the client-SDKs. The issue affected multimodal agentic turns especially.
|
||||||
|
|
||||||
**Full Changelog**: https://github.com/meta-llama/llama-stack/compare/v0.0.62...v0.0.63
|
**Full Changelog**: https://github.com/meta-llama/llama-stack/compare/v0.0.62...v0.0.63
|
||||||
|
|
||||||
---
|
---
|
||||||
|
@ -423,39 +444,40 @@ Published on: 2024-11-22T00:36:09Z
|
||||||
# v0.0.53
|
# v0.0.53
|
||||||
Published on: 2024-11-20T22:18:00Z
|
Published on: 2024-11-20T22:18:00Z
|
||||||
|
|
||||||
🚀 Initial Release Notes for Llama Stack!
|
🚀 Initial Release Notes for Llama Stack!
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
- Resource-oriented design for models, shields, memory banks, datasets and eval tasks
|
- Resource-oriented design for models, shields, memory banks, datasets and eval tasks
|
||||||
- Persistence for registered objects with distribution
|
- Persistence for registered objects with distribution
|
||||||
- Ability to persist memory banks created for FAISS
|
- Ability to persist memory banks created for FAISS
|
||||||
- PostgreSQL KVStore implementation
|
- PostgreSQL KVStore implementation
|
||||||
- Environment variable placeholder support in run.yaml files
|
- Environment variable placeholder support in run.yaml files
|
||||||
- Comprehensive Zero-to-Hero notebooks and quickstart guides
|
- Comprehensive Zero-to-Hero notebooks and quickstart guides
|
||||||
- Support for quantized models in Ollama
|
- Support for quantized models in Ollama
|
||||||
- Vision models support for Together, Fireworks, Meta-Reference, and Ollama, and vLLM
|
- Vision models support for Together, Fireworks, Meta-Reference, and Ollama, and vLLM
|
||||||
- Bedrock distribution with safety shields support
|
- Bedrock distribution with safety shields support
|
||||||
- Evals API with task registration and scoring functions
|
- Evals API with task registration and scoring functions
|
||||||
- MMLU and SimpleQA benchmark scoring functions
|
- MMLU and SimpleQA benchmark scoring functions
|
||||||
- Huggingface dataset provider integration for benchmarks
|
- Huggingface dataset provider integration for benchmarks
|
||||||
- Support for custom dataset registration from local paths
|
- Support for custom dataset registration from local paths
|
||||||
- Benchmark evaluation CLI tools with visualization tables
|
- Benchmark evaluation CLI tools with visualization tables
|
||||||
- RAG evaluation scoring functions and metrics
|
- RAG evaluation scoring functions and metrics
|
||||||
- Local persistence for datasets and eval tasks
|
- Local persistence for datasets and eval tasks
|
||||||
|
|
||||||
### Changed
|
### Changed
|
||||||
- Split safety into distinct providers (llama-guard, prompt-guard, code-scanner)
|
- Split safety into distinct providers (llama-guard, prompt-guard, code-scanner)
|
||||||
- Changed provider naming convention (`impls` → `inline`, `adapters` → `remote`)
|
- Changed provider naming convention (`impls` → `inline`, `adapters` → `remote`)
|
||||||
- Updated API signatures for dataset and eval task registration
|
- Updated API signatures for dataset and eval task registration
|
||||||
- Restructured folder organization for providers
|
- Restructured folder organization for providers
|
||||||
- Enhanced Docker build configuration
|
- Enhanced Docker build configuration
|
||||||
- Added version prefixing for REST API routes
|
- Added version prefixing for REST API routes
|
||||||
- Enhanced evaluation task registration workflow
|
- Enhanced evaluation task registration workflow
|
||||||
- Improved benchmark evaluation output formatting
|
- Improved benchmark evaluation output formatting
|
||||||
- Restructured evals folder organization for better modularity
|
- Restructured evals folder organization for better modularity
|
||||||
|
|
||||||
### Removed
|
### Removed
|
||||||
- `llama stack configure` command
|
- `llama stack configure` command
|
||||||
|
|
||||||
|
|
||||||
---
|
---
|
||||||
|
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue