mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-17 17:09:51 +00:00
fix for pre-commit
This commit is contained in:
parent
21d8c70954
commit
fff18151e6
2 changed files with 7 additions and 18 deletions
|
|
@ -4,12 +4,12 @@
|
||||||
|
|
||||||
Agents API for creating and interacting with agentic systems.
|
Agents API for creating and interacting with agentic systems.
|
||||||
|
|
||||||
Main functionalities provided by this API:
|
Main functionalities provided by this API:
|
||||||
- Create agents with specific instructions and ability to use tools.
|
- Create agents with specific instructions and ability to use tools.
|
||||||
- Interactions with agents are grouped into sessions ("threads"), and each interaction is called a "turn".
|
- Interactions with agents are grouped into sessions ("threads"), and each interaction is called a "turn".
|
||||||
- Agents can be provided with various tools (see the ToolGroups and ToolRuntime APIs for more details).
|
- Agents can be provided with various tools (see the ToolGroups and ToolRuntime APIs for more details).
|
||||||
- Agents can be provided with various shields (see the Safety API for more details).
|
- Agents can be provided with various shields (see the Safety API for more details).
|
||||||
- Agents can also use Memory to retrieve information from knowledge bases. See the RAG Tool and Vector IO APIs for more details.
|
- Agents can also use Memory to retrieve information from knowledge bases. See the RAG Tool and Vector IO APIs for more details.
|
||||||
|
|
||||||
This section contains documentation for all available providers for the **agents** API.
|
This section contains documentation for all available providers for the **agents** API.
|
||||||
|
|
||||||
|
|
|
||||||
|
|
@ -6,9 +6,7 @@
|
||||||
|
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
from openai.types.chat import ChatCompletionContentPartImageParam, ChatCompletionContentPartTextParam
|
from llama_stack.apis.inference import ChatCompletionRequest
|
||||||
|
|
||||||
from llama_stack.apis.inference import ChatCompletionRequest, RerankResponse
|
|
||||||
from llama_stack.providers.utils.inference.litellm_openai_mixin import (
|
from llama_stack.providers.utils.inference.litellm_openai_mixin import (
|
||||||
LiteLLMOpenAIMixin,
|
LiteLLMOpenAIMixin,
|
||||||
)
|
)
|
||||||
|
|
@ -52,12 +50,3 @@ class VertexAIInferenceAdapter(LiteLLMOpenAIMixin):
|
||||||
params.pop("api_key", None)
|
params.pop("api_key", None)
|
||||||
|
|
||||||
return params
|
return params
|
||||||
|
|
||||||
async def rerank(
|
|
||||||
self,
|
|
||||||
model: str,
|
|
||||||
query: str | ChatCompletionContentPartTextParam | ChatCompletionContentPartImageParam,
|
|
||||||
items: list[str | ChatCompletionContentPartTextParam | ChatCompletionContentPartImageParam],
|
|
||||||
max_num_results: int | None = None,
|
|
||||||
) -> RerankResponse:
|
|
||||||
raise NotImplementedError("Reranking is not supported for Vertex AI")
|
|
||||||
|
|
|
||||||
Loading…
Add table
Add a link
Reference in a new issue