mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-29 16:14:45 +00:00
fold openai responses into the Agents API
This commit is contained in:
parent
207224a811
commit
abd6280cb8
25 changed files with 967 additions and 199 deletions
|
|
@ -16,7 +16,6 @@ from llama_stack.apis.files import Files
|
|||
from llama_stack.apis.inference import Inference
|
||||
from llama_stack.apis.inspect import Inspect
|
||||
from llama_stack.apis.models import Models
|
||||
from llama_stack.apis.openai_responses.openai_responses import OpenAIResponses
|
||||
from llama_stack.apis.post_training import PostTraining
|
||||
from llama_stack.apis.providers import Providers as ProvidersAPI
|
||||
from llama_stack.apis.safety import Safety
|
||||
|
|
@ -81,7 +80,6 @@ def api_protocol_map() -> Dict[Api, Any]:
|
|||
Api.tool_groups: ToolGroups,
|
||||
Api.tool_runtime: ToolRuntime,
|
||||
Api.files: Files,
|
||||
Api.openai_responses: OpenAIResponses,
|
||||
}
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -149,8 +149,6 @@ class CommonRoutingTableImpl(RoutingTable):
|
|||
p.benchmark_store = self
|
||||
elif api == Api.tool_runtime:
|
||||
p.tool_store = self
|
||||
elif api == Api.openai_responses:
|
||||
p.model_store = self
|
||||
|
||||
async def shutdown(self) -> None:
|
||||
for p in self.impls_by_provider_id.values():
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue