mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-12-03 18:00:36 +00:00
Merge remote-tracking branch 'upstream/main' into add-mcp-authentication-param
This commit is contained in:
commit
c2bf725a77
275 changed files with 1187 additions and 745 deletions
|
|
@ -13,6 +13,7 @@ from contextlib import contextmanager
|
|||
from io import BytesIO
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack_api import OpenAIFilePurpose
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -9,9 +9,9 @@ from unittest.mock import patch
|
|||
|
||||
import pytest
|
||||
import requests
|
||||
from llama_stack_api import OpenAIFilePurpose
|
||||
|
||||
from llama_stack.core.datatypes import User
|
||||
from llama_stack_api import OpenAIFilePurpose
|
||||
|
||||
purpose = OpenAIFilePurpose.ASSISTANTS
|
||||
|
||||
|
|
|
|||
|
|
@ -15,6 +15,9 @@ that enables routing based on provider_data alone.
|
|||
from unittest.mock import AsyncMock, patch
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||
from llama_stack.core.telemetry.telemetry import MetricEvent
|
||||
from llama_stack_api import (
|
||||
Api,
|
||||
OpenAIAssistantMessageParam,
|
||||
|
|
@ -23,9 +26,6 @@ from llama_stack_api import (
|
|||
OpenAIChoice,
|
||||
)
|
||||
|
||||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||
from llama_stack.core.telemetry.telemetry import MetricEvent
|
||||
|
||||
|
||||
class OpenAIChatCompletionWithMetrics(OpenAIChatCompletion):
|
||||
metrics: list[MetricEvent] | None = None
|
||||
|
|
|
|||
|
|
@ -9,6 +9,8 @@ import time
|
|||
import uuid
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack.log import get_logger
|
||||
from llama_stack_api import (
|
||||
DataConfig,
|
||||
DatasetFormat,
|
||||
|
|
@ -18,8 +20,6 @@ from llama_stack_api import (
|
|||
TrainingConfig,
|
||||
)
|
||||
|
||||
from llama_stack.log import get_logger
|
||||
|
||||
# Configure logging
|
||||
logger = get_logger(name=__name__, category="post_training")
|
||||
|
||||
|
|
|
|||
|
|
@ -12,9 +12,9 @@ import warnings
|
|||
from collections.abc import Generator
|
||||
|
||||
import pytest
|
||||
from llama_stack_api import ViolationLevel
|
||||
|
||||
from llama_stack.models.llama.sku_types import CoreModelId
|
||||
from llama_stack_api import ViolationLevel
|
||||
|
||||
# Llama Guard models available for text and vision shields
|
||||
LLAMA_GUARD_TEXT_MODELS = [CoreModelId.llama_guard_4_12b.value]
|
||||
|
|
|
|||
|
|
@ -7,6 +7,7 @@ import base64
|
|||
import mimetypes
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack_api import ViolationLevel
|
||||
|
||||
CODE_SCANNER_ENABLED_PROVIDERS = {"ollama", "together", "fireworks"}
|
||||
|
|
|
|||
|
|
@ -9,6 +9,7 @@ import mimetypes
|
|||
import os
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack_api import ViolationLevel
|
||||
|
||||
VISION_SHIELD_ENABLED_PROVIDERS = {"together"}
|
||||
|
|
|
|||
|
|
@ -7,9 +7,9 @@
|
|||
import re
|
||||
|
||||
import pytest
|
||||
from llama_stack_api import ToolGroupNotFoundError
|
||||
|
||||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||
from llama_stack_api import ToolGroupNotFoundError
|
||||
from tests.common.mcp import MCP_TOOLGROUP_ID, make_mcp_server
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -8,12 +8,12 @@ import time
|
|||
from io import BytesIO
|
||||
|
||||
import pytest
|
||||
from llama_stack_api import Chunk, ExpiresAfter
|
||||
from llama_stack_client import BadRequestError
|
||||
from openai import BadRequestError as OpenAIBadRequestError
|
||||
|
||||
from llama_stack.core.library_client import LlamaStackAsLibraryClient
|
||||
from llama_stack.log import get_logger
|
||||
from llama_stack_api import Chunk, ExpiresAfter
|
||||
|
||||
from ..conftest import vector_provider_wrapper
|
||||
|
||||
|
|
|
|||
|
|
@ -5,6 +5,7 @@
|
|||
# the root directory of this source tree.
|
||||
|
||||
import pytest
|
||||
|
||||
from llama_stack_api import Chunk
|
||||
|
||||
from ..conftest import vector_provider_wrapper
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue