mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-10-04 04:04:14 +00:00
use llama_stack.log
This commit is contained in:
parent
371de50875
commit
0098d24762
1 changed files with 3 additions and 2 deletions
|
@ -156,7 +156,6 @@ def make_mcp_server(required_auth_token: str | None = None, tools: dict[str, Cal
|
||||||
:param required_auth_token: Optional auth token required for access
|
:param required_auth_token: Optional auth token required for access
|
||||||
:param tools: Dictionary of tool_name -> tool_function. If None, uses default tools.
|
:param tools: Dictionary of tool_name -> tool_function. If None, uses default tools.
|
||||||
"""
|
"""
|
||||||
import logging
|
|
||||||
import threading
|
import threading
|
||||||
import time
|
import time
|
||||||
|
|
||||||
|
@ -168,6 +167,8 @@ def make_mcp_server(required_auth_token: str | None = None, tools: dict[str, Cal
|
||||||
from starlette.responses import Response
|
from starlette.responses import Response
|
||||||
from starlette.routing import Mount, Route
|
from starlette.routing import Mount, Route
|
||||||
|
|
||||||
|
from llama_stack.log import get_logger
|
||||||
|
|
||||||
server = FastMCP("FastMCP Test Server", log_level="WARNING")
|
server = FastMCP("FastMCP Test Server", log_level="WARNING")
|
||||||
|
|
||||||
tools = tools or default_tools()
|
tools = tools or default_tools()
|
||||||
|
@ -212,7 +213,7 @@ def make_mcp_server(required_auth_token: str | None = None, tools: dict[str, Cal
|
||||||
return sock.getsockname()[1]
|
return sock.getsockname()[1]
|
||||||
|
|
||||||
port = get_open_port()
|
port = get_open_port()
|
||||||
logger = logging.getLogger(__name__)
|
logger = get_logger(__name__, category="tests::mcp")
|
||||||
|
|
||||||
# make uvicorn logs be less verbose
|
# make uvicorn logs be less verbose
|
||||||
config = uvicorn.Config(app, host="0.0.0.0", port=port, log_level="warning")
|
config = uvicorn.Config(app, host="0.0.0.0", port=port, log_level="warning")
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue