mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-03 17:29:01 +00:00
clean up prints
This commit is contained in:
parent
57340e7a51
commit
07806f85c7
3 changed files with 0 additions and 7 deletions
|
@ -185,8 +185,6 @@ async def construct_stack(
|
|||
impls = await resolve_impls(
|
||||
run_config, provider_registry or get_provider_registry(), dist_registry
|
||||
)
|
||||
print("impls", impls)
|
||||
print("run_config", run_config)
|
||||
await register_resources(run_config, impls)
|
||||
return impls
|
||||
|
||||
|
|
|
@ -8,7 +8,6 @@ import tempfile
|
|||
|
||||
import pytest
|
||||
import pytest_asyncio
|
||||
from rich.pretty import pprint
|
||||
|
||||
from llama_stack.apis.models import ModelInput, ModelType
|
||||
from llama_stack.distribution.datatypes import Api, Provider
|
||||
|
@ -99,7 +98,6 @@ async def agents_stack(request, inference_model, safety_shield):
|
|||
)
|
||||
)
|
||||
|
||||
pprint(f"fixture providers: {providers}")
|
||||
test_stack = await construct_stack_for_test(
|
||||
[Api.agents, Api.inference, Api.safety, Api.memory],
|
||||
providers,
|
||||
|
|
|
@ -9,7 +9,6 @@ import tempfile
|
|||
from typing import Any, Dict, List, Optional
|
||||
|
||||
from pydantic import BaseModel
|
||||
from termcolor import cprint
|
||||
|
||||
from llama_stack.apis.datasets import DatasetInput
|
||||
from llama_stack.apis.eval_tasks import EvalTaskInput
|
||||
|
@ -58,9 +57,7 @@ async def construct_stack_for_test(
|
|||
scoring_fns=scoring_fns or [],
|
||||
eval_tasks=eval_tasks or [],
|
||||
)
|
||||
cprint(f"run_config BEFORE: {run_config}", "yellow")
|
||||
run_config = parse_and_maybe_upgrade_config(run_config)
|
||||
cprint(f"run_config AFTER: {run_config}", "red")
|
||||
try:
|
||||
remote_config = remote_provider_config(run_config)
|
||||
if not remote_config:
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue