mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-08-03 17:29:01 +00:00
clean up prints
This commit is contained in:
parent
57340e7a51
commit
07806f85c7
3 changed files with 0 additions and 7 deletions
|
@ -185,8 +185,6 @@ async def construct_stack(
|
||||||
impls = await resolve_impls(
|
impls = await resolve_impls(
|
||||||
run_config, provider_registry or get_provider_registry(), dist_registry
|
run_config, provider_registry or get_provider_registry(), dist_registry
|
||||||
)
|
)
|
||||||
print("impls", impls)
|
|
||||||
print("run_config", run_config)
|
|
||||||
await register_resources(run_config, impls)
|
await register_resources(run_config, impls)
|
||||||
return impls
|
return impls
|
||||||
|
|
||||||
|
|
|
@ -8,7 +8,6 @@ import tempfile
|
||||||
|
|
||||||
import pytest
|
import pytest
|
||||||
import pytest_asyncio
|
import pytest_asyncio
|
||||||
from rich.pretty import pprint
|
|
||||||
|
|
||||||
from llama_stack.apis.models import ModelInput, ModelType
|
from llama_stack.apis.models import ModelInput, ModelType
|
||||||
from llama_stack.distribution.datatypes import Api, Provider
|
from llama_stack.distribution.datatypes import Api, Provider
|
||||||
|
@ -99,7 +98,6 @@ async def agents_stack(request, inference_model, safety_shield):
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
|
||||||
pprint(f"fixture providers: {providers}")
|
|
||||||
test_stack = await construct_stack_for_test(
|
test_stack = await construct_stack_for_test(
|
||||||
[Api.agents, Api.inference, Api.safety, Api.memory],
|
[Api.agents, Api.inference, Api.safety, Api.memory],
|
||||||
providers,
|
providers,
|
||||||
|
|
|
@ -9,7 +9,6 @@ import tempfile
|
||||||
from typing import Any, Dict, List, Optional
|
from typing import Any, Dict, List, Optional
|
||||||
|
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
from termcolor import cprint
|
|
||||||
|
|
||||||
from llama_stack.apis.datasets import DatasetInput
|
from llama_stack.apis.datasets import DatasetInput
|
||||||
from llama_stack.apis.eval_tasks import EvalTaskInput
|
from llama_stack.apis.eval_tasks import EvalTaskInput
|
||||||
|
@ -58,9 +57,7 @@ async def construct_stack_for_test(
|
||||||
scoring_fns=scoring_fns or [],
|
scoring_fns=scoring_fns or [],
|
||||||
eval_tasks=eval_tasks or [],
|
eval_tasks=eval_tasks or [],
|
||||||
)
|
)
|
||||||
cprint(f"run_config BEFORE: {run_config}", "yellow")
|
|
||||||
run_config = parse_and_maybe_upgrade_config(run_config)
|
run_config = parse_and_maybe_upgrade_config(run_config)
|
||||||
cprint(f"run_config AFTER: {run_config}", "red")
|
|
||||||
try:
|
try:
|
||||||
remote_config = remote_provider_config(run_config)
|
remote_config = remote_provider_config(run_config)
|
||||||
if not remote_config:
|
if not remote_config:
|
||||||
|
|
Loading…
Add table
Add a link
Reference in a new issue