diff --git a/llama_stack/distribution/utils/dynamic.py b/llama_stack/distribution/utils/dynamic.py index 9818f6a6e..e15ab63d6 100644 --- a/llama_stack/distribution/utils/dynamic.py +++ b/llama_stack/distribution/utils/dynamic.py @@ -48,7 +48,6 @@ async def instantiate_provider( routing_table = provider_config inner_specs = {x.provider_id: x for x in provider_spec.inner_specs} - cprint(f"inner_specs: {inner_specs}", "cyan") inner_impls = [] for routing_entry in routing_table: impl = await instantiate_provider( diff --git a/tests/examples/local-run.yaml b/tests/examples/local-run.yaml new file mode 100644 index 000000000..2ae975cdc --- /dev/null +++ b/tests/examples/local-run.yaml @@ -0,0 +1,87 @@ +built_at: '2024-09-23T00:54:40.551416' +image_name: test-2 +docker_image: null +conda_env: test-2 +apis_to_serve: +- shields +- agents +- models +- memory +- memory_banks +- inference +- safety +api_providers: + inference: + providers: + - meta-reference + safety: + providers: + - meta-reference + agents: + provider_id: meta-reference + config: + persistence_store: + namespace: null + type: sqlite + db_path: /home/xiyan/.llama/runtime/kvstore.db + memory: + providers: + - meta-reference + telemetry: + provider_id: meta-reference + config: {} +routing_table: + inference: + - provider_id: meta-reference + config: + model: Meta-Llama3.1-8B-Instruct + quantization: null + torch_seed: null + max_seq_len: 4096 + max_batch_size: 1 + routing_key: Meta-Llama3.1-8B-Instruct + safety: + - provider_id: meta-reference + config: + llama_guard_shield: + model: Llama-Guard-3-8B + excluded_categories: [] + disable_input_check: false + disable_output_check: false + prompt_guard_shield: + model: Prompt-Guard-86M + routing_key: llama_guard + - provider_id: meta-reference + config: + llama_guard_shield: + model: Llama-Guard-3-8B + excluded_categories: [] + disable_input_check: false + disable_output_check: false + prompt_guard_shield: + model: Prompt-Guard-86M + routing_key: code_scanner_guard + - provider_id: meta-reference + config: + llama_guard_shield: + model: Llama-Guard-3-8B + excluded_categories: [] + disable_input_check: false + disable_output_check: false + prompt_guard_shield: + model: Prompt-Guard-86M + routing_key: injection_shield + - provider_id: meta-reference + config: + llama_guard_shield: + model: Llama-Guard-3-8B + excluded_categories: [] + disable_input_check: false + disable_output_check: false + prompt_guard_shield: + model: Prompt-Guard-86M + routing_key: jailbreak_shield + memory: + - provider_id: meta-reference + config: {} + routing_key: vector diff --git a/tests/examples/router-local-run.yaml b/tests/examples/router-local-run.yaml deleted file mode 100644 index 589566cd2..000000000 --- a/tests/examples/router-local-run.yaml +++ /dev/null @@ -1,78 +0,0 @@ -built_at: '2024-09-18T13:41:17.656743' -image_name: local -docker_image: null -conda_env: local -apis_to_serve: -- inference -- memory -- telemetry -- agents -- safety -- models -api_providers: - inference: - providers: - - meta-reference - - remote::ollama - memory: - providers: - - meta-reference - - remote::pgvector - safety: - providers: - - meta-reference - telemetry: - provider_id: meta-reference - config: {} - agents: - provider_id: meta-reference - config: - persistence_store: - namespace: null - type: sqlite - db_path: /home/xiyan/.llama/runtime/kvstore.db -routing_table: - inference: - - routing_key: Meta-Llama3.1-8B-Instruct - provider_id: meta-reference - config: - model: Meta-Llama3.1-8B-Instruct - quantization: null - torch_seed: null - max_seq_len: 4096 - max_batch_size: 1 - memory: - - routing_key: vector - provider_id: meta-reference - config: {} - safety: - - routing_key: llama_guard - provider_id: meta-reference - config: - llama_guard_shield: - model: Llama-Guard-3-8B - excluded_categories: [] - disable_input_check: false - disable_output_check: false - prompt_guard_shield: - model: Prompt-Guard-86M - - routing_key: prompt_guard - provider_id: meta-reference - config: - llama_guard_shield: - model: Llama-Guard-3-8B - excluded_categories: [] - disable_input_check: false - disable_output_check: false - prompt_guard_shield: - model: Prompt-Guard-86M - - routing_key: injection_shield - provider_id: meta-reference - config: - llama_guard_shield: - model: Llama-Guard-3-8B - excluded_categories: [] - disable_input_check: false - disable_output_check: false - prompt_guard_shield: - model: Prompt-Guard-86M diff --git a/tests/examples/simple-local-run.yaml b/tests/examples/simple-local-run.yaml deleted file mode 100644 index 28a4f3825..000000000 --- a/tests/examples/simple-local-run.yaml +++ /dev/null @@ -1,40 +0,0 @@ -built_at: '2024-09-19T22:50:36.239761' -image_name: simple-local -docker_image: null -conda_env: simple-local -apis_to_serve: -- inference -- safety -- agents -- memory -- models -- telemetry -api_providers: - inference: - provider_id: meta-reference - config: - model: Meta-Llama3.1-8B-Instruct - quantization: null - torch_seed: null - max_seq_len: 4096 - max_batch_size: 1 - safety: - provider_id: meta-reference - config: - llama_guard_shield: - model: Llama-Guard-3-8B - excluded_categories: [] - disable_input_check: false - disable_output_check: false - prompt_guard_shield: - model: Prompt-Guard-86M - agents: - provider_id: meta-reference - config: {} - memory: - provider_id: meta-reference - config: {} - telemetry: - provider_id: meta-reference - config: {} -provider_routing_table: {}