mirror of
https://github.com/meta-llama/llama-stack.git
synced 2025-07-03 12:52:41 +00:00
chore(lint): update Ruff ignores for project conventions and maintainability (#1184)
- Added new ignores from flake8-bugbear (`B007`, `B008`) - Ignored `C901` (high function complexity) for now, pending review - Maintained PyTorch conventions (`N812`, `N817`) - Allowed `E731` (lambda assignments) for flexibility - Consolidated existing ignores (`E402`, `E501`, `F405`, `C408`, `N812`) - Documented rationale for each ignored rule This keeps our linting aligned with project needs while tracking potential fixes. Signed-off-by: Sébastien Han <seb@redhat.com> Signed-off-by: Sébastien Han <seb@redhat.com>
This commit is contained in:
parent
3b57d8ee88
commit
6fa257b475
33 changed files with 113 additions and 145 deletions
|
@ -42,7 +42,7 @@ def code_scanner_shield_id(available_shields):
|
|||
|
||||
@pytest.fixture(scope="session")
|
||||
def model_providers(llama_stack_client):
|
||||
return set([x.provider_id for x in llama_stack_client.providers.list() if x.api == "inference"])
|
||||
return {x.provider_id for x in llama_stack_client.providers.list() if x.api == "inference"}
|
||||
|
||||
|
||||
def test_unsafe_examples(llama_stack_client, llama_guard_text_shield_id):
|
||||
|
|
Loading…
Add table
Add a link
Reference in a new issue