From d27a0f276cc999645086f95167ff6e8ec684315d Mon Sep 17 00:00:00 2001 From: Ashwin Bharambe Date: Sun, 4 May 2025 13:21:06 -0700 Subject: [PATCH] fix: pytest.mark.skip, not pytest.skip --- requirements.txt | 137 ------------------------ tests/integration/agents/test_agents.py | 2 +- 2 files changed, 1 insertion(+), 138 deletions(-) diff --git a/requirements.txt b/requirements.txt index ca200e07d..194af774b 100644 --- a/requirements.txt +++ b/requirements.txt @@ -1,206 +1,69 @@ # This file was autogenerated by uv via the following command: # uv export --frozen --no-hashes --no-emit-project --output-file=requirements.txt annotated-types==0.7.0 - # via pydantic anyio==4.8.0 - # via - # httpx - # llama-stack-client - # openai attrs==25.1.0 - # via - # jsonschema - # referencing blobfile==3.0.0 - # via llama-stack cachetools==5.5.2 - # via google-auth certifi==2025.1.31 - # via - # httpcore - # httpx - # kubernetes - # requests charset-normalizer==3.4.1 - # via requests click==8.1.8 - # via llama-stack-client colorama==0.4.6 ; sys_platform == 'win32' - # via - # click - # tqdm distro==1.9.0 - # via - # llama-stack-client - # openai durationpy==0.9 - # via kubernetes exceptiongroup==1.2.2 ; python_full_version < '3.11' - # via anyio filelock==3.17.0 - # via - # blobfile - # huggingface-hub fire==0.7.0 - # via llama-stack fsspec==2024.12.0 - # via huggingface-hub google-auth==2.38.0 - # via kubernetes h11==0.16.0 - # via - # httpcore - # llama-stack httpcore==1.0.9 - # via httpx httpx==0.28.1 - # via - # llama-stack - # llama-stack-client - # openai huggingface-hub==0.29.0 - # via llama-stack idna==3.10 - # via - # anyio - # httpx - # requests jinja2==3.1.6 - # via llama-stack jiter==0.8.2 - # via openai jsonschema==4.23.0 - # via llama-stack jsonschema-specifications==2024.10.1 - # via jsonschema kubernetes==32.0.1 - # via llama-stack llama-stack-client==0.2.5 - # via llama-stack lxml==5.3.1 - # via blobfile markdown-it-py==3.0.0 - # via rich markupsafe==3.0.2 - # via jinja2 mdurl==0.1.2 - # via markdown-it-py numpy==2.2.3 - # via pandas oauthlib==3.2.2 - # via - # kubernetes - # requests-oauthlib openai==1.71.0 - # via llama-stack packaging==24.2 - # via huggingface-hub pandas==2.2.3 - # via llama-stack-client pillow==11.1.0 - # via llama-stack prompt-toolkit==3.0.50 - # via - # llama-stack - # llama-stack-client pyaml==25.1.0 - # via llama-stack-client pyasn1==0.6.1 - # via - # pyasn1-modules - # rsa pyasn1-modules==0.4.2 - # via google-auth pycryptodomex==3.21.0 - # via blobfile pydantic==2.10.6 - # via - # llama-stack - # llama-stack-client - # openai pydantic-core==2.27.2 - # via pydantic pygments==2.19.1 - # via rich python-dateutil==2.9.0.post0 - # via - # kubernetes - # pandas python-dotenv==1.0.1 - # via llama-stack pytz==2025.1 - # via pandas pyyaml==6.0.2 - # via - # huggingface-hub - # kubernetes - # pyaml referencing==0.36.2 - # via - # jsonschema - # jsonschema-specifications regex==2024.11.6 - # via tiktoken requests==2.32.3 - # via - # huggingface-hub - # kubernetes - # llama-stack - # requests-oauthlib - # tiktoken requests-oauthlib==2.0.0 - # via kubernetes rich==13.9.4 - # via - # llama-stack - # llama-stack-client rpds-py==0.22.3 - # via - # jsonschema - # referencing rsa==4.9 - # via google-auth setuptools==75.8.0 - # via llama-stack six==1.17.0 - # via - # kubernetes - # python-dateutil sniffio==1.3.1 - # via - # anyio - # llama-stack-client - # openai termcolor==2.5.0 - # via - # fire - # llama-stack - # llama-stack-client tiktoken==0.9.0 - # via llama-stack tqdm==4.67.1 - # via - # huggingface-hub - # llama-stack-client - # openai typing-extensions==4.12.2 - # via - # anyio - # huggingface-hub - # llama-stack-client - # openai - # pydantic - # pydantic-core - # referencing - # rich tzdata==2025.1 - # via pandas urllib3==2.3.0 - # via - # blobfile - # kubernetes - # requests wcwidth==0.2.13 - # via prompt-toolkit websocket-client==1.8.0 - # via kubernetes diff --git a/tests/integration/agents/test_agents.py b/tests/integration/agents/test_agents.py index 546674708..367fe1b9a 100644 --- a/tests/integration/agents/test_agents.py +++ b/tests/integration/agents/test_agents.py @@ -552,7 +552,7 @@ def test_rag_agent_with_attachments(llama_stack_client_with_mocked_inference, ag assert "lora" in response.output_message.content.lower() -@pytest.skip(reason="Code interpreter is currently disabled in the Stack") +@pytest.mark.skip(reason="Code interpreter is currently disabled in the Stack") def test_rag_and_code_agent(llama_stack_client_with_mocked_inference, agent_config): if "llama-4" in agent_config["model"].lower(): pytest.xfail("Not working for llama4")