From 6540122844b5b35976664cbb02df4aff85ca627b Mon Sep 17 00:00:00 2001 From: Dillon Laird Date: Tue, 12 Mar 2024 14:07:29 -0700 Subject: [PATCH] fixed flake8 --- pyproject.toml | 3 --- tests/test_llm.py | 10 +++++----- tests/test_lmm.py | 10 +++++----- 3 files changed, 10 insertions(+), 13 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index 1e33b7ba..a823de15 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -52,9 +52,6 @@ log_cli_level = "INFO" log_cli_format = "%(asctime)s [%(levelname)s] %(message)s (%(filename)s:%(lineno)s)" log_cli_date_format = "%Y-%m-%d %H:%M:%S" -[tool.flake8] -exclude = "tests/*" - [tool.black] exclude = '.vscode|.eggs|venv' line-length = 88 # suggested by black official site diff --git a/tests/test_llm.py b/tests/test_llm.py index 18a96fc4..74453a4b 100644 --- a/tests/test_llm.py +++ b/tests/test_llm.py @@ -4,13 +4,13 @@ from vision_agent.tools import CLIP from vision_agent.tools.tools import GroundingDINO -from .fixtures import openai_llm_mock +from .fixtures import openai_llm_mock # noqa: F401 @pytest.mark.parametrize( "openai_llm_mock", ["mocked response"], indirect=["openai_llm_mock"] ) -def test_generate_with_mock(openai_llm_mock): +def test_generate_with_mock(openai_llm_mock): # noqa: F811 llm = OpenAILLM() response = llm.generate("test prompt") assert response == "mocked response" @@ -25,7 +25,7 @@ def test_generate_with_mock(openai_llm_mock): ['{"Parameters": {"prompt": "cat"}}'], indirect=["openai_llm_mock"], ) -def test_generate_classifier(openai_llm_mock): +def test_generate_classifier(openai_llm_mock): # noqa: F811 llm = OpenAILLM() prompt = "Can you generate a cat classifier?" classifier = llm.generate_classifier(prompt) @@ -38,7 +38,7 @@ def test_generate_classifier(openai_llm_mock): ['{"Parameters": {"prompt": "cat"}}'], indirect=["openai_llm_mock"], ) -def test_generate_detector(openai_llm_mock): +def test_generate_detector(openai_llm_mock): # noqa: F811 llm = OpenAILLM() prompt = "Can you generate a cat detector?" detector = llm.generate_detector(prompt) @@ -51,7 +51,7 @@ def test_generate_detector(openai_llm_mock): ['{"Parameters": {"prompt": "cat"}}'], indirect=["openai_llm_mock"], ) -def test_generate_segmentor(openai_llm_mock): +def test_generate_segmentor(openai_llm_mock): # noqa: F811 llm = OpenAILLM() prompt = "Can you generate a cat segmentor?" segmentor = llm.generate_detector(prompt) diff --git a/tests/test_lmm.py b/tests/test_lmm.py index af356de6..97cce581 100644 --- a/tests/test_lmm.py +++ b/tests/test_lmm.py @@ -6,7 +6,7 @@ from vision_agent.lmm.lmm import OpenAILMM from vision_agent.tools import CLIP, GroundingDINO, GroundingSAM -from .fixtures import openai_lmm_mock +from .fixtures import openai_lmm_mock # noqa: F401 def create_temp_image(image_format="jpeg"): @@ -20,7 +20,7 @@ def create_temp_image(image_format="jpeg"): @pytest.mark.parametrize( "openai_lmm_mock", ["mocked response"], indirect=["openai_lmm_mock"] ) -def test_generate_with_mock(openai_lmm_mock): +def test_generate_with_mock(openai_lmm_mock): # noqa: F811 temp_image = create_temp_image() lmm = OpenAILMM() response = lmm.generate("test prompt", image=temp_image) @@ -38,7 +38,7 @@ def test_generate_with_mock(openai_lmm_mock): ['{"Parameters": {"prompt": "cat"}}'], indirect=["openai_lmm_mock"], ) -def test_generate_classifier(openai_lmm_mock): +def test_generate_classifier(openai_lmm_mock): # noqa: F811 lmm = OpenAILMM() prompt = "Can you generate a cat classifier?" classifier = lmm.generate_classifier(prompt) @@ -51,7 +51,7 @@ def test_generate_classifier(openai_lmm_mock): ['{"Parameters": {"prompt": "cat"}}'], indirect=["openai_lmm_mock"], ) -def test_generate_classifier(openai_lmm_mock): +def test_generate_classifier(openai_lmm_mock): # noqa: F811 lmm = OpenAILMM() prompt = "Can you generate a cat classifier?" detector = lmm.generate_detector(prompt) @@ -64,7 +64,7 @@ def test_generate_classifier(openai_lmm_mock): ['{"Parameters": {"prompt": "cat"}}'], indirect=["openai_lmm_mock"], ) -def test_generate_classifier(openai_lmm_mock): +def test_generate_classifier(openai_lmm_mock): # noqa: F811 lmm = OpenAILMM() prompt = "Can you generate a cat classifier?" segmentor = lmm.generate_segmentor(prompt)