X-Git-Url: https://git.madduck.net/etc/vim.git/blobdiff_plain/a2408b3cb23fe252b3674cee484d742496bb3411..8b0c7bcfda60e2200ce1335b5b371593548ab9f8:/tests/test_primer.py diff --git a/tests/test_primer.py b/tests/test_primer.py index 46ccfc1..e7f99fd 100644 --- a/tests/test_primer.py +++ b/tests/test_primer.py @@ -67,10 +67,14 @@ def event_loop() -> Iterator[None]: loop.close() -async def raise_subprocess_error(*args: Any, **kwargs: Any) -> None: +async def raise_subprocess_error_1(*args: Any, **kwargs: Any) -> None: raise CalledProcessError(1, ["unittest", "error"], b"", b"") +async def raise_subprocess_error_123(*args: Any, **kwargs: Any) -> None: + raise CalledProcessError(123, ["unittest", "error"], b"", b"") + + async def return_false(*args: Any, **kwargs: Any) -> bool: return False @@ -102,13 +106,16 @@ class PrimerLibTests(unittest.TestCase): def test_black_run(self) -> None: """Pretend to run Black to ensure we cater for all scenarios""" loop = asyncio.get_event_loop() + project_name = "unittest" repo_path = Path(gettempdir()) project_config = deepcopy(FAKE_PROJECT_CONFIG) results = lib.Results({"failed": 0, "success": 0}, {}) # Test a successful Black run with patch("black_primer.lib._gen_check_output", return_subproccess_output): - loop.run_until_complete(lib.black_run(repo_path, project_config, results)) + loop.run_until_complete( + lib.black_run(project_name, repo_path, project_config, results) + ) self.assertEqual(1, results.stats["success"]) self.assertFalse(results.failed_projects) @@ -116,18 +123,34 @@ class PrimerLibTests(unittest.TestCase): project_config["expect_formatting_changes"] = True results = lib.Results({"failed": 0, "success": 0}, {}) with patch("black_primer.lib._gen_check_output", return_subproccess_output): - loop.run_until_complete(lib.black_run(repo_path, project_config, results)) + loop.run_until_complete( + lib.black_run(project_name, repo_path, project_config, results) + ) self.assertEqual(1, results.stats["failed"]) self.assertTrue(results.failed_projects) # Test a fail based on returning 1 and not expecting formatting changes project_config["expect_formatting_changes"] = False results = lib.Results({"failed": 0, "success": 0}, {}) - with patch("black_primer.lib._gen_check_output", raise_subprocess_error): - loop.run_until_complete(lib.black_run(repo_path, project_config, results)) + with patch("black_primer.lib._gen_check_output", raise_subprocess_error_1): + loop.run_until_complete( + lib.black_run(project_name, repo_path, project_config, results) + ) self.assertEqual(1, results.stats["failed"]) self.assertTrue(results.failed_projects) + # Test a formatting error based on returning 123 + with patch("black_primer.lib._gen_check_output", raise_subprocess_error_123): + loop.run_until_complete( + lib.black_run(project_name, repo_path, project_config, results) + ) + self.assertEqual(2, results.stats["failed"]) + + def test_flatten_cli_args(self) -> None: + fake_long_args = ["--arg", ["really/", "|long", "|regex", "|splitup"], "--done"] + expected = ["--arg", "really/|long|regex|splitup", "--done"] + self.assertEqual(expected, lib._flatten_cli_args(fake_long_args)) + @event_loop() def test_gen_check_output(self) -> None: loop = asyncio.get_event_loop() @@ -166,12 +189,14 @@ class PrimerLibTests(unittest.TestCase): @patch("sys.stdout", new_callable=StringIO) @event_loop() def test_process_queue(self, mock_stdout: Mock) -> None: + """Test the process queue on primer itself + - If you have non black conforming formatting in primer itself this can fail""" loop = asyncio.get_event_loop() config_path = Path(lib.__file__).parent / "primer.json" with patch("black_primer.lib.git_checkout_or_rebase", return_false): with TemporaryDirectory() as td: return_val = loop.run_until_complete( - lib.process_queue(str(config_path), td, 2) + lib.process_queue(str(config_path), Path(td), 2) ) self.assertEqual(0, return_val) @@ -189,9 +214,10 @@ class PrimerCLITests(unittest.TestCase): "rebase": False, "workdir": str(work_dir), "workers": 69, + "no_diff": False, } with patch("black_primer.cli.lib.process_queue", return_zero): - return_val = loop.run_until_complete(cli.async_main(**args)) + return_val = loop.run_until_complete(cli.async_main(**args)) # type: ignore self.assertEqual(0, return_val) def test_handle_debug(self) -> None: