All patches and comments are welcome. Please squash your changes to logical
commits before using git-format-patch and git-send-email to
patches@git.madduck.net.
If you'd read over the Git project's submission guidelines and adhered to them,
I'd be especially grateful.
2 from json.decoder import JSONDecodeError
4 from concurrent.futures import Executor, ThreadPoolExecutor, ProcessPoolExecutor
5 from contextlib import contextmanager
6 from datetime import datetime
9 from multiprocessing import Manager, freeze_support
11 from pathlib import Path
12 from pathspec.patterns.gitwildmatch import GitWildMatchPatternError
33 from dataclasses import replace
36 from black.const import DEFAULT_LINE_LENGTH, DEFAULT_INCLUDES, DEFAULT_EXCLUDES
37 from black.const import STDIN_PLACEHOLDER
38 from black.nodes import STARS, syms, is_simple_decorator_expression
39 from black.lines import Line, EmptyLineTracker
40 from black.linegen import transform_line, LineGenerator, LN
41 from black.comments import normalize_fmt_off
42 from black.mode import Mode, TargetVersion
43 from black.mode import Feature, supports_feature, VERSION_TO_FEATURES
44 from black.cache import read_cache, write_cache, get_cache_info, filter_cached, Cache
45 from black.concurrency import cancel, shutdown, maybe_install_uvloop
46 from black.output import dump_to_file, ipynb_diff, diff, color_diff, out, err
47 from black.report import Report, Changed, NothingChanged
48 from black.files import find_project_root, find_pyproject_toml, parse_pyproject_toml
49 from black.files import gen_python_files, get_gitignore, normalize_path_maybe_ignore
50 from black.files import wrap_stream_for_windows
51 from black.parsing import InvalidInput # noqa F401
52 from black.parsing import lib2to3_parse, parse_ast, stringify_ast
53 from black.handle_ipynb_magics import (
56 remove_trailing_semicolon,
57 put_trailing_semicolon_back,
59 jupyter_dependencies_are_installed,
64 from blib2to3.pytree import Node, Leaf
65 from blib2to3.pgen2 import token
67 from _black_version import version as __version__
75 class WriteBack(Enum):
83 def from_configuration(
84 cls, *, check: bool, diff: bool, color: bool = False
86 if check and not diff:
92 return cls.DIFF if diff else cls.YES
95 # Legacy name, left for integrations.
98 DEFAULT_WORKERS = os.cpu_count()
101 def read_pyproject_toml(
102 ctx: click.Context, param: click.Parameter, value: Optional[str]
104 """Inject Black configuration from "pyproject.toml" into defaults in `ctx`.
106 Returns the path to a successfully found and read configuration file, None
110 value = find_pyproject_toml(ctx.params.get("src", ()))
115 config = parse_pyproject_toml(value)
116 except (OSError, ValueError) as e:
117 raise click.FileError(
118 filename=value, hint=f"Error reading configuration file: {e}"
124 # Sanitize the values to be Click friendly. For more information please see:
125 # https://github.com/psf/black/issues/1458
126 # https://github.com/pallets/click/issues/1567
128 k: str(v) if not isinstance(v, (list, dict)) else v
129 for k, v in config.items()
132 target_version = config.get("target_version")
133 if target_version is not None and not isinstance(target_version, list):
134 raise click.BadOptionUsage(
135 "target-version", "Config key target-version must be a list"
138 default_map: Dict[str, Any] = {}
140 default_map.update(ctx.default_map)
141 default_map.update(config)
143 ctx.default_map = default_map
147 def target_version_option_callback(
148 c: click.Context, p: Union[click.Option, click.Parameter], v: Tuple[str, ...]
149 ) -> List[TargetVersion]:
150 """Compute the target versions from a --target-version flag.
152 This is its own function because mypy couldn't infer the type correctly
153 when it was a lambda, causing mypyc trouble.
155 return [TargetVersion[val.upper()] for val in v]
158 def re_compile_maybe_verbose(regex: str) -> Pattern[str]:
159 """Compile a regular expression string in `regex`.
161 If it contains newlines, use verbose mode.
164 regex = "(?x)" + regex
165 compiled: Pattern[str] = re.compile(regex)
171 param: click.Parameter,
172 value: Optional[str],
173 ) -> Optional[Pattern[str]]:
175 return re_compile_maybe_verbose(value) if value is not None else None
177 raise click.BadParameter("Not a valid regular expression") from None
180 @click.command(context_settings=dict(help_option_names=["-h", "--help"]))
181 @click.option("-c", "--code", type=str, help="Format the code passed in as a string.")
186 default=DEFAULT_LINE_LENGTH,
187 help="How many characters per line to allow.",
193 type=click.Choice([v.name.lower() for v in TargetVersion]),
194 callback=target_version_option_callback,
197 "Python versions that should be supported by Black's output. [default: per-file"
205 "Format all input files like typing stubs regardless of file extension (useful"
206 " when piping source on standard input)."
213 "Format all input files like Jupyter Notebooks regardless of file extension "
214 "(useful when piping source on standard input)."
219 "--skip-string-normalization",
221 help="Don't normalize string quotes or prefixes.",
225 "--skip-magic-trailing-comma",
227 help="Don't use trailing commas as a reason to split lines.",
230 "--experimental-string-processing",
234 "Experimental option that performs more normalization on string literals."
235 " Currently disabled because it leads to some crashes."
242 "Don't write the files back, just return the status. Return code 0 means"
243 " nothing would change. Return code 1 means some files would be reformatted."
244 " Return code 123 means there was an internal error."
250 help="Don't write the files back, just output a diff for each file on stdout.",
253 "--color/--no-color",
255 help="Show colored diff. Only applies when `--diff` is given.",
260 help="If --fast given, skip temporary sanity checks. [default: --safe]",
263 "--required-version",
266 "Require a specific version of Black to be running (useful for unifying results"
267 " across many environments e.g. with a pyproject.toml file)."
273 default=DEFAULT_INCLUDES,
274 callback=validate_regex,
276 "A regular expression that matches files and directories that should be"
277 " included on recursive searches. An empty value means all files are included"
278 " regardless of the name. Use forward slashes for directories on all platforms"
279 " (Windows, too). Exclusions are calculated first, inclusions later."
286 callback=validate_regex,
288 "A regular expression that matches files and directories that should be"
289 " excluded on recursive searches. An empty value means no paths are excluded."
290 " Use forward slashes for directories on all platforms (Windows, too)."
291 " Exclusions are calculated first, inclusions later. [default:"
292 f" {DEFAULT_EXCLUDES}]"
299 callback=validate_regex,
301 "Like --exclude, but adds additional files and directories on top of the"
302 " excluded ones. (Useful if you simply want to add to the default)"
308 callback=validate_regex,
310 "Like --exclude, but files and directories matching this regex will be "
311 "excluded even when they are passed explicitly as arguments."
318 "The name of the file when passing it through stdin. Useful to make "
319 "sure Black will respect --force-exclude option on some "
320 "editors that rely on using stdin."
326 type=click.IntRange(min=1),
327 default=DEFAULT_WORKERS,
329 help="Number of parallel workers",
336 "Don't emit non-error messages to stderr. Errors are still emitted; silence"
337 " those with 2>/dev/null."
345 "Also emit messages to stderr about files that were not changed or were ignored"
346 " due to exclusion patterns."
349 @click.version_option(version=__version__)
354 exists=True, file_okay=True, dir_okay=True, readable=True, allow_dash=True
370 callback=read_pyproject_toml,
371 help="Read configuration from FILE path.",
378 target_version: List[TargetVersion],
385 skip_string_normalization: bool,
386 skip_magic_trailing_comma: bool,
387 experimental_string_processing: bool,
390 required_version: str,
391 include: Pattern[str],
392 exclude: Optional[Pattern[str]],
393 extend_exclude: Optional[Pattern[str]],
394 force_exclude: Optional[Pattern[str]],
395 stdin_filename: Optional[str],
397 src: Tuple[str, ...],
398 config: Optional[str],
400 """The uncompromising code formatter."""
401 if config and verbose:
402 out(f"Using configuration from {config}.", bold=False, fg="blue")
404 error_msg = "Oh no! 💥 💔 💥"
405 if required_version and required_version != __version__:
407 f"{error_msg} The required version `{required_version}` does not match"
408 f" the running version `{__version__}`!"
412 err("Cannot pass both `pyi` and `ipynb` flags!")
415 write_back = WriteBack.from_configuration(check=check, diff=diff, color=color)
417 versions = set(target_version)
419 # We'll autodetect later.
422 target_versions=versions,
423 line_length=line_length,
426 string_normalization=not skip_string_normalization,
427 magic_trailing_comma=not skip_magic_trailing_comma,
428 experimental_string_processing=experimental_string_processing,
432 # Run in quiet mode by default with -c; the extra output isn't useful.
433 # You can still pass -v to get verbose output.
436 report = Report(check=check, diff=diff, quiet=quiet, verbose=verbose)
440 content=code, fast=fast, write_back=write_back, mode=mode, report=report
444 sources = get_sources(
451 extend_exclude=extend_exclude,
452 force_exclude=force_exclude,
454 stdin_filename=stdin_filename,
456 except GitWildMatchPatternError:
461 "No Python files are present to be formatted. Nothing to do 😴",
467 if len(sources) == 1:
471 write_back=write_back,
479 write_back=write_back,
485 if verbose or not quiet:
486 out(error_msg if report.return_code else "All done! ✨ 🍰 ✨")
488 click.echo(str(report), err=True)
489 ctx.exit(report.return_code)
495 src: Tuple[str, ...],
498 include: Pattern[str],
499 exclude: Optional[Pattern[str]],
500 extend_exclude: Optional[Pattern[str]],
501 force_exclude: Optional[Pattern[str]],
503 stdin_filename: Optional[str],
505 """Compute the set of files to be formatted."""
507 root = find_project_root(src)
508 sources: Set[Path] = set()
509 path_empty(src, "No Path provided. Nothing to do 😴", quiet, verbose, ctx)
512 exclude = re_compile_maybe_verbose(DEFAULT_EXCLUDES)
513 gitignore = get_gitignore(root)
518 if s == "-" and stdin_filename:
519 p = Path(stdin_filename)
525 if is_stdin or p.is_file():
526 normalized_path = normalize_path_maybe_ignore(p, root, report)
527 if normalized_path is None:
530 normalized_path = "/" + normalized_path
531 # Hard-exclude any files that matches the `--force-exclude` regex.
533 force_exclude_match = force_exclude.search(normalized_path)
535 force_exclude_match = None
536 if force_exclude_match and force_exclude_match.group(0):
537 report.path_ignored(p, "matches the --force-exclude regular expression")
541 p = Path(f"{STDIN_PLACEHOLDER}{str(p)}")
543 if p.suffix == ".ipynb" and not jupyter_dependencies_are_installed(
544 verbose=verbose, quiet=quiet
567 err(f"invalid path: {s}")
572 src: Sized, msg: str, quiet: bool, verbose: bool, ctx: click.Context
575 Exit if there is no `src` provided for formatting
578 if verbose or not quiet:
584 content: str, fast: bool, write_back: WriteBack, mode: Mode, report: Report
587 Reformat and print out `content` without spawning child processes.
588 Similar to `reformat_one`, but for string content.
590 `fast`, `write_back`, and `mode` options are passed to
591 :func:`format_file_in_place` or :func:`format_stdin_to_stdout`.
593 path = Path("<string>")
596 if format_stdin_to_stdout(
597 content=content, fast=fast, write_back=write_back, mode=mode
599 changed = Changed.YES
600 report.done(path, changed)
601 except Exception as exc:
603 traceback.print_exc()
604 report.failed(path, str(exc))
608 src: Path, fast: bool, write_back: WriteBack, mode: Mode, report: "Report"
610 """Reformat a single file under `src` without spawning child processes.
612 `fast`, `write_back`, and `mode` options are passed to
613 :func:`format_file_in_place` or :func:`format_stdin_to_stdout`.
620 elif str(src).startswith(STDIN_PLACEHOLDER):
622 # Use the original name again in case we want to print something
624 src = Path(str(src)[len(STDIN_PLACEHOLDER) :])
629 if src.suffix == ".pyi":
630 mode = replace(mode, is_pyi=True)
631 elif src.suffix == ".ipynb":
632 mode = replace(mode, is_ipynb=True)
633 if format_stdin_to_stdout(fast=fast, write_back=write_back, mode=mode):
634 changed = Changed.YES
637 if write_back not in (WriteBack.DIFF, WriteBack.COLOR_DIFF):
638 cache = read_cache(mode)
639 res_src = src.resolve()
640 res_src_s = str(res_src)
641 if res_src_s in cache and cache[res_src_s] == get_cache_info(res_src):
642 changed = Changed.CACHED
643 if changed is not Changed.CACHED and format_file_in_place(
644 src, fast=fast, write_back=write_back, mode=mode
646 changed = Changed.YES
647 if (write_back is WriteBack.YES and changed is not Changed.CACHED) or (
648 write_back is WriteBack.CHECK and changed is Changed.NO
650 write_cache(cache, [src], mode)
651 report.done(src, changed)
652 except Exception as exc:
654 traceback.print_exc()
655 report.failed(src, str(exc))
661 write_back: WriteBack,
664 workers: Optional[int],
666 """Reformat multiple files using a ProcessPoolExecutor."""
668 loop = asyncio.get_event_loop()
669 worker_count = workers if workers is not None else DEFAULT_WORKERS
670 if sys.platform == "win32":
671 # Work around https://bugs.python.org/issue26903
672 worker_count = min(worker_count, 60)
674 executor = ProcessPoolExecutor(max_workers=worker_count)
675 except (ImportError, OSError):
676 # we arrive here if the underlying system does not support multi-processing
677 # like in AWS Lambda or Termux, in which case we gracefully fallback to
678 # a ThreadPoolExecutor with just a single worker (more workers would not do us
679 # any good due to the Global Interpreter Lock)
680 executor = ThreadPoolExecutor(max_workers=1)
683 loop.run_until_complete(
687 write_back=write_back,
696 if executor is not None:
700 async def schedule_formatting(
703 write_back: WriteBack,
706 loop: asyncio.AbstractEventLoop,
709 """Run formatting of `sources` in parallel using the provided `executor`.
711 (Use ProcessPoolExecutors for actual parallelism.)
713 `write_back`, `fast`, and `mode` options are passed to
714 :func:`format_file_in_place`.
717 if write_back not in (WriteBack.DIFF, WriteBack.COLOR_DIFF):
718 cache = read_cache(mode)
719 sources, cached = filter_cached(cache, sources)
720 for src in sorted(cached):
721 report.done(src, Changed.CACHED)
726 sources_to_cache = []
728 if write_back in (WriteBack.DIFF, WriteBack.COLOR_DIFF):
729 # For diff output, we need locks to ensure we don't interleave output
730 # from different processes.
732 lock = manager.Lock()
734 asyncio.ensure_future(
735 loop.run_in_executor(
736 executor, format_file_in_place, src, fast, mode, write_back, lock
739 for src in sorted(sources)
741 pending = tasks.keys()
743 loop.add_signal_handler(signal.SIGINT, cancel, pending)
744 loop.add_signal_handler(signal.SIGTERM, cancel, pending)
745 except NotImplementedError:
746 # There are no good alternatives for these on Windows.
749 done, _ = await asyncio.wait(pending, return_when=asyncio.FIRST_COMPLETED)
751 src = tasks.pop(task)
753 cancelled.append(task)
754 elif task.exception():
755 report.failed(src, str(task.exception()))
757 changed = Changed.YES if task.result() else Changed.NO
758 # If the file was written back or was successfully checked as
759 # well-formatted, store this information in the cache.
760 if write_back is WriteBack.YES or (
761 write_back is WriteBack.CHECK and changed is Changed.NO
763 sources_to_cache.append(src)
764 report.done(src, changed)
766 if sys.version_info >= (3, 7):
767 await asyncio.gather(*cancelled, return_exceptions=True)
769 await asyncio.gather(*cancelled, loop=loop, return_exceptions=True)
771 write_cache(cache, sources_to_cache, mode)
774 def format_file_in_place(
778 write_back: WriteBack = WriteBack.NO,
779 lock: Any = None, # multiprocessing.Manager().Lock() is some crazy proxy
781 """Format file under `src` path. Return True if changed.
783 If `write_back` is DIFF, write a diff to stdout. If it is YES, write reformatted
785 `mode` and `fast` options are passed to :func:`format_file_contents`.
787 if src.suffix == ".pyi":
788 mode = replace(mode, is_pyi=True)
789 elif src.suffix == ".ipynb":
790 mode = replace(mode, is_ipynb=True)
792 then = datetime.utcfromtimestamp(src.stat().st_mtime)
793 with open(src, "rb") as buf:
794 src_contents, encoding, newline = decode_bytes(buf.read())
796 dst_contents = format_file_contents(src_contents, fast=fast, mode=mode)
797 except NothingChanged:
799 except JSONDecodeError:
801 f"File '{src}' cannot be parsed as valid Jupyter notebook."
804 if write_back == WriteBack.YES:
805 with open(src, "w", encoding=encoding, newline=newline) as f:
806 f.write(dst_contents)
807 elif write_back in (WriteBack.DIFF, WriteBack.COLOR_DIFF):
808 now = datetime.utcnow()
809 src_name = f"{src}\t{then} +0000"
810 dst_name = f"{src}\t{now} +0000"
812 diff_contents = ipynb_diff(src_contents, dst_contents, src_name, dst_name)
814 diff_contents = diff(src_contents, dst_contents, src_name, dst_name)
816 if write_back == WriteBack.COLOR_DIFF:
817 diff_contents = color_diff(diff_contents)
819 with lock or nullcontext():
820 f = io.TextIOWrapper(
826 f = wrap_stream_for_windows(f)
827 f.write(diff_contents)
833 def format_stdin_to_stdout(
836 content: Optional[str] = None,
837 write_back: WriteBack = WriteBack.NO,
840 """Format file on stdin. Return True if changed.
842 If content is None, it's read from sys.stdin.
844 If `write_back` is YES, write reformatted code back to stdout. If it is DIFF,
845 write a diff to stdout. The `mode` argument is passed to
846 :func:`format_file_contents`.
848 then = datetime.utcnow()
851 src, encoding, newline = decode_bytes(sys.stdin.buffer.read())
853 src, encoding, newline = content, "utf-8", ""
857 dst = format_file_contents(src, fast=fast, mode=mode)
860 except NothingChanged:
864 f = io.TextIOWrapper(
865 sys.stdout.buffer, encoding=encoding, newline=newline, write_through=True
867 if write_back == WriteBack.YES:
868 # Make sure there's a newline after the content
869 if dst and dst[-1] != "\n":
872 elif write_back in (WriteBack.DIFF, WriteBack.COLOR_DIFF):
873 now = datetime.utcnow()
874 src_name = f"STDIN\t{then} +0000"
875 dst_name = f"STDOUT\t{now} +0000"
876 d = diff(src, dst, src_name, dst_name)
877 if write_back == WriteBack.COLOR_DIFF:
879 f = wrap_stream_for_windows(f)
884 def check_stability_and_equivalence(
885 src_contents: str, dst_contents: str, *, mode: Mode
887 """Perform stability and equivalence checks.
889 Raise AssertionError if source and destination contents are not
890 equivalent, or if a second pass of the formatter would format the
893 assert_equivalent(src_contents, dst_contents)
895 # Forced second pass to work around optional trailing commas (becoming
896 # forced trailing commas on pass 2) interacting differently with optional
897 # parentheses. Admittedly ugly.
898 dst_contents_pass2 = format_str(dst_contents, mode=mode)
899 if dst_contents != dst_contents_pass2:
900 dst_contents = dst_contents_pass2
901 assert_equivalent(src_contents, dst_contents, pass_num=2)
902 assert_stable(src_contents, dst_contents, mode=mode)
903 # Note: no need to explicitly call `assert_stable` if `dst_contents` was
904 # the same as `dst_contents_pass2`.
907 def format_file_contents(src_contents: str, *, fast: bool, mode: Mode) -> FileContent:
908 """Reformat contents of a file and return new contents.
910 If `fast` is False, additionally confirm that the reformatted code is
911 valid by calling :func:`assert_equivalent` and :func:`assert_stable` on it.
912 `mode` is passed to :func:`format_str`.
914 if not src_contents.strip():
918 dst_contents = format_ipynb_string(src_contents, fast=fast, mode=mode)
920 dst_contents = format_str(src_contents, mode=mode)
921 if src_contents == dst_contents:
924 if not fast and not mode.is_ipynb:
925 # Jupyter notebooks will already have been checked above.
926 check_stability_and_equivalence(src_contents, dst_contents, mode=mode)
930 def validate_cell(src: str) -> None:
931 """Check that cell does not already contain TransformerManager transformations.
933 If a cell contains ``!ls``, then it'll be transformed to
934 ``get_ipython().system('ls')``. However, if the cell originally contained
935 ``get_ipython().system('ls')``, then it would get transformed in the same way:
937 >>> TransformerManager().transform_cell("get_ipython().system('ls')")
938 "get_ipython().system('ls')\n"
939 >>> TransformerManager().transform_cell("!ls")
940 "get_ipython().system('ls')\n"
942 Due to the impossibility of safely roundtripping in such situations, cells
943 containing transformed magics will be ignored.
945 if any(transformed_magic in src for transformed_magic in TRANSFORMED_MAGICS):
949 def format_cell(src: str, *, fast: bool, mode: Mode) -> str:
950 """Format code in given cell of Jupyter notebook.
954 - if cell has trailing semicolon, remove it;
955 - if cell has IPython magics, mask them;
957 - reinstate IPython magics;
958 - reinstate trailing semicolon (if originally present);
959 - strip trailing newlines.
961 Cells with syntax errors will not be processed, as they
962 could potentially be automagics or multi-line magics, which
963 are currently not supported.
966 src_without_trailing_semicolon, has_trailing_semicolon = remove_trailing_semicolon(
970 masked_src, replacements = mask_cell(src_without_trailing_semicolon)
972 raise NothingChanged from None
973 masked_dst = format_str(masked_src, mode=mode)
975 check_stability_and_equivalence(masked_src, masked_dst, mode=mode)
976 dst_without_trailing_semicolon = unmask_cell(masked_dst, replacements)
977 dst = put_trailing_semicolon_back(
978 dst_without_trailing_semicolon, has_trailing_semicolon
980 dst = dst.rstrip("\n")
982 raise NothingChanged from None
986 def validate_metadata(nb: MutableMapping[str, Any]) -> None:
987 """If notebook is marked as non-Python, don't format it.
989 All notebook metadata fields are optional, see
990 https://nbformat.readthedocs.io/en/latest/format_description.html. So
991 if a notebook has empty metadata, we will try to parse it anyway.
993 language = nb.get("metadata", {}).get("language_info", {}).get("name", None)
994 if language is not None and language != "python":
995 raise NothingChanged from None
998 def format_ipynb_string(src_contents: str, *, fast: bool, mode: Mode) -> FileContent:
999 """Format Jupyter notebook.
1001 Operate cell-by-cell, only on code cells, only for Python notebooks.
1002 If the ``.ipynb`` originally had a trailing newline, it'll be preserved.
1004 trailing_newline = src_contents[-1] == "\n"
1006 nb = json.loads(src_contents)
1007 validate_metadata(nb)
1008 for cell in nb["cells"]:
1009 if cell.get("cell_type", None) == "code":
1011 src = "".join(cell["source"])
1012 dst = format_cell(src, fast=fast, mode=mode)
1013 except NothingChanged:
1016 cell["source"] = dst.splitlines(keepends=True)
1019 dst_contents = json.dumps(nb, indent=1, ensure_ascii=False)
1020 if trailing_newline:
1021 dst_contents = dst_contents + "\n"
1024 raise NothingChanged
1027 def format_str(src_contents: str, *, mode: Mode) -> FileContent:
1028 """Reformat a string and return new contents.
1030 `mode` determines formatting options, such as how many characters per line are
1034 >>> print(black.format_str("def f(arg:str='')->None:...", mode=black.Mode()))
1035 def f(arg: str = "") -> None:
1038 A more complex example:
1041 ... black.format_str(
1042 ... "def f(arg:str='')->None: hey",
1043 ... mode=black.Mode(
1044 ... target_versions={black.TargetVersion.PY36},
1046 ... string_normalization=False,
1057 src_node = lib2to3_parse(src_contents.lstrip(), mode.target_versions)
1059 future_imports = get_future_imports(src_node)
1060 if mode.target_versions:
1061 versions = mode.target_versions
1063 versions = detect_target_versions(src_node)
1065 # TODO: fully drop support and this code hopefully in January 2022 :D
1066 if TargetVersion.PY27 in mode.target_versions or versions == {TargetVersion.PY27}:
1068 "DEPRECATION: Python 2 support will be removed in the first stable release"
1069 "expected in January 2022."
1071 err(msg, fg="yellow", bold=True)
1073 normalize_fmt_off(src_node)
1074 lines = LineGenerator(
1076 remove_u_prefix="unicode_literals" in future_imports
1077 or supports_feature(versions, Feature.UNICODE_LITERALS),
1079 elt = EmptyLineTracker(is_pyi=mode.is_pyi)
1080 empty_line = Line(mode=mode)
1082 split_line_features = {
1084 for feature in {Feature.TRAILING_COMMA_IN_CALL, Feature.TRAILING_COMMA_IN_DEF}
1085 if supports_feature(versions, feature)
1087 for current_line in lines.visit(src_node):
1088 dst_contents.append(str(empty_line) * after)
1089 before, after = elt.maybe_empty_lines(current_line)
1090 dst_contents.append(str(empty_line) * before)
1091 for line in transform_line(
1092 current_line, mode=mode, features=split_line_features
1094 dst_contents.append(str(line))
1095 return "".join(dst_contents)
1098 def decode_bytes(src: bytes) -> Tuple[FileContent, Encoding, NewLine]:
1099 """Return a tuple of (decoded_contents, encoding, newline).
1101 `newline` is either CRLF or LF but `decoded_contents` is decoded with
1102 universal newlines (i.e. only contains LF).
1104 srcbuf = io.BytesIO(src)
1105 encoding, lines = tokenize.detect_encoding(srcbuf.readline)
1107 return "", encoding, "\n"
1109 newline = "\r\n" if b"\r\n" == lines[0][-2:] else "\n"
1111 with io.TextIOWrapper(srcbuf, encoding) as tiow:
1112 return tiow.read(), encoding, newline
1115 def get_features_used(node: Node) -> Set[Feature]: # noqa: C901
1116 """Return a set of (relatively) new Python features used in this file.
1118 Currently looking for:
1120 - underscores in numeric literals;
1121 - trailing commas after * or ** in function signatures and calls;
1122 - positional only arguments in function signatures and lambdas;
1123 - assignment expression;
1124 - relaxed decorator syntax;
1125 - print / exec statements;
1127 features: Set[Feature] = set()
1128 for n in node.pre_order():
1129 if n.type == token.STRING:
1130 value_head = n.value[:2] # type: ignore
1131 if value_head in {'f"', 'F"', "f'", "F'", "rf", "fr", "RF", "FR"}:
1132 features.add(Feature.F_STRINGS)
1134 elif n.type == token.NUMBER:
1135 if "_" in n.value: # type: ignore
1136 features.add(Feature.NUMERIC_UNDERSCORES)
1138 elif n.type == token.SLASH:
1139 if n.parent and n.parent.type in {
1144 features.add(Feature.POS_ONLY_ARGUMENTS)
1146 elif n.type == token.COLONEQUAL:
1147 features.add(Feature.ASSIGNMENT_EXPRESSIONS)
1149 elif n.type == syms.decorator:
1150 if len(n.children) > 1 and not is_simple_decorator_expression(
1153 features.add(Feature.RELAXED_DECORATORS)
1156 n.type in {syms.typedargslist, syms.arglist}
1158 and n.children[-1].type == token.COMMA
1160 if n.type == syms.typedargslist:
1161 feature = Feature.TRAILING_COMMA_IN_DEF
1163 feature = Feature.TRAILING_COMMA_IN_CALL
1165 for ch in n.children:
1166 if ch.type in STARS:
1167 features.add(feature)
1169 if ch.type == syms.argument:
1170 for argch in ch.children:
1171 if argch.type in STARS:
1172 features.add(feature)
1174 elif n.type == token.PRINT_STMT:
1175 features.add(Feature.PRINT_STMT)
1176 elif n.type == token.EXEC_STMT:
1177 features.add(Feature.EXEC_STMT)
1182 def detect_target_versions(node: Node) -> Set[TargetVersion]:
1183 """Detect the version to target based on the nodes used."""
1184 features = get_features_used(node)
1186 version for version in TargetVersion if features <= VERSION_TO_FEATURES[version]
1190 def get_future_imports(node: Node) -> Set[str]:
1191 """Return a set of __future__ imports in the file."""
1192 imports: Set[str] = set()
1194 def get_imports_from_children(children: List[LN]) -> Generator[str, None, None]:
1195 for child in children:
1196 if isinstance(child, Leaf):
1197 if child.type == token.NAME:
1200 elif child.type == syms.import_as_name:
1201 orig_name = child.children[0]
1202 assert isinstance(orig_name, Leaf), "Invalid syntax parsing imports"
1203 assert orig_name.type == token.NAME, "Invalid syntax parsing imports"
1204 yield orig_name.value
1206 elif child.type == syms.import_as_names:
1207 yield from get_imports_from_children(child.children)
1210 raise AssertionError("Invalid syntax parsing imports")
1212 for child in node.children:
1213 if child.type != syms.simple_stmt:
1216 first_child = child.children[0]
1217 if isinstance(first_child, Leaf):
1218 # Continue looking if we see a docstring; otherwise stop.
1220 len(child.children) == 2
1221 and first_child.type == token.STRING
1222 and child.children[1].type == token.NEWLINE
1228 elif first_child.type == syms.import_from:
1229 module_name = first_child.children[1]
1230 if not isinstance(module_name, Leaf) or module_name.value != "__future__":
1233 imports |= set(get_imports_from_children(first_child.children[3:]))
1240 def assert_equivalent(src: str, dst: str, *, pass_num: int = 1) -> None:
1241 """Raise AssertionError if `src` and `dst` aren't equivalent."""
1243 src_ast = parse_ast(src)
1244 except Exception as exc:
1245 raise AssertionError(
1246 "cannot use --safe with this file; failed to parse source file."
1250 dst_ast = parse_ast(dst)
1251 except Exception as exc:
1252 log = dump_to_file("".join(traceback.format_tb(exc.__traceback__)), dst)
1253 raise AssertionError(
1254 f"INTERNAL ERROR: Black produced invalid code on pass {pass_num}: {exc}. "
1255 "Please report a bug on https://github.com/psf/black/issues. "
1256 f"This invalid output might be helpful: {log}"
1259 src_ast_str = "\n".join(stringify_ast(src_ast))
1260 dst_ast_str = "\n".join(stringify_ast(dst_ast))
1261 if src_ast_str != dst_ast_str:
1262 log = dump_to_file(diff(src_ast_str, dst_ast_str, "src", "dst"))
1263 raise AssertionError(
1264 "INTERNAL ERROR: Black produced code that is not equivalent to the"
1265 f" source on pass {pass_num}. Please report a bug on "
1266 f"https://github.com/psf/black/issues. This diff might be helpful: {log}"
1270 def assert_stable(src: str, dst: str, mode: Mode) -> None:
1271 """Raise AssertionError if `dst` reformats differently the second time."""
1272 newdst = format_str(dst, mode=mode)
1276 diff(src, dst, "source", "first pass"),
1277 diff(dst, newdst, "first pass", "second pass"),
1279 raise AssertionError(
1280 "INTERNAL ERROR: Black produced different code on the second pass of the"
1281 " formatter. Please report a bug on https://github.com/psf/black/issues."
1282 f" This diff might be helpful: {log}"
1287 def nullcontext() -> Iterator[None]:
1288 """Return an empty context manager.
1290 To be used like `nullcontext` in Python 3.7.
1295 def patch_click() -> None:
1296 """Make Click not crash on Python 3.6 with LANG=C.
1298 On certain misconfigured environments, Python 3 selects the ASCII encoding as the
1299 default which restricts paths that it can access during the lifetime of the
1300 application. Click refuses to work in this scenario by raising a RuntimeError.
1302 In case of Black the likelihood that non-ASCII characters are going to be used in
1303 file paths is minimal since it's Python source code. Moreover, this crash was
1304 spurious on Python 3.7 thanks to PEP 538 and PEP 540.
1307 from click import core
1308 from click import _unicodefun
1309 except ModuleNotFoundError:
1312 for module in (core, _unicodefun):
1313 if hasattr(module, "_verify_python3_env"):
1314 module._verify_python3_env = lambda: None # type: ignore
1315 if hasattr(module, "_verify_python_env"):
1316 module._verify_python_env = lambda: None # type: ignore
1319 def patched_main() -> None:
1320 maybe_install_uvloop()
1326 if __name__ == "__main__":