import abc import os import stat from collections import deque from os import stat_result from typing import (Any, Dict, Iterable, Iterator, List, MutableMapping, Optional, Set, Text, Tuple, TYPE_CHECKING) from . import jsonlib from .utils import git # Cannot do `from ..gitignore import gitignore` because # relative import beyond toplevel throws *ImportError*! from gitignore import gitignore # type: ignore if TYPE_CHECKING: from .manifest import Manifest # avoid cyclic import GitIgnoreCacheType = MutableMapping[bytes, bool] def get_tree(tests_root: Text, manifest: "Manifest", manifest_path: Optional[Text], cache_root: Optional[Text], working_copy: bool = True, rebuild: bool = False) -> "FileSystem": tree = None if cache_root is None: cache_root = os.path.join(tests_root, ".wptcache") if not os.path.exists(cache_root): try: os.makedirs(cache_root) except OSError: cache_root = None if not working_copy: raise ValueError("working_copy=False unsupported") if tree is None: tree = FileSystem(tests_root, manifest.url_base, manifest_path=manifest_path, cache_path=cache_root, rebuild=rebuild) return tree class GitHasher: def __init__(self, path: Text) -> None: self.git = git(path) def _local_changes(self) -> Set[Text]: """get a set of files which have changed between HEAD and working copy""" assert self.git is not None # note that git runs the command with tests_root as the cwd, which may # not be the root of the git repo (e.g., within a browser repo) cmd = ["diff-index", "--relative", "--no-renames", "--name-only", "-z", "HEAD"] data = self.git(*cmd) return set(data.split("\0")) def hash_cache(self) -> Dict[Text, Optional[Text]]: """ A dict of rel_path -> current git object id if the working tree matches HEAD else None """ hash_cache: Dict[Text, Optional[Text]] = {} if self.git is None: return hash_cache # note that git runs the command with tests_root as the cwd, which may # not be the root of the git repo (e.g., within a browser repo) cmd = ["ls-tree", "-r", "-z", "HEAD"] local_changes = self._local_changes() for result in self.git(*cmd).split("\0")[:-1]: # type: Text data, rel_path = result.rsplit("\t", 1) hash_cache[rel_path] = None if rel_path in local_changes else data.split(" ", 3)[2] return hash_cache class FileSystem: def __init__(self, tests_root: Text, url_base: Text, cache_path: Optional[Text], manifest_path: Optional[Text] = None, rebuild: bool = False) -> None: self.tests_root = tests_root self.url_base = url_base self.ignore_cache = None self.mtime_cache = None tests_root_bytes = tests_root.encode("utf8") if cache_path is not None: if manifest_path is not None: self.mtime_cache = MtimeCache(cache_path, tests_root, manifest_path, rebuild) if gitignore.has_ignore(tests_root_bytes): self.ignore_cache = GitIgnoreCache(cache_path, tests_root, rebuild) self.path_filter = gitignore.PathFilter(tests_root_bytes, extras=[b".git/"], cache=self.ignore_cache) git = GitHasher(tests_root) self.hash_cache = git.hash_cache() def __iter__(self) -> Iterator[Tuple[Text, Optional[Text], bool]]: mtime_cache = self.mtime_cache for dirpath, dirnames, filenames in self.path_filter( walk(self.tests_root.encode("utf8"))): for filename, path_stat in filenames: path = os.path.join(dirpath, filename).decode("utf8") if mtime_cache is None or mtime_cache.updated(path, path_stat): file_hash = self.hash_cache.get(path, None) yield path, file_hash, True else: yield path, None, False def dump_caches(self) -> None: for cache in [self.mtime_cache, self.ignore_cache]: if cache is not None: cache.dump() class CacheFile(metaclass=abc.ABCMeta): def __init__(self, cache_root: Text, tests_root: Text, rebuild: bool = False) -> None: self.tests_root = tests_root if not os.path.exists(cache_root): os.makedirs(cache_root) self.path = os.path.join(cache_root, self.file_name) self.modified = False self.data = self.load(rebuild) @abc.abstractproperty def file_name(self) -> Text: pass def dump(self) -> None: if not self.modified: return with open(self.path, 'w') as f: jsonlib.dump_local(self.data, f) def load(self, rebuild: bool = False) -> Dict[Text, Any]: data: Dict[Text, Any] = {} try: if not rebuild: with open(self.path) as f: try: data = jsonlib.load(f) except ValueError: pass data = self.check_valid(data) except OSError: pass return data def check_valid(self, data: Dict[Text, Any]) -> Dict[Text, Any]: """Check if the cached data is valid and return an updated copy of the cache containing only data that can be used.""" return data class MtimeCache(CacheFile): file_name = "mtime.json" def __init__(self, cache_root: Text, tests_root: Text, manifest_path: Text, rebuild: bool = False) -> None: self.manifest_path = manifest_path super().__init__(cache_root, tests_root, rebuild) def updated(self, rel_path: Text, stat: stat_result) -> bool: """Return a boolean indicating whether the file changed since the cache was last updated. This implicitly updates the cache with the new mtime data.""" mtime = stat.st_mtime if mtime != self.data.get(rel_path): self.modified = True self.data[rel_path] = mtime return True return False def check_valid(self, data: Dict[Any, Any]) -> Dict[Any, Any]: if data.get("/tests_root") != self.tests_root: self.modified = True else: if self.manifest_path is not None and os.path.exists(self.manifest_path): mtime = os.path.getmtime(self.manifest_path) if data.get("/manifest_path") != [self.manifest_path, mtime]: self.modified = True else: self.modified = True if self.modified: data = {} data["/tests_root"] = self.tests_root return data def dump(self) -> None: if self.manifest_path is None: raise ValueError if not os.path.exists(self.manifest_path): return mtime = os.path.getmtime(self.manifest_path) self.data["/manifest_path"] = [self.manifest_path, mtime] self.data["/tests_root"] = self.tests_root super().dump() class GitIgnoreCache(CacheFile, GitIgnoreCacheType): file_name = "gitignore2.json" def check_valid(self, data: Dict[Any, Any]) -> Dict[Any, Any]: ignore_path = os.path.join(self.tests_root, ".gitignore") mtime = os.path.getmtime(ignore_path) if data.get("/gitignore_file") != [ignore_path, mtime]: self.modified = True data = {} data["/gitignore_file"] = [ignore_path, mtime] return data def __contains__(self, key: Any) -> bool: try: key = key.decode("utf-8") except Exception: return False return key in self.data def __getitem__(self, key: bytes) -> bool: real_key = key.decode("utf-8") v = self.data[real_key] assert isinstance(v, bool) return v def __setitem__(self, key: bytes, value: bool) -> None: real_key = key.decode("utf-8") if self.data.get(real_key) != value: self.modified = True self.data[real_key] = value def __delitem__(self, key: bytes) -> None: real_key = key.decode("utf-8") del self.data[real_key] def __iter__(self) -> Iterator[bytes]: return (key.encode("utf-8") for key in self.data) def __len__(self) -> int: return len(self.data) def walk(root: bytes) -> Iterable[Tuple[bytes, List[Tuple[bytes, stat_result]], List[Tuple[bytes, stat_result]]]]: """Re-implementation of os.walk. Returns an iterator over (dirpath, dirnames, filenames), with some semantic differences to os.walk. This has a similar interface to os.walk, with the important difference that instead of lists of filenames and directory names, it yields lists of tuples of the form [(name, stat)] where stat is the result of os.stat for the file. That allows reusing the same stat data in the caller. It also always returns the dirpath relative to the root, with the root iself being returned as the empty string. Unlike os.walk the implementation is not recursive.""" get_stat = os.stat is_dir = stat.S_ISDIR is_link = stat.S_ISLNK join = os.path.join listdir = os.listdir relpath = os.path.relpath root = os.path.abspath(root) stack = deque([(root, b"")]) while stack: dir_path, rel_path = stack.popleft() try: # Note that listdir and error are globals in this module due # to earlier import-*. names = listdir(dir_path) except OSError: continue dirs, non_dirs = [], [] for name in names: path = join(dir_path, name) try: path_stat = get_stat(path) except OSError: continue if is_dir(path_stat.st_mode): dirs.append((name, path_stat)) else: non_dirs.append((name, path_stat)) yield rel_path, dirs, non_dirs for name, path_stat in dirs: new_path = join(dir_path, name) if not is_link(path_stat.st_mode): stack.append((new_path, relpath(new_path, root)))