Source code for swh.model.from_disk

# Copyright (C) 2017-2022 The Software Heritage developers
# See the AUTHORS file at the top-level directory of this distribution
# License: GNU General Public License version 3, or any later version
# See top-level LICENSE file for more information

"""Conversion from filesystem tree to SWH objects.

This module allows reading a tree of directories and files from a local
filesystem, and convert them to in-memory data structures, which can then
be exported to SWH data model objects, as defined in :mod:`swh.model.model`.
"""

import enum
import fnmatch
import glob
import os
import re
import stat
from typing import (
    Any,
    Callable,
    Dict,
    Iterable,
    Iterator,
    List,
    Optional,
    Pattern,
    Tuple,
    Union,
    cast,
)
import warnings

import attr
from deprecated import deprecated
from typing_extensions import Final

from . import model
from .exceptions import InvalidDirectoryPath
from .git_objects import directory_entry_sort_key
from .hashutil import MultiHash, hash_to_hex
from .merkle import MerkleLeaf, MerkleNode
from .swhids import CoreSWHID
from .swhids import ObjectType as SWHIDType


[docs] class FromDiskType(model._StringCompatibleEnum): """Possible object types for "from disk" object.""" CONTENT = "content" DIRECTORY = "directory" def __eq__(self, other): # stay compatible with legacy string comparison (for now) if isinstance(other, str): # note: we should issue deprecation warning at some point return self.value == other return super().__eq__(other) def __str__(self): # preserve interpolation property (for now) return self.value def __hash__(self): # make sure we don't confuse dictionary key matching (for now) return hash(str(self.value))
# There is an handful of other module that test for # DiskBackedContent.object_type in conjunction of Content.object_type, give # them a hand to let them handle compatibility in a smoother way. # # Remove this compatibility trick once this user have been migrated
[docs] @deprecated(version="v6.13.0", reason="Use model.Content.object_type instead") def DiskBackedContent(*args, **kwargs): return model.Content(*args, **kwargs)
[docs] @attr.s(frozen=True, slots=True) class DiskBackedData: path = attr.ib(type=bytes) def __call__(self) -> bytes: with open(self.path, "rb") as fd: return fd.read()
[docs] class DentryPerms(enum.IntEnum): """Admissible permissions for directory entries.""" content = 0o100644 """Content""" executable_content = 0o100755 """Executable content (e.g. executable script)""" symlink = 0o120000 """Symbolic link""" directory = 0o040000 """Directory""" revision = 0o160000 """Revision (e.g. submodule)"""
[docs] def mode_to_perms(mode): """Convert a file mode to a permission compatible with Software Heritage directory entries Args: mode (int): a file mode as returned by :func:`os.stat` in :attr:`os.stat_result.st_mode` Returns: DentryPerms: one of the following values: :const:`DentryPerms.content`: plain file :const:`DentryPerms.executable_content`: executable file :const:`DentryPerms.symlink`: symbolic link :const:`DentryPerms.directory`: directory """ if stat.S_ISLNK(mode): return DentryPerms.symlink if stat.S_ISDIR(mode): return DentryPerms.directory else: # file is executable in any way if mode & (0o111): return DentryPerms.executable_content else: return DentryPerms.content
[docs] class Content(MerkleLeaf): """Representation of a Software Heritage content as a node in a Merkle tree. The current Merkle hash for the Content nodes is the `sha1_git`, which makes it consistent with what :class:`Directory` uses for its own hash computation. """ __slots__: List[str] = [] object_type: Final = FromDiskType.CONTENT
[docs] @classmethod def from_bytes(cls, *, mode, data): """Convert data (raw :class:`bytes`) to a Software Heritage content entry Args: mode (int): a file mode (passed to :func:`mode_to_perms`) data (bytes): raw contents of the file """ ret = MultiHash.from_data(data).digest() ret["length"] = len(data) ret["perms"] = mode_to_perms(mode) ret["data"] = data ret["status"] = "visible" return cls(ret)
[docs] @classmethod def from_file(cls, *, path, max_content_length=None): """Compute the Software Heritage content entry corresponding to an on-disk file. The returned dictionary contains keys useful for both: - loading the content in the archive (hashes, `length`) - using the content as a directory entry in a directory Args: save_path (bool): add the file path to the entry max_content_length (Optional[int]): if given, all contents larger than this will be skipped. """ file_stat = os.lstat(path) mode = file_stat.st_mode length = file_stat.st_size too_large = max_content_length is not None and length > max_content_length if stat.S_ISLNK(mode): # Symbolic link: return a file whose contents are the link target if too_large: # Unlike large contents, we can't stream symlinks to # MultiHash, and we don't want to fit them in memory if # they exceed max_content_length either. # Thankfully, this should not happen for reasonable values of # max_content_length because of OS/filesystem limitations, # so let's just raise an error. raise Exception(f"Symlink too large ({length} bytes)") return cls.from_symlink(path=path, mode=mode) elif not stat.S_ISREG(mode): # not a regular file: return the empty file instead return cls.from_bytes(mode=mode, data=b"") if too_large: skip_reason = "Content too large" else: skip_reason = None hashes = MultiHash.from_path(path).digest() if skip_reason: ret = { **hashes, "status": "absent", "reason": skip_reason, } else: ret = { **hashes, "status": "visible", } ret["path"] = path ret["perms"] = mode_to_perms(mode) ret["length"] = length obj = cls(ret) return obj
[docs] def swhid(self) -> CoreSWHID: """Return node identifier as a SWHID""" return CoreSWHID(object_type=SWHIDType.CONTENT, object_id=self.hash)
def __repr__(self): return "Content(id=%s)" % hash_to_hex(self.hash)
[docs] def compute_hash(self): return self.data["sha1_git"]
[docs] def to_model(self) -> model.BaseContent: """Builds a `model.BaseContent` object based on this leaf.""" data = self.get_data().copy() data.pop("perms", None) path = data.pop("path", None) if data["status"] == "absent": return model.SkippedContent.from_dict(data) elif "data" not in data: data["get_data"] = DiskBackedData(path=path) return model.Content.from_dict(data)
[docs] def accept_all_directories( dirpath: bytes, dirname: bytes, entries: Optional[Iterable[Any]] ) -> bool: """Default filter for :func:`Directory.from_disk` accepting all directories Args: dirname (bytes): directory name entries (list): directory entries """ warnings.warn( "`accept_all_directories` is deprecated, use `accept_all_paths`", DeprecationWarning, ) return True
[docs] def accept_all_paths( path: bytes, name: bytes, entries: Optional[Iterable[Any]] ) -> bool: """Default filter for :func:`Directory.from_disk` accepting all paths""" return True
[docs] def ignore_empty_directories( dirpath: bytes, dirname: bytes, entries: Optional[Iterable[Any]] ) -> bool: """Filter for :func:`directory_to_objects` ignoring empty directories Args: dirname (bytes): directory name entries (list): directory entries Returns: True if the directory is not empty, false if the directory is empty """ if entries is None: # Files are not ignored return True return bool(entries)
[docs] def ignore_named_directories(names, *, case_sensitive=True): """Filter for :func:`directory_to_objects` to ignore directories named one of names. Args: names (list of bytes): names to ignore case_sensitive (bool): whether to do the filtering in a case sensitive way Returns: a directory filter for :func:`directory_to_objects` """ if not case_sensitive: names = [name.lower() for name in names] def named_filter( dirpath: str, dirname: str, entries: Iterable[Any], names: Iterable[Any] = names, case_sensitive: bool = case_sensitive, ): if entries is None: # Files are not ignored return True if case_sensitive: return dirname not in names else: return dirname.lower() not in names return named_filter
# TODO: `extract_regex_objs` has been copied and adapted from `swh.scanner`. # In the future `swh.scanner` should use the `swh.model` version and remove its own.
[docs] def extract_regex_objs( root_path: bytes, patterns: Iterable[bytes] ) -> Iterator[Pattern[bytes]]: """Generates a regex object for each pattern given in input and checks if the path is a subdirectory or relative to the root path. Args: root_path (bytes): path to the root directory patterns (list of byte): shell patterns to match Yields: an SRE_Pattern object """ absolute_root_path = os.path.abspath(root_path) for pattern in patterns: if os.path.isabs(pattern): pattern = os.path.relpath(pattern, root_path) # python 3.10 has a `root_dir` argument for glob, but not the previous # version. So we adjust the pattern test_pattern = os.path.join(absolute_root_path, pattern) for path in glob.glob(test_pattern): if os.path.isabs(path) and not path.startswith(absolute_root_path): error_msg = ( b'The path "' + path + b'" is not a subdirectory or relative ' b'to the root directory path: "' + root_path + b'"' ) raise InvalidDirectoryPath(error_msg) regex = fnmatch.translate((pattern.decode())) yield re.compile(regex.encode())
[docs] def ignore_directories_patterns(root_path: bytes, patterns: Iterable[bytes]): """Filter for :func:`directory_to_objects` to ignore directories matching certain patterns. Args: root_path (bytes): path of the root directory patterns (list of bytes): patterns to ignore Returns: a directory filter for :func:`directory_to_objects` """ sre_patterns = set(extract_regex_objs(root_path, patterns)) def pattern_filter( dirpath: bytes, dirname: bytes, entries: Iterable[Any], patterns: Iterable[Any] = sre_patterns, root_path: bytes = os.path.abspath(root_path), ): full_path = os.path.abspath(os.path.join(dirpath, dirname)) relative_path = os.path.relpath(full_path, root_path) return not any([pattern.match(relative_path) for pattern in patterns]) return pattern_filter
[docs] def iter_directory( directory: "Directory", ) -> Tuple[List[model.Content], List[model.SkippedContent], List[model.Directory]]: """Return the directory listing from a disk-memory directory instance. Raises: TypeError in case an unexpected object type is listed. Returns: Tuple of respectively iterable of content, skipped content and directories. """ contents: List[model.Content] = [] skipped_contents: List[model.SkippedContent] = [] directories: List[model.Directory] = [] for i_obj in directory.iter_tree(): if isinstance(i_obj, Directory): directories.append(i_obj.to_model()) elif isinstance(i_obj, Content): obj = i_obj.to_model() if isinstance(obj, model.SkippedContent): skipped_contents.append(obj) else: # FIXME: read the data from disk later (when the # storage buffer is flushed). # c_obj = cast(model.Content, obj) contents.append(c_obj.with_data()) else: raise TypeError(f"Unexpected object type from disk: {obj}") return contents, skipped_contents, directories
[docs] class Directory(MerkleNode): """Representation of a Software Heritage directory as a node in a Merkle Tree. This class can be used to generate, from an on-disk directory, all the objects that need to be sent to the Software Heritage archive. The :func:`from_disk` constructor allows you to generate the data structure from a directory on disk. The resulting :class:`Directory` can then be manipulated as a dictionary, using the path as key. The :func:`collect` method is used to retrieve all the objects that need to be added to the Software Heritage archive since the last collection, by class (contents and directories). When using the dict-like methods to update the contents of the directory, the affected levels of hierarchy are reset and can be collected again using the same method. This enables the efficient collection of updated nodes, for instance when the client is applying diffs. """ __slots__ = ["__entries", "__model_object"] object_type: Final = FromDiskType.DIRECTORY
[docs] @classmethod def from_disk( cls, *, path: bytes, path_filter: Callable[ [bytes, bytes, Optional[List[bytes]]], bool ] = accept_all_paths, max_content_length: Optional[int] = None, progress_callback: Optional[Callable[[int], None]] = None, ) -> "Directory": """Compute the Software Heritage objects for a given directory tree Args: path (bytes): the directory to traverse data (bool): whether to add the data to the content objects save_path (bool): whether to add the path to the content objects path_filter (function): a filter to ignore some paths. Takes three arguments: `path`, `name` and `entries`. `entries` is `None` for files, and a (possibly empty) list of names for directories. Returns True if the path should be added, False if the path should be ignored. max_content_length (Optional[int]): if given, all contents larger than this will be skipped. progress_callback (Optional function): if given, returns for each non empty directories traversed the number of computed entries. """ # the top might have been specified with a final slash. This will # confuse various code. # # We should prevent '/' as is however if 1 < len(path) and path[-1:] == b"/": path = path[0:1] + path[1:].rstrip(b"/") assert len(path) <= 1 or path[-1:] != b"/" top_path = path top_path_prefix_size = len(top_path) + 1 dirs: Dict[bytes, Directory] = {} dirs[top_path] = cls({"name": os.path.basename(top_path), "path": top_path}) filtered = [] to_visit = [path] while to_visit: root = to_visit.pop() path, name = os.path.split(root) with os.scandir(root) as it: entries_list = list(it) if root != top_path and not path_filter( path, name, [entry.path for entry in entries_list] ): # we should not traverse the current directory, so stop right now, # but also mark it as removed (for later cleanup) filtered.append(root) continue entries = {} for entry in entries_list: if not entry.is_dir(follow_symlinks=False): if not path_filter(root, entry.name, None): continue content = Content.from_file( path=entry.path, max_content_length=max_content_length ) entries[entry.name] = content else: entries[entry.name] = cls({"name": entry.name, "path": entry.path}) dirs[entry.path] = entries[entry.name] to_visit.append(entry.path) dirs[root].update(entries) if progress_callback is not None: if len(entries) > 0: progress_callback(len(entries)) top_dir = dirs[top_path] for path in reversed(filtered): path = path[top_path_prefix_size:] del top_dir[path] # a bit sad but now we have to traverse the gathered tree structure to # filter it again (e.g. for the ignore_empty_directory filter to work # recursively) todo: List[Tuple[bytes, Directory]] = [(b"", top_dir)] traversal = [] while todo: cpath, cdir = todo.pop(0) traversal.append(cpath) for dirname, subdir in cdir.items(): if subdir.object_type == FromDiskType.DIRECTORY: spath = cpath + b"/" + dirname todo.append((spath, subdir)) for dirpath in reversed(traversal): node = top_dir[dirpath] assert node.object_type == FromDiskType.DIRECTORY path, name = os.path.split(dirpath) if dirpath and not path_filter(path, name, list(node.keys())): # should be filtered del top_dir[dirpath] top_dir.update_hash(force=True) return top_dir
def __init__(self, data=None): super().__init__(data=data) self.__entries = None self.__model_object = None # note: the overwrite could probably be done by parametrysing the # MerkelNode type, but that is a much bigger rework than the series # introducing this change.
[docs] def iter_tree(self, dedup=True) -> Iterator[Union["Directory", "Content"]]: """Yields all children nodes, recursively. Common nodes are deduplicated by default (deduplication can be turned off setting the given argument 'dedup' to False). """ tree = super().iter_tree(dedup=dedup) yield from cast(Iterator[Union["Directory", "Content"]], tree)
[docs] def invalidate_hash(self): self.__entries = None self.__model_object = None super().invalidate_hash()
[docs] @staticmethod def child_to_directory_entry(name, child): if child.object_type == FromDiskType.DIRECTORY: return { "type": "dir", "perms": DentryPerms.directory, "target": child.hash, "name": name, } elif child.object_type == FromDiskType.CONTENT: return { "type": "file", "perms": child.data["perms"], "target": child.hash, "name": name, } else: raise ValueError(f"unknown child {child}")
[docs] def get_data(self, **kwargs): return { "id": self.hash, "entries": self.entries, }
@property def entries(self): """Child nodes, sorted by name in the same way :func:`swh.model.git_objects.directory_git_object` does.""" if self.__entries is None: self.__entries = sorted( ( self.child_to_directory_entry(name, child) for name, child in self.items() ), key=directory_entry_sort_key, ) return self.__entries
[docs] def swhid(self) -> CoreSWHID: """Return node identifier as a SWHID""" return CoreSWHID(object_type=SWHIDType.DIRECTORY, object_id=self.hash)
[docs] def compute_hash(self): return self.to_model().id
[docs] def to_model(self) -> model.Directory: """Builds a `model.Directory` object based on this node; ignoring its children.""" if self.__model_object is None: DirectoryEntry = model.DirectoryEntry entries = [] for name, child in self.items(): if child.object_type == FromDiskType.DIRECTORY: e = DirectoryEntry( type="dir", perms=DentryPerms.directory, target=child.hash, name=name, ) elif child.object_type == FromDiskType.CONTENT: e = DirectoryEntry( type="file", perms=child.data["perms"], target=child.hash, name=name, ) else: raise ValueError(f"unknown child {child}") entries.append(e) entries.sort(key=directory_entry_sort_key) self.__model_object = model.Directory(entries=tuple(entries)) return self.__model_object
def __getitem__(self, key): if not isinstance(key, bytes): raise ValueError("Can only get a bytes from Directory") # Convenience shortcut if key == b"": return self if b"/" not in key: return super().__getitem__(key) else: key1, key2 = key.split(b"/", 1) return self.__getitem__(key1)[key2] def __setitem__(self, key, value): if not isinstance(key, bytes): raise ValueError("Can only set a bytes Directory entry") if not isinstance(value, (Content, Directory)): raise ValueError( "Can only set a Directory entry to a Content or " "Directory" ) if key == b"": raise ValueError("Directory entry must have a name") if b"\x00" in key: raise ValueError("Directory entry name must not contain nul bytes") if b"/" not in key: return super().__setitem__(key, value) else: key1, key2 = key.rsplit(b"/", 1) self[key1].__setitem__(key2, value) def __delitem__(self, key): if not isinstance(key, bytes): raise ValueError("Can only delete a bytes Directory entry") if b"/" not in key: super().__delitem__(key) else: key1, key2 = key.rsplit(b"/", 1) del self[key1][key2] def __contains__(self, key): if b"/" not in key: return super().__contains__(key) else: key1, key2 = key.split(b"/", 1) return super().__contains__(key1) and self[key1].__contains__(key2) def __repr__(self): return "Directory(id=%s, entries=[%s])" % ( hash_to_hex(self.hash), ", ".join(str(entry) for entry in self), )