summaryrefslogtreecommitdiffstats
path: root/ansible_collections/community/general/plugins/modules/archive.py
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-13 12:04:41 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-13 12:04:41 +0000
commit975f66f2eebe9dadba04f275774d4ab83f74cf25 (patch)
tree89bd26a93aaae6a25749145b7e4bca4a1e75b2be /ansible_collections/community/general/plugins/modules/archive.py
parentInitial commit. (diff)
downloadansible-975f66f2eebe9dadba04f275774d4ab83f74cf25.tar.xz
ansible-975f66f2eebe9dadba04f275774d4ab83f74cf25.zip
Adding upstream version 7.7.0+dfsg.upstream/7.7.0+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'ansible_collections/community/general/plugins/modules/archive.py')
-rw-r--r--ansible_collections/community/general/plugins/modules/archive.py686
1 files changed, 686 insertions, 0 deletions
diff --git a/ansible_collections/community/general/plugins/modules/archive.py b/ansible_collections/community/general/plugins/modules/archive.py
new file mode 100644
index 000000000..8748fb8a3
--- /dev/null
+++ b/ansible_collections/community/general/plugins/modules/archive.py
@@ -0,0 +1,686 @@
+#!/usr/bin/python
+# -*- coding: utf-8 -*-
+
+# Copyright (c) 2016, Ben Doherty <bendohmv@gmail.com>
+# Sponsored by Oomph, Inc. http://www.oomphinc.com
+# Copyright (c) 2017, Ansible Project
+# GNU General Public License v3.0+ (see LICENSES/GPL-3.0-or-later.txt or https://www.gnu.org/licenses/gpl-3.0.txt)
+# SPDX-License-Identifier: GPL-3.0-or-later
+
+from __future__ import absolute_import, division, print_function
+__metaclass__ = type
+
+DOCUMENTATION = r'''
+---
+module: archive
+short_description: Creates a compressed archive of one or more files or trees
+extends_documentation_fragment:
+ - files
+ - community.general.attributes
+description:
+ - Creates or extends an archive.
+ - The source and archive are on the remote host, and the archive I(is not) copied to the local host.
+ - Source files can be deleted after archival by specifying I(remove=True).
+attributes:
+ check_mode:
+ support: full
+ diff_mode:
+ support: none
+options:
+ path:
+ description:
+ - Remote absolute path, glob, or list of paths or globs for the file or files to compress or archive.
+ type: list
+ elements: path
+ required: true
+ format:
+ description:
+ - The type of compression to use.
+ - Support for xz was added in Ansible 2.5.
+ type: str
+ choices: [ bz2, gz, tar, xz, zip ]
+ default: gz
+ dest:
+ description:
+ - The file name of the destination archive. The parent directory must exists on the remote host.
+ - This is required when C(path) refers to multiple files by either specifying a glob, a directory or multiple paths in a list.
+ - If the destination archive already exists, it will be truncated and overwritten.
+ type: path
+ exclude_path:
+ description:
+ - Remote absolute path, glob, or list of paths or globs for the file or files to exclude from I(path) list and glob expansion.
+ - Use I(exclusion_patterns) to instead exclude files or subdirectories below any of the paths from the I(path) list.
+ type: list
+ elements: path
+ default: []
+ exclusion_patterns:
+ description:
+ - Glob style patterns to exclude files or directories from the resulting archive.
+ - This differs from I(exclude_path) which applies only to the source paths from I(path).
+ type: list
+ elements: path
+ version_added: 3.2.0
+ force_archive:
+ description:
+ - Allows you to force the module to treat this as an archive even if only a single file is specified.
+ - By default when a single file is specified it is compressed only (not archived).
+ - Enable this if you want to use M(ansible.builtin.unarchive) on an archive of a single file created with this module.
+ type: bool
+ default: false
+ remove:
+ description:
+ - Remove any added source files and trees after adding to archive.
+ type: bool
+ default: false
+notes:
+ - Can produce I(gzip), I(bzip2), I(lzma), and I(zip) compressed files or archives.
+ - This module uses C(tarfile), C(zipfile), C(gzip), and C(bz2) packages on the target host to create archives.
+ These are part of the Python standard library for Python 2 and 3.
+requirements:
+ - Requires C(lzma) (standard library of Python 3) or L(backports.lzma, https://pypi.org/project/backports.lzma/) (Python 2) if using C(xz) format.
+seealso:
+ - module: ansible.builtin.unarchive
+author:
+ - Ben Doherty (@bendoh)
+'''
+
+EXAMPLES = r'''
+- name: Compress directory /path/to/foo/ into /path/to/foo.tgz
+ community.general.archive:
+ path: /path/to/foo
+ dest: /path/to/foo.tgz
+
+- name: Compress regular file /path/to/foo into /path/to/foo.gz and remove it
+ community.general.archive:
+ path: /path/to/foo
+ remove: true
+
+- name: Create a zip archive of /path/to/foo
+ community.general.archive:
+ path: /path/to/foo
+ format: zip
+
+- name: Create a bz2 archive of multiple files, rooted at /path
+ community.general.archive:
+ path:
+ - /path/to/foo
+ - /path/wong/foo
+ dest: /path/file.tar.bz2
+ format: bz2
+
+- name: Create a bz2 archive of a globbed path, while excluding specific dirnames
+ community.general.archive:
+ path:
+ - /path/to/foo/*
+ dest: /path/file.tar.bz2
+ exclude_path:
+ - /path/to/foo/bar
+ - /path/to/foo/baz
+ format: bz2
+
+- name: Create a bz2 archive of a globbed path, while excluding a glob of dirnames
+ community.general.archive:
+ path:
+ - /path/to/foo/*
+ dest: /path/file.tar.bz2
+ exclude_path:
+ - /path/to/foo/ba*
+ format: bz2
+
+- name: Use gzip to compress a single archive (i.e don't archive it first with tar)
+ community.general.archive:
+ path: /path/to/foo/single.file
+ dest: /path/file.gz
+ format: gz
+
+- name: Create a tar.gz archive of a single file.
+ community.general.archive:
+ path: /path/to/foo/single.file
+ dest: /path/file.tar.gz
+ format: gz
+ force_archive: true
+'''
+
+RETURN = r'''
+state:
+ description:
+ The state of the input C(path).
+ type: str
+ returned: always
+dest_state:
+ description:
+ - The state of the I(dest) file.
+ - C(absent) when the file does not exist.
+ - C(archive) when the file is an archive.
+ - C(compress) when the file is compressed, but not an archive.
+ - C(incomplete) when the file is an archive, but some files under I(path) were not found.
+ type: str
+ returned: success
+ version_added: 3.4.0
+missing:
+ description: Any files that were missing from the source.
+ type: list
+ returned: success
+archived:
+ description: Any files that were compressed or added to the archive.
+ type: list
+ returned: success
+arcroot:
+ description: The archive root.
+ type: str
+ returned: always
+expanded_paths:
+ description: The list of matching paths from paths argument.
+ type: list
+ returned: always
+expanded_exclude_paths:
+ description: The list of matching exclude paths from the exclude_path argument.
+ type: list
+ returned: always
+'''
+
+import abc
+import bz2
+import glob
+import gzip
+import io
+import os
+import re
+import shutil
+import tarfile
+import zipfile
+from fnmatch import fnmatch
+from sys import version_info
+from traceback import format_exc
+from zlib import crc32
+
+from ansible.module_utils.basic import AnsibleModule, missing_required_lib
+from ansible.module_utils.common.text.converters import to_bytes, to_native
+from ansible.module_utils import six
+
+try: # python 3.2+
+ from zipfile import BadZipFile # type: ignore[attr-defined]
+except ImportError: # older python
+ from zipfile import BadZipfile as BadZipFile
+
+LZMA_IMP_ERR = None
+if six.PY3:
+ try:
+ import lzma
+ HAS_LZMA = True
+ except ImportError:
+ LZMA_IMP_ERR = format_exc()
+ HAS_LZMA = False
+else:
+ try:
+ from backports import lzma
+ HAS_LZMA = True
+ except ImportError:
+ LZMA_IMP_ERR = format_exc()
+ HAS_LZMA = False
+
+PY27 = version_info[0:2] >= (2, 7)
+
+STATE_ABSENT = 'absent'
+STATE_ARCHIVED = 'archive'
+STATE_COMPRESSED = 'compress'
+STATE_INCOMPLETE = 'incomplete'
+
+
+def common_path(paths):
+ empty = b'' if paths and isinstance(paths[0], six.binary_type) else ''
+
+ return os.path.join(
+ os.path.dirname(os.path.commonprefix([os.path.join(os.path.dirname(p), empty) for p in paths])), empty
+ )
+
+
+def expand_paths(paths):
+ expanded_path = []
+ is_globby = False
+ for path in paths:
+ b_path = _to_bytes(path)
+ if b'*' in b_path or b'?' in b_path:
+ e_paths = glob.glob(b_path)
+ is_globby = True
+ else:
+ e_paths = [b_path]
+ expanded_path.extend(e_paths)
+ return expanded_path, is_globby
+
+
+def matches_exclusion_patterns(path, exclusion_patterns):
+ return any(fnmatch(path, p) for p in exclusion_patterns)
+
+
+def is_archive(path):
+ return re.search(br'\.(tar|tar\.(gz|bz2|xz)|tgz|tbz2|zip)$', os.path.basename(path), re.IGNORECASE)
+
+
+def strip_prefix(prefix, string):
+ return string[len(prefix):] if string.startswith(prefix) else string
+
+
+def _to_bytes(s):
+ return to_bytes(s, errors='surrogate_or_strict')
+
+
+def _to_native(s):
+ return to_native(s, errors='surrogate_or_strict')
+
+
+def _to_native_ascii(s):
+ return to_native(s, errors='surrogate_or_strict', encoding='ascii')
+
+
+@six.add_metaclass(abc.ABCMeta)
+class Archive(object):
+ def __init__(self, module):
+ self.module = module
+
+ self.destination = _to_bytes(module.params['dest']) if module.params['dest'] else None
+ self.exclusion_patterns = module.params['exclusion_patterns'] or []
+ self.format = module.params['format']
+ self.must_archive = module.params['force_archive']
+ self.remove = module.params['remove']
+
+ self.changed = False
+ self.destination_state = STATE_ABSENT
+ self.errors = []
+ self.file = None
+ self.successes = []
+ self.targets = []
+ self.not_found = []
+
+ paths = module.params['path']
+ self.expanded_paths, has_globs = expand_paths(paths)
+ self.expanded_exclude_paths = expand_paths(module.params['exclude_path'])[0]
+
+ self.paths = sorted(set(self.expanded_paths) - set(self.expanded_exclude_paths))
+
+ if not self.paths:
+ module.fail_json(
+ path=', '.join(paths),
+ expanded_paths=_to_native(b', '.join(self.expanded_paths)),
+ expanded_exclude_paths=_to_native(b', '.join(self.expanded_exclude_paths)),
+ msg='Error, no source paths were found'
+ )
+
+ self.root = common_path(self.paths)
+
+ if not self.must_archive:
+ self.must_archive = any([has_globs, os.path.isdir(self.paths[0]), len(self.paths) > 1])
+
+ if not self.destination and not self.must_archive:
+ self.destination = b'%s.%s' % (self.paths[0], _to_bytes(self.format))
+
+ if self.must_archive and not self.destination:
+ module.fail_json(
+ dest=_to_native(self.destination),
+ path=', '.join(paths),
+ msg='Error, must specify "dest" when archiving multiple files or trees'
+ )
+
+ if self.remove:
+ self._check_removal_safety()
+
+ self.original_checksums = self.destination_checksums()
+ self.original_size = self.destination_size()
+
+ def add(self, path, archive_name):
+ try:
+ self._add(_to_native_ascii(path), _to_native(archive_name))
+ if self.contains(_to_native(archive_name)):
+ self.successes.append(path)
+ except Exception as e:
+ self.errors.append('%s: %s' % (_to_native_ascii(path), _to_native(e)))
+
+ def add_single_target(self, path):
+ if self.format in ('zip', 'tar'):
+ self.open()
+ self.add(path, strip_prefix(self.root, path))
+ self.close()
+ self.destination_state = STATE_ARCHIVED
+ else:
+ try:
+ f_out = self._open_compressed_file(_to_native_ascii(self.destination), 'wb')
+ with open(path, 'rb') as f_in:
+ shutil.copyfileobj(f_in, f_out)
+ f_out.close()
+ self.successes.append(path)
+ self.destination_state = STATE_COMPRESSED
+ except (IOError, OSError) as e:
+ self.module.fail_json(
+ path=_to_native(path),
+ dest=_to_native(self.destination),
+ msg='Unable to write to compressed file: %s' % _to_native(e), exception=format_exc()
+ )
+
+ def add_targets(self):
+ self.open()
+ try:
+ for target in self.targets:
+ if os.path.isdir(target):
+ for directory_path, directory_names, file_names in os.walk(target, topdown=True):
+ for directory_name in directory_names:
+ full_path = os.path.join(directory_path, directory_name)
+ self.add(full_path, strip_prefix(self.root, full_path))
+
+ for file_name in file_names:
+ full_path = os.path.join(directory_path, file_name)
+ self.add(full_path, strip_prefix(self.root, full_path))
+ else:
+ self.add(target, strip_prefix(self.root, target))
+ except Exception as e:
+ if self.format in ('zip', 'tar'):
+ archive_format = self.format
+ else:
+ archive_format = 'tar.' + self.format
+ self.module.fail_json(
+ msg='Error when writing %s archive at %s: %s' % (
+ archive_format, _to_native(self.destination), _to_native(e)
+ ),
+ exception=format_exc()
+ )
+ self.close()
+
+ if self.errors:
+ self.module.fail_json(
+ msg='Errors when writing archive at %s: %s' % (_to_native(self.destination), '; '.join(self.errors))
+ )
+
+ def is_different_from_original(self):
+ if self.original_checksums is None:
+ return self.original_size != self.destination_size()
+ else:
+ return self.original_checksums != self.destination_checksums()
+
+ def destination_checksums(self):
+ if self.destination_exists() and self.destination_readable():
+ return self._get_checksums(self.destination)
+ return None
+
+ def destination_exists(self):
+ return self.destination and os.path.exists(self.destination)
+
+ def destination_readable(self):
+ return self.destination and os.access(self.destination, os.R_OK)
+
+ def destination_size(self):
+ return os.path.getsize(self.destination) if self.destination_exists() else 0
+
+ def find_targets(self):
+ for path in self.paths:
+ if not os.path.lexists(path):
+ self.not_found.append(path)
+ else:
+ self.targets.append(path)
+
+ def has_targets(self):
+ return bool(self.targets)
+
+ def has_unfound_targets(self):
+ return bool(self.not_found)
+
+ def remove_single_target(self, path):
+ try:
+ os.remove(path)
+ except OSError as e:
+ self.module.fail_json(
+ path=_to_native(path),
+ msg='Unable to remove source file: %s' % _to_native(e), exception=format_exc()
+ )
+
+ def remove_targets(self):
+ for path in self.successes:
+ if os.path.exists(path):
+ try:
+ if os.path.isdir(path):
+ shutil.rmtree(path)
+ else:
+ os.remove(path)
+ except OSError:
+ self.errors.append(_to_native(path))
+ for path in self.paths:
+ try:
+ if os.path.isdir(path):
+ shutil.rmtree(path)
+ except OSError:
+ self.errors.append(_to_native(path))
+
+ if self.errors:
+ self.module.fail_json(
+ dest=_to_native(self.destination), msg='Error deleting some source files: ', files=self.errors
+ )
+
+ def update_permissions(self):
+ file_args = self.module.load_file_common_arguments(self.module.params, path=self.destination)
+ self.changed = self.module.set_fs_attributes_if_different(file_args, self.changed)
+
+ @property
+ def result(self):
+ return {
+ 'archived': [_to_native(p) for p in self.successes],
+ 'dest': _to_native(self.destination),
+ 'dest_state': self.destination_state,
+ 'changed': self.changed,
+ 'arcroot': _to_native(self.root),
+ 'missing': [_to_native(p) for p in self.not_found],
+ 'expanded_paths': [_to_native(p) for p in self.expanded_paths],
+ 'expanded_exclude_paths': [_to_native(p) for p in self.expanded_exclude_paths],
+ }
+
+ def _check_removal_safety(self):
+ for path in self.paths:
+ if os.path.isdir(path) and self.destination.startswith(os.path.join(path, b'')):
+ self.module.fail_json(
+ path=b', '.join(self.paths),
+ msg='Error, created archive can not be contained in source paths when remove=true'
+ )
+
+ def _open_compressed_file(self, path, mode):
+ f = None
+ if self.format == 'gz':
+ f = gzip.open(path, mode)
+ elif self.format == 'bz2':
+ f = bz2.BZ2File(path, mode)
+ elif self.format == 'xz':
+ f = lzma.LZMAFile(path, mode)
+ else:
+ self.module.fail_json(msg="%s is not a valid format" % self.format)
+
+ return f
+
+ @abc.abstractmethod
+ def close(self):
+ pass
+
+ @abc.abstractmethod
+ def contains(self, name):
+ pass
+
+ @abc.abstractmethod
+ def open(self):
+ pass
+
+ @abc.abstractmethod
+ def _add(self, path, archive_name):
+ pass
+
+ @abc.abstractmethod
+ def _get_checksums(self, path):
+ pass
+
+
+class ZipArchive(Archive):
+ def __init__(self, module):
+ super(ZipArchive, self).__init__(module)
+
+ def close(self):
+ self.file.close()
+
+ def contains(self, name):
+ try:
+ self.file.getinfo(name)
+ except KeyError:
+ return False
+ return True
+
+ def open(self):
+ self.file = zipfile.ZipFile(_to_native_ascii(self.destination), 'w', zipfile.ZIP_DEFLATED, True)
+
+ def _add(self, path, archive_name):
+ if not matches_exclusion_patterns(path, self.exclusion_patterns):
+ self.file.write(path, archive_name)
+
+ def _get_checksums(self, path):
+ try:
+ archive = zipfile.ZipFile(_to_native_ascii(path), 'r')
+ checksums = set((info.filename, info.CRC) for info in archive.infolist())
+ archive.close()
+ except BadZipFile:
+ checksums = set()
+ return checksums
+
+
+class TarArchive(Archive):
+ def __init__(self, module):
+ super(TarArchive, self).__init__(module)
+ self.fileIO = None
+
+ def close(self):
+ self.file.close()
+ if self.format == 'xz':
+ with lzma.open(_to_native(self.destination), 'wb') as f:
+ f.write(self.fileIO.getvalue())
+ self.fileIO.close()
+
+ def contains(self, name):
+ try:
+ self.file.getmember(name)
+ except KeyError:
+ return False
+ return True
+
+ def open(self):
+ if self.format in ('gz', 'bz2'):
+ self.file = tarfile.open(_to_native_ascii(self.destination), 'w|' + self.format)
+ # python3 tarfile module allows xz format but for python2 we have to create the tarfile
+ # in memory and then compress it with lzma.
+ elif self.format == 'xz':
+ self.fileIO = io.BytesIO()
+ self.file = tarfile.open(fileobj=self.fileIO, mode='w')
+ elif self.format == 'tar':
+ self.file = tarfile.open(_to_native_ascii(self.destination), 'w')
+ else:
+ self.module.fail_json(msg="%s is not a valid archive format" % self.format)
+
+ def _add(self, path, archive_name):
+ def py27_filter(tarinfo):
+ return None if matches_exclusion_patterns(tarinfo.name, self.exclusion_patterns) else tarinfo
+
+ def py26_filter(path):
+ return matches_exclusion_patterns(path, self.exclusion_patterns)
+
+ if PY27:
+ self.file.add(path, archive_name, recursive=False, filter=py27_filter)
+ else:
+ self.file.add(path, archive_name, recursive=False, exclude=py26_filter)
+
+ def _get_checksums(self, path):
+ if HAS_LZMA:
+ LZMAError = lzma.LZMAError
+ else:
+ # Just picking another exception that's also listed below
+ LZMAError = tarfile.ReadError
+ try:
+ if self.format == 'xz':
+ with lzma.open(_to_native_ascii(path), 'r') as f:
+ archive = tarfile.open(fileobj=f)
+ checksums = set((info.name, info.chksum) for info in archive.getmembers())
+ archive.close()
+ else:
+ archive = tarfile.open(_to_native_ascii(path), 'r|' + self.format)
+ checksums = set((info.name, info.chksum) for info in archive.getmembers())
+ archive.close()
+ except (LZMAError, tarfile.ReadError, tarfile.CompressionError):
+ try:
+ # The python implementations of gzip, bz2, and lzma do not support restoring compressed files
+ # to their original names so only file checksum is returned
+ f = self._open_compressed_file(_to_native_ascii(path), 'r')
+ checksum = 0
+ while True:
+ chunk = f.read(16 * 1024 * 1024)
+ if not chunk:
+ break
+ checksum = crc32(chunk, checksum)
+ checksums = set([(b'', checksum)])
+ f.close()
+ except Exception:
+ checksums = set()
+ return checksums
+
+
+def get_archive(module):
+ if module.params['format'] == 'zip':
+ return ZipArchive(module)
+ else:
+ return TarArchive(module)
+
+
+def main():
+ module = AnsibleModule(
+ argument_spec=dict(
+ path=dict(type='list', elements='path', required=True),
+ format=dict(type='str', default='gz', choices=['bz2', 'gz', 'tar', 'xz', 'zip']),
+ dest=dict(type='path'),
+ exclude_path=dict(type='list', elements='path', default=[]),
+ exclusion_patterns=dict(type='list', elements='path'),
+ force_archive=dict(type='bool', default=False),
+ remove=dict(type='bool', default=False),
+ ),
+ add_file_common_args=True,
+ supports_check_mode=True,
+ )
+
+ if not HAS_LZMA and module.params['format'] == 'xz':
+ module.fail_json(
+ msg=missing_required_lib("lzma or backports.lzma", reason="when using xz format"), exception=LZMA_IMP_ERR
+ )
+
+ check_mode = module.check_mode
+
+ archive = get_archive(module)
+ archive.find_targets()
+
+ if not archive.has_targets():
+ if archive.destination_exists():
+ archive.destination_state = STATE_ARCHIVED if is_archive(archive.destination) else STATE_COMPRESSED
+ elif archive.has_targets() and archive.must_archive:
+ if check_mode:
+ archive.changed = True
+ else:
+ archive.add_targets()
+ archive.destination_state = STATE_INCOMPLETE if archive.has_unfound_targets() else STATE_ARCHIVED
+ archive.changed |= archive.is_different_from_original()
+ if archive.remove:
+ archive.remove_targets()
+ else:
+ if check_mode:
+ if not archive.destination_exists():
+ archive.changed = True
+ else:
+ path = archive.paths[0]
+ archive.add_single_target(path)
+ archive.changed |= archive.is_different_from_original()
+ if archive.remove:
+ archive.remove_single_target(path)
+
+ if archive.destination_exists():
+ archive.update_permissions()
+
+ module.exit_json(**archive.result)
+
+
+if __name__ == '__main__':
+ main()