#!/usr/bin/python # -*- coding: utf-8 -*- DOCUMENTATION = ''' --- module: archive version_added: 2.2 short_description: Creates a compressed archive of one or more files or trees. extends_documentation_fragment: files description: - The M(archive) module packs an archive. It is the opposite of the unarchive module. By default, it assumes the compression source exists on the target. It will not copy the source file from the local system to the target before archiving - set copy=yes to pack an archive which does not already exist on the target. The source files are deleted after archiving. options: path: description: - Remote absolute path, glob, or list of paths or globs for the file or files to archive or compress. required: false default: null compression: description: - The type of compression to use. Can be 'gz', 'bz2', or 'zip'. choices: [ 'gz', 'bz2', 'zip' ] creates: description: - The file name of the destination archive. When it already exists, this step will B(not) be run. This is required when 'path' refers to multiple files by either specifying a glob, a directory or multiple paths in a list. required: false default: null author: "Ben Doherty (@bendoh)" notes: - requires tarfile, zipfile, gzip, and bzip2 packages on target host - can product I(gzip), I(bzip2) and I(zip) compressed files or archives - removes source files by default ''' EXAMPLES = ''' # Compress directory /path/to/foo/ into /path/to/foo.tgz - archive: path=/path/to/foo creates=/path/to/foo.tgz # Compress regular file /path/to/foo into /path/to/foo.gz - archive: path=/path/to/foo # Create a zip archive of /path/to/foo - archive: path=/path/to/foo compression=zip # Create a bz2 archive of multiple files, rooted at /path - archive: path: - /path/to/foo - /path/wong/foo creates: /path/file.tar.bz2 compression: bz2 ''' import stat import os import errno import glob import shutil import gzip import bz2 import zipfile import tarfile def main(): module = AnsibleModule( argument_spec = dict( path = dict(required=True), compression = dict(choices=['gz', 'bz2', 'zip'], default='gz', required=False), creates = dict(required=False), remove = dict(required=False, default=True, type='bool'), ), add_file_common_args=True, supports_check_mode=True, ) params = module.params paths = params['path'] creates = params['creates'] remove = params['remove'] expanded_paths = [] compression = params['compression'] globby = False changed = False state = 'absent' # Simple or archive file compression (inapplicable with 'zip') archive = False successes = [] if isinstance(paths, basestring): paths = [paths] for i, path in enumerate(paths): path = os.path.expanduser(params['path']) # Detect glob-like characters if any((c in set('*?')) for c in path): expanded_paths = expanded_paths + glob.glob(path) else: expanded_paths.append(path) if len(expanded_paths) == 0: module.fail_json(path, msg='Error, no source paths were found') # If we actually matched multiple files or TRIED to, then # treat this as a multi-file archive archive = globby or len(expanded_paths) > 1 or any(os.path.isdir(path) for path in expanded_paths) # Default created file name (for single-file archives) to # . if not archive and not creates: creates = '%s.%s' % (expanded_paths[0], compression) # Force archives to specify 'creates' if archive and not creates: module.fail_json(creates=creates, path=', '.join(paths), msg='Error, must specify "creates" when archiving multiple files or trees') archive_paths = [] missing = [] arcroot = '' for path in expanded_paths: # Use the longest common directory name among all the files # as the archive root path if arcroot == '': arcroot = os.path.dirname(path) else: for i in xrange(len(arcroot)): if path[i] != arcroot[i]: break if i < len(arcroot): arcroot = os.path.dirname(arcroot[0:i+1]) if path == creates: # Don't allow the archive to specify itself! this is an error. module.fail_json(path=', '.join(paths), msg='Error, created archive would be included in archive') if os.path.lexists(path): archive_paths.append(path) else: missing.append(path) # No source files were found but the named archive exists: are we 'compress' or 'archive' now? if len(missing) == len(expanded_paths) and creates and os.path.exists(creates): # Just check the filename to know if it's an archive or simple compressed file if re.search(r'(\.tar\.gz|\.tgz|.tbz2|\.tar\.bz2|\.zip)$', os.path.basename(creates), re.IGNORECASE): state = 'archive' else: state = 'compress' # Multiple files, or globbiness elif archive: if len(archive_paths) == 0: # No source files were found, but the archive is there. if os.path.lexists(creates): state = 'archive' elif len(missing) > 0: # SOME source files were found, but not all of them state = 'incomplete' archive = None size = 0 errors = [] if os.path.lexists(creates): size = os.path.getsize(creates) if state != 'archive': try: if compression == 'gz' or compression == 'bz2': archive = tarfile.open(creates, 'w|' + compression) for path in archive_paths: archive.add(path, path[len(arcroot):]) successes.append(path) elif compression == 'zip': archive = zipfile.ZipFile(creates, 'w') for path in archive_paths: archive.write(path, path[len(arcroot):]) successes.append(path) except OSError: e = get_exception() module.fail_json(msg='Error when writing zip archive at %s: %s' % (creates, str(e))) if archive: archive.close() if state != 'archive' and remove: for path in successes: try: if os.path.isdir(path): shutil.rmtree(path) else: os.remove(path) except OSError: e = get_exception() errors.append(path) if len(errors) > 0: module.fail_json(creates=creates, msg='Error deleting some source files: ' + str(e), files=errors) # Rudimentary check: If size changed then file changed. Not perfect, but easy. if os.path.getsize(creates) != size: changed = True if len(successes) and state != 'incomplete': state = 'archive' # Simple, single-file compression else: path = expanded_paths[0] # No source or compressed file if not (os.path.exists(path) or os.path.lexists(creates)): state = 'absent' # if it already exists and the source file isn't there, consider this done elif not os.path.lexists(path) and os.path.lexists(creates): state = 'compress' else: if module.check_mode: if not os.path.exists(creates): changed = True else: size = 0 f_in = f_out = archive = None if os.path.lexists(creates): size = os.path.getsize(creates) try: if compression == 'zip': archive = zipfile.ZipFile(creates, 'wb') archive.write(path, path[len(arcroot):]) archive.close() state = 'archive' # because all zip files are archives else: f_in = open(path, 'rb') if compression == 'gz': f_out = gzip.open(creates, 'wb') elif compression == 'bz2': f_out = bz2.BZ2File(creates, 'wb') else: raise OSError("Invalid compression") shutil.copyfileobj(f_in, f_out) except OSError: e = get_exception() module.fail_json(path=path, creates=creates, msg='Unable to write to compressed file: %s' % str(e)) if archive: archive.close() if f_in: f_in.close() if f_out: f_out.close() # Rudimentary check: If size changed then file changed. Not perfect, but easy. if os.path.getsize(creates) != size: changed = True state = 'compress' if remove: try: os.remove(path) except OSError: e = get_exception() module.fail_json(path=path, msg='Unable to remove source file: %s' % str(e)) module.exit_json(archived=successes, creates=creates, changed=changed, state=state, arcroot=arcroot, missing=missing, expanded_paths=expanded_paths) # import module snippets from ansible.module_utils.basic import * if __name__ == '__main__': main()