Initial commit of extras/archive module.
This manages compressed files or archives of many compressed files. You can maintain or update .gz, .bz2 compressed files, .zip archives, or tarballs compressed with gzip or bzip2. Possible use cases: * Back up user home directories * Ensure large text files are always compressed * Archive trees for distribution
This commit is contained in:
parent
f408a4db02
commit
2e1ce01dff
1 changed files with 285 additions and 0 deletions
285
lib/ansible/modules/extras/files/archive.py
Normal file
285
lib/ansible/modules/extras/files/archive.py
Normal file
|
@ -0,0 +1,285 @@
|
|||
#!/usr/bin/python
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
DOCUMENTATION = '''
|
||||
---
|
||||
module: archive
|
||||
version_added: 2.2
|
||||
short_description: Creates a compressed archive of one or more files or trees.
|
||||
extends_documentation_fragment: files
|
||||
description:
|
||||
- The M(archive) module packs an archive. It is the opposite of the unarchive module. By default, it assumes the compression source exists on the target. It will not copy the source file from the local system to the target before archiving - set copy=yes to pack an archive which does not already exist on the target. The source files are deleted after archiving.
|
||||
options:
|
||||
path:
|
||||
description:
|
||||
- Remote absolute path, glob, or list of paths or globs for the file or files to archive or compress.
|
||||
required: false
|
||||
default: null
|
||||
compression:
|
||||
description:
|
||||
- "The type of compression to use. Can be 'gz', 'bz2', or 'zip'.
|
||||
choices: [ 'gz', 'bz2', 'zip' ]
|
||||
creates:
|
||||
description:
|
||||
- The file name of the destination archive. When it already exists, this step will B(not) be run. This is required when 'path' refers to multiple files by either specifying a glob, a directory or multiple paths in a list.
|
||||
required: false
|
||||
default: null
|
||||
author: "Ben Doherty (@bendoh)"
|
||||
notes:
|
||||
- requires tarfile, zipfile, gzip, and bzip2 packages on target host
|
||||
- can product I(gzip), I(bzip2) and I(zip) compressed files or archives
|
||||
- removes source files by default
|
||||
'''
|
||||
|
||||
EXAMPLES = '''
|
||||
# Compress directory /path/to/foo/ into /path/to/foo.tgz
|
||||
- archive: path=/path/to/foo creates=/path/to/foo.tgz
|
||||
|
||||
# Compress regular file /path/to/foo into /path/to/foo.gz
|
||||
- archive: path=/path/to/foo
|
||||
|
||||
# Create a zip archive of /path/to/foo
|
||||
- archive: path=/path/to/foo compression=zip
|
||||
|
||||
# Create a bz2 archive of multiple files, rooted at /path
|
||||
- archive:
|
||||
path:
|
||||
- /path/to/foo
|
||||
- /path/wong/foo
|
||||
creates: /path/file.tar.bz2
|
||||
compression: bz2
|
||||
'''
|
||||
|
||||
import stat
|
||||
import os
|
||||
import errno
|
||||
import glob
|
||||
import shutil
|
||||
import gzip
|
||||
import bz2
|
||||
import zipfile
|
||||
import tarfile
|
||||
|
||||
def main():
|
||||
module = AnsibleModule(
|
||||
argument_spec = dict(
|
||||
path = dict(required=True),
|
||||
compression = dict(choices=['gz', 'bz2', 'zip'], default='gz', required=False),
|
||||
creates = dict(required=False),
|
||||
remove = dict(required=False, default=True, type='bool'),
|
||||
),
|
||||
add_file_common_args=True,
|
||||
supports_check_mode=True,
|
||||
)
|
||||
|
||||
params = module.params
|
||||
paths = params['path']
|
||||
creates = params['creates']
|
||||
remove = params['remove']
|
||||
expanded_paths = []
|
||||
compression = params['compression']
|
||||
globby = False
|
||||
changed = False
|
||||
state = 'absent'
|
||||
|
||||
# Simple or archive file compression (inapplicable with 'zip')
|
||||
archive = False
|
||||
successes = []
|
||||
|
||||
if isinstance(paths, basestring):
|
||||
paths = [paths]
|
||||
|
||||
for i, path in enumerate(paths):
|
||||
path = os.path.expanduser(params['path'])
|
||||
|
||||
# Detect glob-like characters
|
||||
if any((c in set('*?')) for c in path):
|
||||
expanded_paths = expanded_paths + glob.glob(path)
|
||||
else:
|
||||
expanded_paths.append(path)
|
||||
|
||||
if len(expanded_paths) == 0:
|
||||
module.fail_json(path, msg='Error, no source paths were found')
|
||||
|
||||
# If we actually matched multiple files or TRIED to, then
|
||||
# treat this as a multi-file archive
|
||||
archive = globby or len(expanded_paths) > 1 or any(os.path.isdir(path) for path in expanded_paths)
|
||||
|
||||
# Default created file name (for single-file archives) to
|
||||
# <file>.<compression>
|
||||
if not archive and not creates:
|
||||
creates = '%s.%s' % (expanded_paths[0], compression)
|
||||
|
||||
# Force archives to specify 'creates'
|
||||
if archive and not creates:
|
||||
module.fail_json(creates=creates, path=', '.join(paths), msg='Error, must specify "creates" when archiving multiple files or trees')
|
||||
|
||||
archive_paths = []
|
||||
missing = []
|
||||
arcroot = ''
|
||||
|
||||
for path in expanded_paths:
|
||||
# Use the longest common directory name among all the files
|
||||
# as the archive root path
|
||||
if arcroot == '':
|
||||
arcroot = os.path.dirname(path)
|
||||
else:
|
||||
for i in xrange(len(arcroot)):
|
||||
if path[i] != arcroot[i]:
|
||||
break
|
||||
|
||||
if i < len(arcroot):
|
||||
arcroot = os.path.dirname(arcroot[0:i+1])
|
||||
|
||||
if path == creates:
|
||||
# Don't allow the archive to specify itself! this is an error.
|
||||
module.fail_json(path=', '.join(paths), msg='Error, created archive would be included in archive')
|
||||
|
||||
if os.path.lexists(path):
|
||||
archive_paths.append(path)
|
||||
else:
|
||||
missing.append(path)
|
||||
|
||||
# No source files were found but the named archive exists: are we 'compress' or 'archive' now?
|
||||
if len(missing) == len(expanded_paths) and creates and os.path.exists(creates):
|
||||
# Just check the filename to know if it's an archive or simple compressed file
|
||||
if re.search(r'(\.tar\.gz|\.tgz|.tbz2|\.tar\.bz2|\.zip)$', os.path.basename(creates), re.IGNORECASE):
|
||||
state = 'archive'
|
||||
else:
|
||||
state = 'compress'
|
||||
|
||||
# Multiple files, or globbiness
|
||||
elif archive:
|
||||
if len(archive_paths) == 0:
|
||||
# No source files were found, but the archive is there.
|
||||
if os.path.lexists(creates):
|
||||
state = 'archive'
|
||||
elif len(missing) > 0:
|
||||
# SOME source files were found, but not all of them
|
||||
state = 'incomplete'
|
||||
|
||||
archive = None
|
||||
size = 0
|
||||
errors = []
|
||||
|
||||
if os.path.lexists(creates):
|
||||
size = os.path.getsize(creates)
|
||||
|
||||
if state != 'archive':
|
||||
try:
|
||||
if compression == 'gz' or compression == 'bz2':
|
||||
archive = tarfile.open(creates, 'w|' + compression)
|
||||
|
||||
for path in archive_paths:
|
||||
archive.add(path, path[len(arcroot):])
|
||||
successes.append(path)
|
||||
|
||||
elif compression == 'zip':
|
||||
archive = zipfile.ZipFile(creates, 'wb')
|
||||
|
||||
for path in archive_paths:
|
||||
archive.write(path, path[len(arcroot):])
|
||||
successes.append(path)
|
||||
|
||||
except OSError:
|
||||
e = get_exception()
|
||||
module.fail_json(msg='Error when writing zip archive at %s: %s' % (creates, str(e)))
|
||||
|
||||
if archive:
|
||||
archive.close()
|
||||
|
||||
if state != 'archive' and remove:
|
||||
for path in successes:
|
||||
try:
|
||||
if os.path.isdir(path):
|
||||
shutil.rmtree(path)
|
||||
else:
|
||||
os.remove(path)
|
||||
except OSError:
|
||||
e = get_exception()
|
||||
errors.append(path)
|
||||
|
||||
if len(errors) > 0:
|
||||
module.fail_json(creates=creates, msg='Error deleting some source files: ' + str(e), files=errors)
|
||||
|
||||
# Rudimentary check: If size changed then file changed. Not perfect, but easy.
|
||||
if os.path.getsize(creates) != size:
|
||||
changed = True
|
||||
|
||||
if len(successes) and state != 'incomplete':
|
||||
state = 'archive'
|
||||
|
||||
# Simple, single-file compression
|
||||
else:
|
||||
path = expanded_paths[0]
|
||||
|
||||
# No source or compressed file
|
||||
if not (os.path.exists(path) or os.path.lexists(creates)):
|
||||
state = 'absent'
|
||||
|
||||
# if it already exists and the source file isn't there, consider this done
|
||||
elif not os.path.lexists(path) and os.path.lexists(creates):
|
||||
state = 'compress'
|
||||
|
||||
else:
|
||||
if module.check_mode:
|
||||
if not os.path.exists(creates):
|
||||
changed = True
|
||||
else:
|
||||
size = 0
|
||||
f_in = f_out = archive = None
|
||||
|
||||
if os.path.lexists(creates):
|
||||
size = os.path.getsize(creates)
|
||||
|
||||
try:
|
||||
if compression == 'zip':
|
||||
archive = zipfile.ZipFile(creates, 'wb')
|
||||
archive.write(path, path[len(arcroot):])
|
||||
archive.close()
|
||||
state = 'archive' # because all zip files are archives
|
||||
|
||||
else:
|
||||
f_in = open(path, 'rb')
|
||||
|
||||
if compression == 'gz':
|
||||
f_out = gzip.open(creates, 'wb')
|
||||
elif compression == 'bz2':
|
||||
f_out = bz2.BZ2File(creates, 'wb')
|
||||
else:
|
||||
raise OSError("Invalid compression")
|
||||
|
||||
shutil.copyfileobj(f_in, f_out)
|
||||
|
||||
except OSError:
|
||||
e = get_exception()
|
||||
|
||||
module.fail_json(path=path, creates=creates, msg='Unable to write to compressed file: %s' % str(e))
|
||||
|
||||
if archive:
|
||||
archive.close()
|
||||
if f_in:
|
||||
f_in.close()
|
||||
if f_out:
|
||||
f_out.close()
|
||||
|
||||
# Rudimentary check: If size changed then file changed. Not perfect, but easy.
|
||||
if os.path.getsize(creates) != size:
|
||||
changed = True
|
||||
|
||||
state = 'compress'
|
||||
|
||||
if remove:
|
||||
try:
|
||||
os.remove(path)
|
||||
|
||||
except OSError:
|
||||
e = get_exception()
|
||||
module.fail_json(path=path, msg='Unable to remove source file: %s' % str(e))
|
||||
|
||||
module.exit_json(archived=successes, creates=creates, changed=changed, state=state, arcroot=arcroot, missing=missing, expanded_paths=expanded_paths)
|
||||
|
||||
# import module snippets
|
||||
from ansible.module_utils.basic import *
|
||||
if __name__ == '__main__':
|
||||
main()
|
Loading…
Reference in a new issue