Added functionality for md5 checksum together with a new overwrite param
to indicate that remote file should be overwritten if a file with the same filename exists.
This commit is contained in:
parent
2f944d87f6
commit
455fd612d3
1 changed files with 45 additions and 16 deletions
57
library/s3
57
library/s3
|
@ -45,9 +45,16 @@ options:
|
||||||
required: false
|
required: false
|
||||||
default: 600
|
default: 600
|
||||||
aliases: []
|
aliases: []
|
||||||
|
overwrite:
|
||||||
|
description:
|
||||||
|
- force overwrite if a file with the same name already exists, values true/false/yes/no. Does not support files uploaded to s3 with multipart upload.
|
||||||
|
required: false
|
||||||
|
default: false
|
||||||
examples:
|
examples:
|
||||||
- code: 's3 bucket=mybucket path=/path/to/file state=present'
|
- code: 's3 bucket=mybucket path=/path/to/file state=present'
|
||||||
description: "Simple playbook example"
|
description: "Simple playbook example"
|
||||||
|
- code: 's3 bucket=mybucket path=/path/to/file state=present overwrite=yes'
|
||||||
|
description: "Will overwrite only if remote and local checksums do not match. Does not support files uploaded to s3 with multipart upload."
|
||||||
requirements: [ "boto" ]
|
requirements: [ "boto" ]
|
||||||
author: Lester Wade
|
author: Lester Wade
|
||||||
'''
|
'''
|
||||||
|
@ -58,10 +65,21 @@ import urlparse
|
||||||
|
|
||||||
try:
|
try:
|
||||||
import boto
|
import boto
|
||||||
|
import hashlib
|
||||||
except ImportError:
|
except ImportError:
|
||||||
print "failed=True msg='boto required for this module'"
|
print "failed=True msg='boto required for this module'"
|
||||||
sys.exit(1)
|
sys.exit(1)
|
||||||
|
|
||||||
|
def upload_s3file(module, s3, bucket, key_name, path, expiry):
|
||||||
|
try:
|
||||||
|
key = bucket.new_key(key_name)
|
||||||
|
key.set_contents_from_filename(path)
|
||||||
|
url = key.generate_url(expiry)
|
||||||
|
module.exit_json(msg="Put operation complete", url=url, changed=True)
|
||||||
|
sys.exit(0)
|
||||||
|
except s3.provider.storage_copy_error, e:
|
||||||
|
module.fail_json(msg= str(e))
|
||||||
|
|
||||||
def main():
|
def main():
|
||||||
module = AnsibleModule(
|
module = AnsibleModule(
|
||||||
argument_spec = dict(
|
argument_spec = dict(
|
||||||
|
@ -72,6 +90,7 @@ def main():
|
||||||
s3_url = dict(aliases=['S3_URL']),
|
s3_url = dict(aliases=['S3_URL']),
|
||||||
ec2_secret_key = dict(aliases=['EC2_SECRET_KEY']),
|
ec2_secret_key = dict(aliases=['EC2_SECRET_KEY']),
|
||||||
ec2_access_key = dict(aliases=['EC2_ACCESS_KEY']),
|
ec2_access_key = dict(aliases=['EC2_ACCESS_KEY']),
|
||||||
|
overwrite = dict(default="false", choices=BOOLEANS),
|
||||||
),
|
),
|
||||||
required_together=[ ['bucket', 'path', 'state'] ],
|
required_together=[ ['bucket', 'path', 'state'] ],
|
||||||
)
|
)
|
||||||
|
@ -83,9 +102,9 @@ def main():
|
||||||
s3_url = module.params.get('s3_url')
|
s3_url = module.params.get('s3_url')
|
||||||
ec2_secret_key = module.params.get('ec2_secret_key')
|
ec2_secret_key = module.params.get('ec2_secret_key')
|
||||||
ec2_access_key = module.params.get('ec2_access_key')
|
ec2_access_key = module.params.get('ec2_access_key')
|
||||||
|
overwrite = module.boolean( module.params.get('overwrite') )
|
||||||
|
|
||||||
# allow eucarc environment variables to be used if ansible vars aren't set
|
# allow eucarc environment variables to be used if ansible vars aren't set
|
||||||
|
|
||||||
if not s3_url and 'S3_URL' in os.environ:
|
if not s3_url and 'S3_URL' in os.environ:
|
||||||
s3_url = os.environ['S3_URL']
|
s3_url = os.environ['S3_URL']
|
||||||
if not ec2_secret_key and 'EC2_SECRET_KEY' in os.environ:
|
if not ec2_secret_key and 'EC2_SECRET_KEY' in os.environ:
|
||||||
|
@ -144,13 +163,31 @@ def main():
|
||||||
except s3.provider.storage_response_error, e:
|
except s3.provider.storage_response_error, e:
|
||||||
module.fail_json(msg= str(e))
|
module.fail_json(msg= str(e))
|
||||||
|
|
||||||
|
if key_exists is True and overwrite is True:
|
||||||
|
# Retrieve MD5 Checksums.
|
||||||
|
md5_remote = key_check.etag[1:-1] # Strip Quotation marks from etag: https://code.google.com/p/boto/issues/detail?id=391
|
||||||
|
etag_multipart = md5_remote.find('-')!=-1 # Find out if this is a multipart upload -> etag is not md5: https://forums.aws.amazon.com/message.jspa?messageID=222158
|
||||||
|
if etag_multipart is True:
|
||||||
|
module.fail_json(msg="Files uploaded with multipart to s3 are not supported with checksum. They do not contain a valid md5 checksum, use overwrite=no instead.")
|
||||||
|
sys.exit(0)
|
||||||
|
md5_local = hashlib.md5(open(path, 'rb').read()).hexdigest()
|
||||||
|
md5_equal = md5_local == md5_remote
|
||||||
|
|
||||||
if state == 'present':
|
if state == 'present':
|
||||||
if bucket_exists is True and key_exists is True:
|
if bucket_exists is True and key_exists is True:
|
||||||
exists = True
|
if overwrite is False:
|
||||||
changed = False
|
exists = True
|
||||||
module.exit_json(msg="Bucket and key already exist", changed=changed)
|
changed = False
|
||||||
|
module.exit_json(msg="Bucket and key already exist", changed=changed)
|
||||||
|
if overwrite is True:
|
||||||
|
if md5_equal is True:
|
||||||
|
module.exit_json(msg="Remote and local file checksums identical.", changed=False)
|
||||||
|
if md5_equal is False:
|
||||||
|
upload_s3file(module, s3, bucket, key_name, path, expiry)
|
||||||
sys.exit(0)
|
sys.exit(0)
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
# If bucket exists, there cannot be a key within, lets create it ...
|
# If bucket exists, there cannot be a key within, lets create it ...
|
||||||
if state == 'present':
|
if state == 'present':
|
||||||
if bucket_exists is False:
|
if bucket_exists is False:
|
||||||
|
@ -162,18 +199,10 @@ def main():
|
||||||
except s3.provider.storage_create_error, e:
|
except s3.provider.storage_create_error, e:
|
||||||
module.fail_json(msg = str(e))
|
module.fail_json(msg = str(e))
|
||||||
|
|
||||||
# TO-DO, md5sum of key and local file to be confident that its valid.
|
# If bucket now exists but key doesn't or overwrite is True, create the key
|
||||||
# If bucket now exists but key doesn't, create the key
|
|
||||||
if state == 'present':
|
if state == 'present':
|
||||||
if bucket_exists is True and key_exists is False:
|
if bucket_exists is True and key_exists is False:
|
||||||
try:
|
upload_s3file(module, s3, bucket, key_name, path, expiry)
|
||||||
key = bucket.new_key(key_name)
|
|
||||||
key.set_contents_from_filename(path)
|
|
||||||
url = key.generate_url(expiry)
|
|
||||||
module.exit_json(msg="Put operation complete", url=url, changed=True)
|
|
||||||
sys.exit(0)
|
|
||||||
except s3.provider.storage_copy_error, e:
|
|
||||||
module.fail_json(msg= str(e))
|
|
||||||
|
|
||||||
# If state is absent and the bucket exists (doesn't matter about key since the bucket is the container), delete it.
|
# If state is absent and the bucket exists (doesn't matter about key since the bucket is the container), delete it.
|
||||||
if state == 'absent':
|
if state == 'absent':
|
||||||
|
|
Loading…
Reference in a new issue