2013-03-19 17:07:36 +01:00
#!/usr/bin/python
2013-03-16 12:18:43 +01:00
# This file is part of Ansible
#
# Ansible is free software: you can redistribute it and/or modify
# it under the terms of the GNU General Public License as published by
# the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# Ansible is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
DOCUMENTATION = '''
- - -
module : s3
2014-09-09 01:02:08 +02:00
short_description : S3 module putting a file into S3 .
2013-03-16 12:18:43 +01:00
description :
2013-06-14 11:53:43 +02:00
- This module allows the user to dictate the presence of a given file in an S3 bucket . If or once the key ( file ) exists in the bucket , it returns a time - expired download URL . This module has a dependency on python - boto .
2013-03-16 12:18:43 +01:00
version_added : " 1.1 "
options :
bucket :
description :
2013-07-04 16:55:06 +02:00
- Bucket name .
2013-03-16 12:18:43 +01:00
required : true
default : null
aliases : [ ]
2013-07-04 16:55:06 +02:00
object :
2013-03-16 12:18:43 +01:00
description :
2013-07-04 16:55:06 +02:00
- Keyname of the object inside the bucket . Can be used to create " virtual directories " , see examples .
2013-03-16 12:18:43 +01:00
required : false
default : null
aliases : [ ]
2013-07-04 16:55:06 +02:00
version_added : " 1.3 "
src :
2013-05-02 21:27:12 +02:00
description :
2013-07-04 16:55:06 +02:00
- The source file path when performing a PUT operation .
2013-05-02 21:27:12 +02:00
required : false
default : null
aliases : [ ]
2013-07-04 16:55:06 +02:00
version_added : " 1.3 "
dest :
2013-03-16 12:18:43 +01:00
description :
2013-07-04 16:55:06 +02:00
- The destination file path when downloading an object / key with a GET operation .
2013-03-16 12:18:43 +01:00
required : false
aliases : [ ]
2013-07-04 16:55:06 +02:00
version_added : " 1.3 "
2013-04-02 19:19:11 +02:00
overwrite :
description :
2013-07-04 16:55:06 +02:00
- Force overwrite either locally on the filesystem or remotely with the object / key . Used with PUT and GET operations .
2013-04-02 19:19:11 +02:00
required : false
2013-10-09 18:56:56 +02:00
default : true
2013-04-05 21:18:31 +02:00
version_added : " 1.2 "
2013-07-04 16:55:06 +02:00
mode :
description :
2013-08-03 20:25:27 +02:00
- Switches the module behaviour between put ( upload ) , get ( download ) , geturl ( return download url ( Ansible 1.3 + ) , getstr ( download object as string ( 1.3 + ) ) , create ( bucket ) and delete ( bucket ) .
2013-07-04 16:55:06 +02:00
required : true
default : null
aliases : [ ]
2013-10-08 14:42:03 +02:00
expiration :
2013-07-04 16:55:06 +02:00
description :
- Time limit ( in seconds ) for the URL generated and returned by S3 / Walrus when performing a mode = put or mode = geturl operation .
required : false
2013-10-08 14:42:03 +02:00
default : 600
2013-07-04 16:55:06 +02:00
aliases : [ ]
2013-08-06 18:23:25 +02:00
s3_url :
description :
2014-12-01 21:16:22 +01:00
- " S3 URL endpoint for usage with Eucalypus, fakes3, etc. Otherwise assumes AWS "
2013-08-06 18:23:25 +02:00
default : null
aliases : [ S3_URL ]
2013-08-13 15:30:56 +02:00
aws_secret_key :
2013-08-06 18:23:25 +02:00
description :
2013-08-13 15:30:56 +02:00
- AWS secret key . If not set then the value of the AWS_SECRET_KEY environment variable is used .
2013-08-06 18:23:25 +02:00
required : false
default : null
2013-08-13 15:30:56 +02:00
aliases : [ ' ec2_secret_key ' , ' secret_key ' ]
aws_access_key :
2013-08-06 18:23:25 +02:00
description :
2013-08-13 15:30:56 +02:00
- AWS access key . If not set then the value of the AWS_ACCESS_KEY environment variable is used .
2013-08-06 18:23:25 +02:00
required : false
default : null
2013-08-13 15:30:56 +02:00
aliases : [ ' ec2_access_key ' , ' access_key ' ]
2013-11-01 15:23:01 +01:00
metadata :
description :
- Metadata for PUT operation , as a dictionary of ' key=value ' and ' key=value,key=value ' .
required : false
default : null
2014-03-11 19:48:39 +01:00
version_added : " 1.6 "
2014-09-09 01:02:08 +02:00
region :
description :
2014-09-09 22:53:33 +02:00
- " AWS region to create the bucket in. If not set then the value of the EC2_REGION and AWS_REGION environment variables are checked, followed by the aws_region and ec2_region settings in the Boto config file. If none of those are set the region defaults to the S3 Location: US Standard. Prior to ansible 1.8 this parameter could be specified but had no effect. "
2014-09-09 01:02:08 +02:00
required : false
default : null
version_added : " 1.8 "
2014-03-11 19:48:39 +01:00
2013-03-16 12:18:43 +01:00
requirements : [ " boto " ]
2013-05-02 21:27:12 +02:00
author : Lester Wade , Ralph Tice
2013-03-16 12:18:43 +01:00
'''
2013-04-26 04:22:48 +02:00
EXAMPLES = '''
# Simple PUT operation
2013-07-04 16:55:06 +02:00
- s3 : bucket = mybucket object = / my / desired / key . txt src = / usr / local / myfile . txt mode = put
2014-12-01 21:14:57 +01:00
2013-07-04 16:55:06 +02:00
# Simple GET operation
- s3 : bucket = mybucket object = / my / desired / key . txt dest = / usr / local / myfile . txt mode = get
2014-12-01 21:14:57 +01:00
2013-11-01 15:23:01 +01:00
# PUT/upload with metadata
2014-03-03 03:45:53 +01:00
- s3 : bucket = mybucket object = / my / desired / key . txt src = / usr / local / myfile . txt mode = put metadata = ' Content-Encoding=gzip,Cache-Control=no-cache '
2014-12-01 21:14:57 +01:00
2013-07-04 16:55:06 +02:00
# Create an empty bucket
- s3 : bucket = mybucket mode = create
2014-12-01 21:14:57 +01:00
# Create a bucket with key as directory, in the EU region
- s3 : bucket = mybucket object = / my / directory / path mode = create region = eu - west - 1
2013-07-04 16:55:06 +02:00
# Delete a bucket and all contents
- s3 : bucket = mybucket mode = delete
2013-04-26 04:22:48 +02:00
'''
2013-03-16 12:18:43 +01:00
import sys
import os
import urlparse
2013-07-04 16:55:06 +02:00
import hashlib
2013-03-16 12:18:43 +01:00
2015-03-03 02:12:54 +01:00
from boto . s3 . connection import OrdinaryCallingFormat
2013-03-16 12:18:43 +01:00
try :
import boto
2014-09-09 00:59:24 +02:00
from boto . s3 . connection import Location
2013-03-16 12:18:43 +01:00
except ImportError :
2014-08-20 19:02:56 +02:00
print " failed=True msg= ' boto required for this module ' "
sys . exit ( 1 )
2013-03-16 12:18:43 +01:00
2013-07-04 16:55:06 +02:00
def key_check ( module , s3 , bucket , obj ) :
try :
bucket = s3 . lookup ( bucket )
key_check = bucket . get_key ( obj )
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
if key_check :
return True
else :
return False
def keysum ( module , s3 , bucket , obj ) :
bucket = s3 . lookup ( bucket )
key_check = bucket . get_key ( obj )
2014-03-28 21:47:46 +01:00
if not key_check :
return None
md5_remote = key_check . etag [ 1 : - 1 ]
etag_multipart = ' - ' in md5_remote # Check for multipart, etag is not md5
if etag_multipart is True :
module . fail_json ( msg = " Files uploaded with multipart of s3 are not supported with checksum, unable to compute checksum. " )
2013-07-04 16:55:06 +02:00
return md5_remote
def bucket_check ( module , s3 , bucket ) :
try :
result = s3 . lookup ( bucket )
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
if result :
return True
else :
return False
2014-09-09 00:59:24 +02:00
def create_bucket ( module , s3 , bucket , location = Location . DEFAULT ) :
2013-07-04 16:55:06 +02:00
try :
2014-09-09 00:59:24 +02:00
bucket = s3 . create_bucket ( bucket , location = location )
2013-07-04 16:55:06 +02:00
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
if bucket :
return True
def delete_bucket ( module , s3 , bucket ) :
try :
bucket = s3 . lookup ( bucket )
bucket_contents = bucket . list ( )
bucket . delete_keys ( [ key . name for key in bucket_contents ] )
bucket . delete ( )
return True
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
def delete_key ( module , s3 , bucket , obj ) :
try :
bucket = s3 . lookup ( bucket )
bucket . delete_key ( obj )
module . exit_json ( msg = " Object deleted from bucket %s " % bucket , changed = True )
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
2013-07-31 12:51:57 +02:00
def create_dirkey ( module , s3 , bucket , obj ) :
2013-07-04 16:55:06 +02:00
try :
bucket = s3 . lookup ( bucket )
2013-07-31 12:51:57 +02:00
key = bucket . new_key ( obj )
key . set_contents_from_string ( ' ' )
module . exit_json ( msg = " Virtual directory %s created in bucket %s " % ( obj , bucket . name ) , changed = True )
2013-07-04 16:55:06 +02:00
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
def upload_file_check ( src ) :
if os . path . exists ( src ) :
file_exists is True
else :
file_exists is False
if os . path . isdir ( src ) :
module . fail_json ( msg = " Specifying a directory is not a valid source for upload. " , failed = True )
return file_exists
def path_check ( path ) :
if os . path . exists ( path ) :
return True
else :
return False
2013-11-01 15:23:01 +01:00
def upload_s3file ( module , s3 , bucket , obj , src , expiry , metadata ) :
2013-04-02 19:19:11 +02:00
try :
2013-07-04 16:55:06 +02:00
bucket = s3 . lookup ( bucket )
2013-11-01 15:23:01 +01:00
key = bucket . new_key ( obj )
if metadata :
for meta_key in metadata . keys ( ) :
key . set_metadata ( meta_key , metadata [ meta_key ] )
2013-07-04 16:55:06 +02:00
key . set_contents_from_filename ( src )
2013-04-02 19:19:11 +02:00
url = key . generate_url ( expiry )
2013-07-04 16:55:06 +02:00
module . exit_json ( msg = " PUT operation complete " , url = url , changed = True )
except s3 . provider . storage_copy_error , e :
module . fail_json ( msg = str ( e ) )
def download_s3file ( module , s3 , bucket , obj , dest ) :
try :
bucket = s3 . lookup ( bucket )
key = bucket . lookup ( obj )
key . get_contents_to_filename ( dest )
module . exit_json ( msg = " GET operation complete " , changed = True )
2013-04-02 19:19:11 +02:00
except s3 . provider . storage_copy_error , e :
module . fail_json ( msg = str ( e ) )
2013-07-23 03:03:00 +02:00
def download_s3str ( module , s3 , bucket , obj ) :
try :
bucket = s3 . lookup ( bucket )
key = bucket . lookup ( obj )
contents = key . get_contents_as_string ( )
module . exit_json ( msg = " GET operation complete " , contents = contents , changed = True )
except s3 . provider . storage_copy_error , e :
module . fail_json ( msg = str ( e ) )
2013-10-09 18:10:33 +02:00
def get_download_url ( module , s3 , bucket , obj , expiry , changed = True ) :
2013-07-04 16:55:06 +02:00
try :
bucket = s3 . lookup ( bucket )
key = bucket . lookup ( obj )
url = key . generate_url ( expiry )
2013-10-09 18:10:33 +02:00
module . exit_json ( msg = " Download url: " , url = url , expiry = expiry , changed = changed )
2013-07-04 16:55:06 +02:00
except s3 . provider . storage_response_error , e :
module . fail_json ( msg = str ( e ) )
2014-01-21 11:27:50 +01:00
def is_fakes3 ( s3_url ) :
""" Return True if s3_url has scheme fakes3:// """
if s3_url is not None :
return urlparse . urlparse ( s3_url ) . scheme == ' fakes3 '
else :
return False
2013-10-16 19:19:01 +02:00
def is_walrus ( s3_url ) :
""" Return True if it ' s Walrus endpoint, not S3
We assume anything other than * . amazonaws . com is Walrus """
if s3_url is not None :
o = urlparse . urlparse ( s3_url )
return not o . hostname . endswith ( ' amazonaws.com ' )
else :
return False
2013-03-16 12:18:43 +01:00
def main ( ) :
2014-03-12 01:19:03 +01:00
argument_spec = ec2_argument_spec ( )
2014-02-09 00:35:26 +01:00
argument_spec . update ( dict (
2013-07-04 16:55:06 +02:00
bucket = dict ( required = True ) ,
object = dict ( ) ,
src = dict ( ) ,
2013-10-02 09:57:54 +02:00
dest = dict ( default = None ) ,
2013-07-23 03:03:00 +02:00
mode = dict ( choices = [ ' get ' , ' put ' , ' delete ' , ' create ' , ' geturl ' , ' getstr ' ] , required = True ) ,
2013-06-01 18:09:04 +02:00
expiry = dict ( default = 600 , aliases = [ ' expiration ' ] ) ,
s3_url = dict ( aliases = [ ' S3_URL ' ] ) ,
2013-10-09 18:56:56 +02:00
overwrite = dict ( aliases = [ ' force ' ] , default = True , type = ' bool ' ) ,
2013-11-01 15:23:01 +01:00
metadata = dict ( type = ' dict ' ) ,
) ,
2013-03-16 12:18:43 +01:00
)
2014-02-09 00:35:26 +01:00
module = AnsibleModule ( argument_spec = argument_spec )
2013-03-16 12:18:43 +01:00
2013-07-04 16:55:06 +02:00
bucket = module . params . get ( ' bucket ' )
obj = module . params . get ( ' object ' )
src = module . params . get ( ' src ' )
2013-10-02 09:57:54 +02:00
if module . params . get ( ' dest ' ) :
dest = os . path . expanduser ( module . params . get ( ' dest ' ) )
2013-07-04 16:55:06 +02:00
mode = module . params . get ( ' mode ' )
2013-03-16 12:18:43 +01:00
expiry = int ( module . params [ ' expiry ' ] )
s3_url = module . params . get ( ' s3_url ' )
2013-06-01 18:09:04 +02:00
overwrite = module . params . get ( ' overwrite ' )
2013-11-01 15:23:01 +01:00
metadata = module . params . get ( ' metadata ' )
2013-11-12 20:13:05 +01:00
ec2_url , aws_access_key , aws_secret_key , region = get_ec2_creds ( module )
2014-09-09 00:59:24 +02:00
if region in ( ' us-east-1 ' , ' ' , None ) :
# S3ism for the US Standard region
location = Location . DEFAULT
else :
# Boto uses symbolic names for locations but region strings will
# actually work fine for everything except us-east-1 (US Standard)
location = region
2013-07-04 16:55:06 +02:00
if module . params . get ( ' object ' ) :
obj = os . path . expanduser ( module . params [ ' object ' ] )
2013-03-16 12:18:43 +01:00
# allow eucarc environment variables to be used if ansible vars aren't set
if not s3_url and ' S3_URL ' in os . environ :
s3_url = os . environ [ ' S3_URL ' ]
2013-08-13 15:30:56 +02:00
2014-01-21 11:27:50 +01:00
# Look at s3_url and tweak connection settings
# if connecting to Walrus or fakes3
if is_fakes3 ( s3_url ) :
try :
fakes3 = urlparse . urlparse ( s3_url )
s3 = boto . connect_s3 (
aws_access_key ,
aws_secret_key ,
is_secure = False ,
host = fakes3 . hostname ,
port = fakes3 . port ,
calling_format = OrdinaryCallingFormat ( ) )
except boto . exception . NoAuthHandlerFound , e :
module . fail_json ( msg = str ( e ) )
elif is_walrus ( s3_url ) :
2013-03-16 12:18:43 +01:00
try :
walrus = urlparse . urlparse ( s3_url ) . hostname
2013-08-13 15:30:56 +02:00
s3 = boto . connect_walrus ( walrus , aws_access_key , aws_secret_key )
2013-03-16 12:18:43 +01:00
except boto . exception . NoAuthHandlerFound , e :
module . fail_json ( msg = str ( e ) )
else :
try :
2015-02-25 14:46:13 +01:00
s3 = boto . s3 . connect_to_region ( location , aws_access_key_id = aws_access_key , aws_secret_access_key = aws_secret_key , is_secure = True , calling_format = OrdinaryCallingFormat ( ) )
2013-03-16 12:18:43 +01:00
except boto . exception . NoAuthHandlerFound , e :
module . fail_json ( msg = str ( e ) )
2015-03-03 02:12:54 +01:00
2013-07-04 16:55:06 +02:00
# If our mode is a GET operation (download), go through the procedure as appropriate ...
if mode == ' get ' :
2015-03-03 02:12:54 +01:00
2013-07-04 16:55:06 +02:00
# First, we check to see if the bucket exists, we get "bucket" returned.
bucketrtn = bucket_check ( module , s3 , bucket )
if bucketrtn is False :
module . fail_json ( msg = " Target bucket cannot be found " , failed = True )
2013-03-16 12:18:43 +01:00
2013-07-04 16:55:06 +02:00
# Next, we check to see if the key in the bucket exists. If it exists, it also returns key_matches md5sum check.
2015-03-03 02:12:54 +01:00
keyrtn = key_check ( module , s3 , bucket , obj )
2013-07-04 16:55:06 +02:00
if keyrtn is False :
module . fail_json ( msg = " Target key cannot be found " , failed = True )
2013-03-16 12:18:43 +01:00
2013-07-04 16:55:06 +02:00
# If the destination path doesn't exist, no need to md5um etag check, so just download.
pathrtn = path_check ( dest )
if pathrtn is False :
download_s3file ( module , s3 , bucket , obj , dest )
2013-05-02 21:27:12 +02:00
2013-07-04 16:55:06 +02:00
# Compare the remote MD5 sum of the object with the local dest md5sum, if it already exists.
if pathrtn is True :
md5_remote = keysum ( module , s3 , bucket , obj )
md5_local = hashlib . md5 ( open ( dest , ' rb ' ) . read ( ) ) . hexdigest ( )
if md5_local == md5_remote :
sum_matches = True
if overwrite is True :
download_s3file ( module , s3 , bucket , obj , dest )
else :
module . exit_json ( msg = " Local and remote object are identical, ignoring. Use overwrite parameter to force. " , changed = False )
2013-03-16 12:18:43 +01:00
else :
2013-07-04 16:55:06 +02:00
sum_matches = False
if overwrite is True :
download_s3file ( module , s3 , bucket , obj , dest )
else :
2015-03-11 16:59:38 +01:00
module . exit_json ( msg = " WARNING: Checksums do not match. Use overwrite parameter to force download. " )
2013-07-04 16:55:06 +02:00
# Firstly, if key_matches is TRUE and overwrite is not enabled, we EXIT with a helpful message.
if sum_matches is True and overwrite is False :
module . exit_json ( msg = " Local and remote object are identical, ignoring. Use overwrite parameter to force. " , changed = False )
# At this point explicitly define the overwrite condition.
if sum_matches is True and pathrtn is True and overwrite is True :
download_s3file ( module , s3 , bucket , obj , dest )
# If sum does not match but the destination exists, we
2015-03-11 16:59:38 +01:00
# if our mode is a PUT operation (upload), go through the procedure as appropriate ...
2013-07-04 16:55:06 +02:00
if mode == ' put ' :
2013-04-02 19:19:11 +02:00
2013-07-04 16:55:06 +02:00
# Use this snippet to debug through conditionals:
# module.exit_json(msg="Bucket return %s"%bucketrtn)
# sys.exit(0)
2013-04-02 19:19:11 +02:00
2013-07-04 16:55:06 +02:00
# Lets check the src path.
pathrtn = path_check ( src )
if pathrtn is False :
module . fail_json ( msg = " Local object for PUT does not exist " , failed = True )
2015-03-11 16:59:38 +01:00
2013-07-04 16:55:06 +02:00
# Lets check to see if bucket exists to get ground truth.
bucketrtn = bucket_check ( module , s3 , bucket )
2013-10-11 17:01:54 +02:00
if bucketrtn is True :
keyrtn = key_check ( module , s3 , bucket , obj )
2013-07-04 16:55:06 +02:00
# Lets check key state. Does it exist and if it does, compute the etag md5sum.
if bucketrtn is True and keyrtn is True :
md5_remote = keysum ( module , s3 , bucket , obj )
md5_local = hashlib . md5 ( open ( src , ' rb ' ) . read ( ) ) . hexdigest ( )
if md5_local == md5_remote :
sum_matches = True
if overwrite is True :
2013-11-01 15:23:01 +01:00
upload_s3file ( module , s3 , bucket , obj , src , expiry , metadata )
2013-07-04 16:55:06 +02:00
else :
2013-10-09 18:25:09 +02:00
get_download_url ( module , s3 , bucket , obj , expiry , changed = False )
2013-07-04 16:55:06 +02:00
else :
sum_matches = False
if overwrite is True :
2013-11-01 15:23:01 +01:00
upload_s3file ( module , s3 , bucket , obj , src , expiry , metadata )
2013-07-04 16:55:06 +02:00
else :
2015-03-11 16:59:38 +01:00
module . exit_json ( msg = " WARNING: Checksums do not match. Use overwrite parameter to force upload. " )
2014-09-09 00:59:24 +02:00
2013-07-04 16:55:06 +02:00
# If neither exist (based on bucket existence), we can create both.
2014-09-09 00:59:24 +02:00
if bucketrtn is False and pathrtn is True :
create_bucket ( module , s3 , bucket , location )
2013-11-01 15:23:01 +01:00
upload_s3file ( module , s3 , bucket , obj , src , expiry , metadata )
2013-03-16 12:18:43 +01:00
2013-07-04 16:55:06 +02:00
# If bucket exists but key doesn't, just upload.
if bucketrtn is True and pathrtn is True and keyrtn is False :
2013-11-01 15:23:01 +01:00
upload_s3file ( module , s3 , bucket , obj , src , expiry , metadata )
2014-09-09 00:59:24 +02:00
2013-07-04 16:55:06 +02:00
# Support for deleting an object if we have both params.
if mode == ' delete ' :
if bucket :
bucketrtn = bucket_check ( module , s3 , bucket )
if bucketrtn is True :
deletertn = delete_bucket ( module , s3 , bucket )
if deletertn is True :
module . exit_json ( msg = " Bucket %s and all keys have been deleted. " % bucket , changed = True )
else :
2013-10-07 14:59:03 +02:00
module . fail_json ( msg = " Bucket does not exist. " , changed = False )
2013-03-16 12:18:43 +01:00
else :
2013-07-04 16:55:06 +02:00
module . fail_json ( msg = " Bucket parameter is required. " , failed = True )
# Need to research how to create directories without "populating" a key, so this should just do bucket creation for now.
# WE SHOULD ENABLE SOME WAY OF CREATING AN EMPTY KEY TO CREATE "DIRECTORY" STRUCTURE, AWS CONSOLE DOES THIS.
if mode == ' create ' :
2013-07-31 12:51:57 +02:00
if bucket and not obj :
2013-07-04 16:55:06 +02:00
bucketrtn = bucket_check ( module , s3 , bucket )
if bucketrtn is True :
module . exit_json ( msg = " Bucket already exists. " , changed = False )
else :
2014-09-09 00:59:24 +02:00
module . exit_json ( msg = " Bucket created successfully " , changed = create_bucket ( module , s3 , bucket , location ) )
2013-07-04 16:55:06 +02:00
if bucket and obj :
2013-07-31 12:51:57 +02:00
bucketrtn = bucket_check ( module , s3 , bucket )
if obj . endswith ( ' / ' ) :
dirobj = obj
else :
dirobj = obj + " / "
if bucketrtn is True :
keyrtn = key_check ( module , s3 , bucket , dirobj )
if keyrtn is True :
module . exit_json ( msg = " Bucket %s and key %s already exists. " % ( bucket , obj ) , changed = False )
else :
create_dirkey ( module , s3 , bucket , dirobj )
if bucketrtn is False :
2014-09-09 00:59:24 +02:00
created = create_bucket ( module , s3 , bucket , location )
2013-07-31 12:51:57 +02:00
create_dirkey ( module , s3 , bucket , dirobj )
2013-07-04 16:55:06 +02:00
# Support for grabbing the time-expired URL for an object in S3/Walrus.
if mode == ' geturl ' :
if bucket and obj :
bucketrtn = bucket_check ( module , s3 , bucket )
if bucketrtn is False :
module . fail_json ( msg = " Bucket %s does not exist. " % bucket , failed = True )
else :
keyrtn = key_check ( module , s3 , bucket , obj )
if keyrtn is True :
get_download_url ( module , s3 , bucket , obj , expiry )
else :
module . fail_json ( msg = " Key %s does not exist. " % obj , failed = True )
else :
module . fail_json ( msg = " Bucket and Object parameters must be set " , failed = True )
2013-04-02 19:19:11 +02:00
2013-07-23 03:03:00 +02:00
if mode == ' getstr ' :
if bucket and obj :
bucketrtn = bucket_check ( module , s3 , bucket )
if bucketrtn is False :
module . fail_json ( msg = " Bucket %s does not exist. " % bucket , failed = True )
else :
keyrtn = key_check ( module , s3 , bucket , obj )
if keyrtn is True :
download_s3str ( module , s3 , bucket , obj )
else :
module . fail_json ( msg = " Key %s does not exist. " % obj , failed = True )
2013-10-11 17:28:47 +02:00
module . exit_json ( failed = False )
2013-11-12 20:13:05 +01:00
# import module snippets
from ansible . module_utils . basic import *
from ansible . module_utils . ec2 import *
2013-03-16 12:18:43 +01:00
main ( )