a87b2e38a0
Ceph Object Gateway (Ceph RGW) is an object storage interface built on top of librados to provide applications with a RESTful gateway to Ceph Storage Clusters: http://docs.ceph.com/docs/master/radosgw/ This patch adds the required bits to handle buckets with the RGW S3 RESTful API properly. It sticks to the AWS behaviour where possible while avoiding not yet implemented features in the Ceph RGW API. Signed-off-by: Javier M. Mellid <jmunhoz@igalia.com>
449 lines
14 KiB
Python
449 lines
14 KiB
Python
#!/usr/bin/python
|
|
#
|
|
# This is a free software: you can redistribute it and/or modify
|
|
# it under the terms of the GNU General Public License as published by
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
# (at your option) any later version.
|
|
#
|
|
# This Ansible library is distributed in the hope that it will be useful,
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
# GNU General Public License for more details.
|
|
#
|
|
# You should have received a copy of the GNU General Public License
|
|
# along with this library. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
DOCUMENTATION = '''
|
|
---
|
|
module: s3_bucket
|
|
short_description: Manage S3 buckets in AWS, Ceph, Walrus and FakeS3
|
|
description:
|
|
- Manage S3 buckets in AWS, Ceph, Walrus and FakeS3
|
|
version_added: "2.0"
|
|
author: "Rob White (@wimnat)"
|
|
options:
|
|
force:
|
|
description:
|
|
- When trying to delete a bucket, delete all keys in the bucket first (an s3 bucket must be empty for a successful deletion)
|
|
required: false
|
|
default: no
|
|
choices: [ 'yes', 'no' ]
|
|
name:
|
|
description:
|
|
- Name of the s3 bucket
|
|
required: true
|
|
default: null
|
|
policy:
|
|
description:
|
|
- The JSON policy as a string.
|
|
required: false
|
|
default: null
|
|
s3_url:
|
|
description:
|
|
- S3 URL endpoint for usage with Ceph, Eucalypus, fakes3, etc. Otherwise assumes AWS
|
|
default: null
|
|
aliases: [ S3_URL ]
|
|
ceph:
|
|
description:
|
|
- Enable API compatibility with Ceph. It takes into account the S3 API subset working with Ceph in order to provide the same module behaviour where possible.
|
|
version_added: "2.2"
|
|
requester_pays:
|
|
description:
|
|
- With Requester Pays buckets, the requester instead of the bucket owner pays the cost of the request and the data download from the bucket.
|
|
required: false
|
|
default: no
|
|
choices: [ 'yes', 'no' ]
|
|
state:
|
|
description:
|
|
- Create or remove the s3 bucket
|
|
required: false
|
|
default: present
|
|
choices: [ 'present', 'absent' ]
|
|
tags:
|
|
description:
|
|
- tags dict to apply to bucket
|
|
required: false
|
|
default: null
|
|
versioning:
|
|
description:
|
|
- Whether versioning is enabled or disabled (note that once versioning is enabled, it can only be suspended)
|
|
required: false
|
|
default: no
|
|
choices: [ 'yes', 'no' ]
|
|
extends_documentation_fragment:
|
|
- aws
|
|
- ec2
|
|
'''
|
|
|
|
EXAMPLES = '''
|
|
# Note: These examples do not set authentication details, see the AWS Guide for details.
|
|
|
|
# Create a simple s3 bucket
|
|
- s3_bucket:
|
|
name: mys3bucket
|
|
|
|
# Create a simple s3 bucket on Ceph Rados Gateway
|
|
- s3_bucket:
|
|
name: mys3bucket
|
|
s3_url: http://your-ceph-rados-gateway-server.xxx
|
|
ceph: true
|
|
|
|
# Remove an s3 bucket and any keys it contains
|
|
- s3_bucket:
|
|
name: mys3bucket
|
|
state: absent
|
|
force: yes
|
|
|
|
# Create a bucket, add a policy from a file, enable requester pays, enable versioning and tag
|
|
- s3_bucket:
|
|
name: mys3bucket
|
|
policy: "{{ lookup('file','policy.json') }}"
|
|
requester_pays: yes
|
|
versioning: yes
|
|
tags:
|
|
example: tag1
|
|
another: tag2
|
|
|
|
'''
|
|
|
|
import xml.etree.ElementTree as ET
|
|
import urlparse
|
|
|
|
try:
|
|
import boto.ec2
|
|
from boto.s3.connection import OrdinaryCallingFormat, Location
|
|
from boto.s3.tagging import Tags, TagSet
|
|
from boto.exception import BotoServerError, S3CreateError, S3ResponseError
|
|
HAS_BOTO = True
|
|
except ImportError:
|
|
HAS_BOTO = False
|
|
|
|
def get_request_payment_status(bucket):
|
|
|
|
response = bucket.get_request_payment()
|
|
root = ET.fromstring(response)
|
|
for message in root.findall('.//{http://s3.amazonaws.com/doc/2006-03-01/}Payer'):
|
|
payer = message.text
|
|
|
|
if payer == "BucketOwner":
|
|
return False
|
|
else:
|
|
return True
|
|
|
|
def create_tags_container(tags):
|
|
|
|
tag_set = TagSet()
|
|
tags_obj = Tags()
|
|
for key, val in tags.iteritems():
|
|
tag_set.add_tag(key, val)
|
|
|
|
tags_obj.add_tag_set(tag_set)
|
|
return tags_obj
|
|
|
|
def _create_bucket(connection, module, location):
|
|
|
|
policy = module.params.get("policy")
|
|
name = module.params.get("name")
|
|
requester_pays = module.params.get("requester_pays")
|
|
tags = module.params.get("tags")
|
|
versioning = module.params.get("versioning")
|
|
changed = False
|
|
|
|
try:
|
|
bucket = connection.get_bucket(name)
|
|
except S3ResponseError, e:
|
|
try:
|
|
bucket = connection.create_bucket(name, location=location)
|
|
changed = True
|
|
except S3CreateError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
# Versioning
|
|
versioning_status = bucket.get_versioning_status()
|
|
if not versioning_status and versioning:
|
|
try:
|
|
bucket.configure_versioning(versioning)
|
|
changed = True
|
|
versioning_status = bucket.get_versioning_status()
|
|
except S3ResponseError, e:
|
|
module.fail_json(msg=e.message)
|
|
elif not versioning_status and not versioning:
|
|
# do nothing
|
|
pass
|
|
else:
|
|
if versioning_status['Versioning'] == "Enabled" and not versioning:
|
|
bucket.configure_versioning(versioning)
|
|
changed = True
|
|
versioning_status = bucket.get_versioning_status()
|
|
elif ( (versioning_status['Versioning'] == "Disabled" and versioning) or (versioning_status['Versioning'] == "Suspended" and versioning) ):
|
|
bucket.configure_versioning(versioning)
|
|
changed = True
|
|
versioning_status = bucket.get_versioning_status()
|
|
|
|
# Requester pays
|
|
requester_pays_status = get_request_payment_status(bucket)
|
|
if requester_pays_status != requester_pays:
|
|
if requester_pays:
|
|
bucket.set_request_payment(payer='Requester')
|
|
changed = True
|
|
requester_pays_status = get_request_payment_status(bucket)
|
|
else:
|
|
bucket.set_request_payment(payer='BucketOwner')
|
|
changed = True
|
|
requester_pays_status = get_request_payment_status(bucket)
|
|
|
|
# Policy
|
|
try:
|
|
current_policy = bucket.get_policy()
|
|
except S3ResponseError, e:
|
|
if e.error_code == "NoSuchBucketPolicy":
|
|
current_policy = None
|
|
else:
|
|
module.fail_json(msg=e.message)
|
|
|
|
if current_policy is not None and policy is not None:
|
|
if policy is not None:
|
|
policy = json.dumps(policy)
|
|
|
|
if json.loads(current_policy) != json.loads(policy):
|
|
try:
|
|
bucket.set_policy(policy)
|
|
changed = True
|
|
current_policy = bucket.get_policy()
|
|
except S3ResponseError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
elif current_policy is None and policy is not None:
|
|
policy = json.dumps(policy)
|
|
|
|
try:
|
|
bucket.set_policy(policy)
|
|
changed = True
|
|
current_policy = bucket.get_policy()
|
|
except S3ResponseError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
elif current_policy is not None and policy is None:
|
|
try:
|
|
bucket.delete_policy()
|
|
changed = True
|
|
current_policy = bucket.get_policy()
|
|
except S3ResponseError, e:
|
|
if e.error_code == "NoSuchBucketPolicy":
|
|
current_policy = None
|
|
else:
|
|
module.fail_json(msg=e.message)
|
|
|
|
####
|
|
## Fix up json of policy so it's not escaped
|
|
####
|
|
|
|
# Tags
|
|
try:
|
|
current_tags = bucket.get_tags()
|
|
tag_set = TagSet()
|
|
except S3ResponseError, e:
|
|
if e.error_code == "NoSuchTagSet":
|
|
current_tags = None
|
|
else:
|
|
module.fail_json(msg=e.message)
|
|
|
|
if current_tags is not None or tags is not None:
|
|
|
|
if current_tags is None:
|
|
current_tags_dict = {}
|
|
else:
|
|
current_tags_dict = dict((t.key, t.value) for t in current_tags[0])
|
|
|
|
if current_tags_dict != tags:
|
|
try:
|
|
if tags:
|
|
bucket.set_tags(create_tags_container(tags))
|
|
else:
|
|
bucket.delete_tags()
|
|
current_tags_dict = tags
|
|
changed = True
|
|
except S3ResponseError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
module.exit_json(changed=changed, name=bucket.name, versioning=versioning_status, requester_pays=requester_pays_status, policy=current_policy, tags=current_tags_dict)
|
|
|
|
def _destroy_bucket(connection, module):
|
|
|
|
force = module.params.get("force")
|
|
name = module.params.get("name")
|
|
changed = False
|
|
|
|
try:
|
|
bucket = connection.get_bucket(name)
|
|
except S3ResponseError, e:
|
|
if e.error_code != "NoSuchBucket":
|
|
module.fail_json(msg=e.message)
|
|
else:
|
|
# Bucket already absent
|
|
module.exit_json(changed=changed)
|
|
|
|
if force:
|
|
try:
|
|
# Empty the bucket
|
|
for key in bucket.list():
|
|
key.delete()
|
|
|
|
except BotoServerError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
try:
|
|
bucket = connection.delete_bucket(name)
|
|
changed = True
|
|
except S3ResponseError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
module.exit_json(changed=changed)
|
|
|
|
def _create_bucket_ceph(connection, module, location):
|
|
|
|
name = module.params.get("name")
|
|
|
|
changed = False
|
|
|
|
try:
|
|
bucket = connection.get_bucket(name)
|
|
except S3ResponseError, e:
|
|
try:
|
|
bucket = connection.create_bucket(name, location=location)
|
|
changed = True
|
|
except S3CreateError, e:
|
|
module.fail_json(msg=e.message)
|
|
|
|
module.exit_json(changed=changed)
|
|
|
|
def _destroy_bucket_ceph(connection, module):
|
|
|
|
_destroy_bucket(connection, module)
|
|
|
|
def create_bucket(connection, module, location, flavour='aws'):
|
|
if flavour == 'ceph':
|
|
_create_bucket_ceph(connection, module, location)
|
|
else:
|
|
_create_bucket(connection, module, location)
|
|
|
|
def destroy_bucket(connection, module, flavour='aws'):
|
|
if flavour == 'ceph':
|
|
_destroy_bucket_ceph(connection, module)
|
|
else:
|
|
_destroy_bucket(connection, module)
|
|
|
|
def is_fakes3(s3_url):
|
|
""" Return True if s3_url has scheme fakes3:// """
|
|
if s3_url is not None:
|
|
return urlparse.urlparse(s3_url).scheme in ('fakes3', 'fakes3s')
|
|
else:
|
|
return False
|
|
|
|
def is_walrus(s3_url):
|
|
""" Return True if it's Walrus endpoint, not S3
|
|
|
|
We assume anything other than *.amazonaws.com is Walrus"""
|
|
if s3_url is not None:
|
|
o = urlparse.urlparse(s3_url)
|
|
return not o.hostname.endswith('amazonaws.com')
|
|
else:
|
|
return False
|
|
|
|
def main():
|
|
|
|
argument_spec = ec2_argument_spec()
|
|
argument_spec.update(
|
|
dict(
|
|
force = dict(required=False, default='no', type='bool'),
|
|
policy = dict(required=False, default=None),
|
|
name = dict(required=True),
|
|
requester_pays = dict(default='no', type='bool'),
|
|
s3_url = dict(aliases=['S3_URL']),
|
|
state = dict(default='present', choices=['present', 'absent']),
|
|
tags = dict(required=None, default={}, type='dict'),
|
|
versioning = dict(default='no', type='bool'),
|
|
ceph = dict(default='no', type='bool')
|
|
)
|
|
)
|
|
|
|
module = AnsibleModule(argument_spec=argument_spec)
|
|
|
|
if not HAS_BOTO:
|
|
module.fail_json(msg='boto required for this module')
|
|
|
|
region, ec2_url, aws_connect_params = get_aws_connection_info(module)
|
|
|
|
if region in ('us-east-1', '', None):
|
|
# S3ism for the US Standard region
|
|
location = Location.DEFAULT
|
|
else:
|
|
# Boto uses symbolic names for locations but region strings will
|
|
# actually work fine for everything except us-east-1 (US Standard)
|
|
location = region
|
|
|
|
s3_url = module.params.get('s3_url')
|
|
|
|
# allow eucarc environment variables to be used if ansible vars aren't set
|
|
if not s3_url and 'S3_URL' in os.environ:
|
|
s3_url = os.environ['S3_URL']
|
|
|
|
ceph = module.params.get('ceph')
|
|
|
|
if ceph and not s3_url:
|
|
module.fail_json(msg='ceph flavour requires s3_url')
|
|
|
|
flavour = 'aws'
|
|
|
|
# Look at s3_url and tweak connection settings
|
|
# if connecting to Walrus or fakes3
|
|
try:
|
|
if s3_url and ceph:
|
|
ceph = urlparse.urlparse(s3_url)
|
|
connection = boto.connect_s3(
|
|
host=ceph.hostname,
|
|
port=ceph.port,
|
|
is_secure=ceph.scheme == 'https',
|
|
calling_format=OrdinaryCallingFormat(),
|
|
**aws_connect_params
|
|
)
|
|
flavour = 'ceph'
|
|
elif is_fakes3(s3_url):
|
|
fakes3 = urlparse.urlparse(s3_url)
|
|
connection = S3Connection(
|
|
is_secure=fakes3.scheme == 'fakes3s',
|
|
host=fakes3.hostname,
|
|
port=fakes3.port,
|
|
calling_format=OrdinaryCallingFormat(),
|
|
**aws_connect_params
|
|
)
|
|
elif is_walrus(s3_url):
|
|
walrus = urlparse.urlparse(s3_url).hostname
|
|
connection = boto.connect_walrus(walrus, **aws_connect_params)
|
|
else:
|
|
connection = boto.s3.connect_to_region(location, is_secure=True, calling_format=OrdinaryCallingFormat(), **aws_connect_params)
|
|
# use this as fallback because connect_to_region seems to fail in boto + non 'classic' aws accounts in some cases
|
|
if connection is None:
|
|
connection = boto.connect_s3(**aws_connect_params)
|
|
|
|
except boto.exception.NoAuthHandlerFound, e:
|
|
module.fail_json(msg='No Authentication Handler found: %s ' % str(e))
|
|
except Exception, e:
|
|
module.fail_json(msg='Failed to connect to S3: %s' % str(e))
|
|
|
|
if connection is None: # this should never happen
|
|
module.fail_json(msg ='Unknown error, failed to create s3 connection, no information from boto.')
|
|
|
|
state = module.params.get("state")
|
|
|
|
if state == 'present':
|
|
create_bucket(connection, module, location, flavour=flavour)
|
|
elif state == 'absent':
|
|
destroy_bucket(connection, module, flavour=flavour)
|
|
|
|
from ansible.module_utils.basic import *
|
|
from ansible.module_utils.ec2 import *
|
|
import urlparse
|
|
|
|
if __name__ == '__main__':
|
|
main()
|