From 8cd395f59518be9fa3a1ea9b345996233f6a3eab Mon Sep 17 00:00:00 2001 From: Abhijeet Kasurde Date: Wed, 14 Nov 2018 15:06:48 +0530 Subject: [PATCH] VMware: Refactor vmware_cluster (#37072) * Update cluster logic * Refactor code * Documentation update Signed-off-by: Abhijeet Kasurde --- lib/ansible/module_utils/vmware.py | 11 - .../modules/cloud/vmware/vmware_cluster.py | 483 ++++++++++++++---- 2 files changed, 384 insertions(+), 110 deletions(-) diff --git a/lib/ansible/module_utils/vmware.py b/lib/ansible/module_utils/vmware.py index 5fca2eaeccb..22f147be3eb 100644 --- a/lib/ansible/module_utils/vmware.py +++ b/lib/ansible/module_utils/vmware.py @@ -113,17 +113,6 @@ def find_dvspg_by_name(dv_switch, portgroup_name): return None -# Maintain for legacy, or remove with 2.1 ? -# Should be replaced with find_cluster_by_name -def find_cluster_by_name_datacenter(datacenter, cluster_name): - - host_folder = datacenter.hostFolder - for folder in host_folder.childEntity: - if folder.name == cluster_name: - return folder - return None - - def find_object_by_name(content, name, obj_type, folder=None, recurse=True): if not isinstance(obj_type, list): obj_type = [obj_type] diff --git a/lib/ansible/modules/cloud/vmware/vmware_cluster.py b/lib/ansible/modules/cloud/vmware/vmware_cluster.py index d0d2315d05b..937ba3da93d 100644 --- a/lib/ansible/modules/cloud/vmware/vmware_cluster.py +++ b/lib/ansible/modules/cloud/vmware/vmware_cluster.py @@ -2,60 +2,180 @@ # -*- coding: utf-8 -*- # Copyright: (c) 2015, Joseph Callen +# Copyright: (c) 2018, Ansible Project +# # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import absolute_import, division, print_function __metaclass__ = type -ANSIBLE_METADATA = {'metadata_version': '1.1', - 'status': ['preview'], - 'supported_by': 'community'} +ANSIBLE_METADATA = { + 'metadata_version': '1.1', + 'status': ['preview'], + 'supported_by': 'community' +} DOCUMENTATION = r''' --- module: vmware_cluster short_description: Manage VMware vSphere clusters description: - - Add or remove VMware vSphere clusters. + - This module can be used to add, remove and update VMware vSphere clusters and its configurations. + - Module can manage HA, DRS and VSAN related configurations. + - All values and VMware object names are case sensitive. version_added: '2.0' author: - Joseph Callen (@jcpowermac) +- Abhijeet Kasurde (@Akasurde) requirements: - - Tested on ESXi 5.5 - - PyVmomi installed + - Tested on ESXi 5.5 and 6.5. + - PyVmomi installed. options: cluster_name: - description: - - The name of the cluster that will be created. - required: yes - datacenter_name: - description: - - The name of the datacenter the cluster will be created in. - required: yes + description: + - The name of the cluster to be managed. + required: yes + datacenter: + description: + - The name of the datacenter. + required: yes + aliases: [ datacenter_name ] enable_drs: - description: - - If set to C(yes) will enable DRS when the cluster is created. - type: bool - default: 'no' + description: + - If set to C(yes), will enable DRS when the cluster is created. + type: bool + default: 'no' + drs_enable_vm_behavior_overrides: + description: + - Determines whether DRS Behavior overrides for individual virtual machines are enabled. + - If set to C(True), overrides C(drs_default_vm_behavior). + type: bool + default: True + version_added: 2.8 + drs_default_vm_behavior: + description: + - Specifies the cluster-wide default DRS behavior for virtual machines. + - If set to C(partiallyAutomated), then vCenter generate recommendations for virtual machine migration and + for the placement with a host. vCenter automatically implement placement at power on. + - If set to C(manual), then vCenter generate recommendations for virtual machine migration and + for the placement with a host. vCenter should not implement the recommendations automatically. + - If set to C(fullyAutomated), then vCenter should automate both the migration of virtual machines + and their placement with a host at power on. + default: fullyAutomated + choices: [ fullyAutomated, manual, partiallyAutomated ] + version_added: 2.8 + drs_vmotion_rate: + description: + - Threshold for generated ClusterRecommendations. + default: 3 + choices: [ 1, 2, 3, 4, 5 ] + version_added: 2.8 enable_ha: - description: - - If set to C(yes) will enable HA when the cluster is created. - type: bool - default: 'no' + description: + - If set to C(yes) will enable HA when the cluster is created. + type: bool + default: 'no' + ha_host_monitoring: + description: + - Indicates whether HA restarts virtual machines after a host fails. + - If set to C(enabled), HA restarts virtual machines after a host fails. + - If set to C(disabled), HA does not restart virtual machines after a host fails. + - If C(enable_ha) is set to C(no), then this value is ignored. + choices: [ 'enabled', 'disabled' ] + default: 'enabled' + version_added: 2.8 + ha_vm_monitoring: + description: + - Indicates the state of virtual machine health monitoring service. + - If set to C(vmAndAppMonitoring), HA response to both virtual machine and application heartbeat failure. + - If set to C(vmMonitoringDisabled), virtual machine health monitoring is disabled. + - If set to C(vmMonitoringOnly), HA response to virtual machine heartbeat failure. + - If C(enable_ha) is set to C(no), then this value is ignored. + choices: ['vmAndAppMonitoring', 'vmMonitoringOnly', 'vmMonitoringDisabled'] + default: 'vmMonitoringDisabled' + version_added: 2.8 + ha_failover_level: + description: + - Number of host failures that should be tolerated, still guaranteeing sufficient resources to + restart virtual machines on available hosts. + - Accepts integer values only. + default: 2 + version_added: 2.8 + ha_admission_control_enabled: + description: + - Determines if strict admission control is enabled. + - It is recommended to set this parameter to C(True), please refer documentation for more details. + default: True + type: bool + version_added: 2.8 + ha_vm_failure_interval: + description: + - The number of seconds after which virtual machine is declared as failed + if no heartbeat has been received. + - This setting is only valid if C(ha_vm_monitoring) is set to, either C(vmAndAppMonitoring) or C(vmMonitoringOnly). + - Unit is seconds. + default: 30 + version_added: 2.8 + ha_vm_min_up_time: + description: + - The number of seconds for the virtual machine's heartbeats to stabilize after + the virtual machine has been powered on. + - This setting is only valid if C(ha_vm_monitoring) is set to, either C(vmAndAppMonitoring) or C(vmMonitoringOnly). + - Unit is seconds. + default: 120 + version_added: 2.8 + ha_vm_max_failures: + description: + - Maximum number of failures and automated resets allowed during the time + that C(ha_vm_max_failure_window) specifies. + - This setting is only valid if C(ha_vm_monitoring) is set to, either C(vmAndAppMonitoring) or C(vmMonitoringOnly). + default: 3 + version_added: 2.8 + ha_vm_max_failure_window: + description: + - The number of seconds for the window during which up to C(ha_vm_max_failures) resets + can occur before automated responses stop. + - This setting is only valid if C(ha_vm_monitoring) is set to, either C(vmAndAppMonitoring) or C(vmMonitoringOnly). + - Unit is seconds. + - Default specifies no failure window. + default: -1 + version_added: 2.8 + ha_restart_priority: + description: + - Determines the preference that HA gives to a virtual machine if sufficient capacity is not available + to power on all failed virtual machines. + - This setting is only valid if C(ha_vm_monitoring) is set to, either C(vmAndAppMonitoring) or C(vmMonitoringOnly). + - If set to C(disabled), then HA is disabled for this virtual machine. + - If set to C(high), then virtual machine with this priority have a higher chance of powering on after a failure, + when there is insufficient capacity on hosts to meet all virtual machine needs. + - If set to C(medium), then virtual machine with this priority have an intermediate chance of powering on after a failure, + when there is insufficient capacity on hosts to meet all virtual machine needs. + - If set to C(low), then virtual machine with this priority have a lower chance of powering on after a failure, + when there is insufficient capacity on hosts to meet all virtual machine needs. + default: 'medium' + version_added: 2.8 + choices: [ 'disabled', 'high', 'low', 'medium' ] enable_vsan: - description: - - If set to C(yes) will enable vSAN when the cluster is created. - type: bool - default: 'no' + description: + - If set to C(yes) will enable vSAN when the cluster is created. + type: bool + default: 'no' + vsan_auto_claim_storage: + description: + - Determines whether the VSAN service is configured to automatically claim local storage + on VSAN-enabled hosts in the cluster. + type: bool + default: False + version_added: 2.8 state: - description: - - Create (C(present)) or remove (C(absent)) a VMware vSphere cluster. - choices: [absent, present] - default: present + description: + - Create C(present) or remove C(absent) a VMware vSphere cluster. + choices: [ absent, present ] + default: present extends_documentation_fragment: vmware.documentation ''' -EXAMPLES = r''' +EXAMPLES = r""" - name: Create Cluster vmware_cluster: hostname: '{{ vcenter_hostname }}' @@ -67,47 +187,73 @@ EXAMPLES = r''' enable_drs: yes enable_vsan: yes delegate_to: localhost -''' + +- name: Create Cluster with additional changes + vmware_cluster: + hostname: "{{ vcenter_server }}" + username: "{{ vcenter_user }}" + password: "{{ vcenter_pass }}" + validate_certs: no + datacenter_name: DC0 + cluster_name: "{{ cluster_name }}" + enable_ha: True + ha_vm_monitoring: vmMonitoringOnly + enable_drs: True + drs_default_vm_behavior: partiallyAutomated + enable_vsan: True + register: cl_result + delegate_to: localhost + +- name: Delete Cluster + vmware_cluster: + hostname: "{{ vcenter_server }}" + username: "{{ vcenter_user }}" + password: "{{ vcenter_pass }}" + datacenter_name: datacenter + cluster_name: cluster + enable_ha: yes + enable_drs: yes + enable_vsan: yes + state: absent +""" + +RETURN = r"""# +""" try: from pyVmomi import vim, vmodl - HAS_PYVMOMI = True except ImportError: - HAS_PYVMOMI = False + pass from ansible.module_utils.basic import AnsibleModule -from ansible.module_utils.vmware import (HAS_PYVMOMI, - TaskError, - connect_to_api, - find_cluster_by_name_datacenter, - find_datacenter_by_name, - vmware_argument_spec, - wait_for_task - ) +from ansible.module_utils.vmware import (PyVmomi, TaskError, find_datacenter_by_name, + vmware_argument_spec, wait_for_task) +from ansible.module_utils._text import to_native -class VMwareCluster(object): +class VMwareCluster(PyVmomi): def __init__(self, module): - self.module = module + super(VMwareCluster, self).__init__(module) self.cluster_name = module.params['cluster_name'] - self.datacenter_name = module.params['datacenter_name'] + self.datacenter_name = module.params['datacenter'] self.enable_drs = module.params['enable_drs'] self.enable_ha = module.params['enable_ha'] self.enable_vsan = module.params['enable_vsan'] self.desired_state = module.params['state'] self.datacenter = None self.cluster = None - self.content = connect_to_api(module) def process_state(self): + """ + Manage internal states of cluster + """ cluster_states = { 'absent': { 'present': self.state_destroy_cluster, 'absent': self.state_exit_unchanged, }, 'present': { - 'update': self.state_update_cluster, - 'present': self.state_exit_unchanged, + 'present': self.state_update_cluster, 'absent': self.state_create_cluster, } } @@ -117,27 +263,71 @@ class VMwareCluster(object): cluster_states[self.desired_state][current_state]() def configure_ha(self): + """ + Manage HA Configuration + Returns: Cluster DAS configuration spec + + """ das_config = vim.cluster.DasConfigInfo() das_config.enabled = self.enable_ha das_config.admissionControlPolicy = vim.cluster.FailoverLevelAdmissionControlPolicy() - das_config.admissionControlPolicy.failoverLevel = 2 + das_config.admissionControlPolicy.failoverLevel = self.params.get('ha_failover_level') + + ha_vm_monitoring = self.params.get('ha_vm_monitoring') + das_vm_config = None + if ha_vm_monitoring in ['vmMonitoringOnly', 'vmAndAppMonitoring']: + vm_tool_spec = vim.cluster.VmToolsMonitoringSettings() + vm_tool_spec.enabled = True + vm_tool_spec.vmMonitoring = ha_vm_monitoring + vm_tool_spec.failureInterval = self.params.get('ha_vm_failure_interval') + vm_tool_spec.minUpTime = self.params.get('ha_vm_min_up_time') + vm_tool_spec.maxFailures = self.params.get('ha_vm_max_failures') + vm_tool_spec.maxFailureWindow = self.params.get('ha_vm_max_failure_window') + + das_vm_config = vim.cluster.DasVmSettings() + das_vm_config.restartPriority = self.params.get('ha_restart_priority') + das_vm_config.isolationResponse = None + das_vm_config.vmToolsMonitoringSettings = vm_tool_spec + + das_config.admissionControlEnabled = self.params.get('ha_admission_control_enabled') + + das_config.hostMonitoring = self.params.get('ha_host_monitoring') + das_config.vmMonitoring = ha_vm_monitoring + das_config.defaultVmSettings = das_vm_config + return das_config def configure_drs(self): + """ + Manage DRS configuration + Returns: Cluster DRS configuration spec + + """ drs_config = vim.cluster.DrsConfigInfo() + drs_config.enabled = self.enable_drs - # Set to partially automated - drs_config.vmotionRate = 3 + drs_config.enableVmBehaviorOverrides = self.params.get('drs_enable_vm_behavior_overrides') + drs_config.defaultVmBehavior = self.params.get('drs_default_vm_behavior') + drs_config.vmotionRate = self.params.get('drs_vmotion_rate') + return drs_config def configure_vsan(self): + """ + Manage VSAN configuration + Returns: Cluster VSAN configuration spec + + """ vsan_config = vim.vsan.cluster.ConfigInfo() vsan_config.enabled = self.enable_vsan vsan_config.defaultConfig = vim.vsan.cluster.ConfigInfo.HostDefaultInfo() - vsan_config.defaultConfig.autoClaimStorage = False + vsan_config.defaultConfig.autoClaimStorage = self.params.get('vsan_auto_claim_storage') return vsan_config def state_create_cluster(self): + """ + Create cluster with given configuration + """ try: cluster_config_spec = vim.cluster.ConfigSpecEx() cluster_config_spec.dasConfig = self.configure_ha() @@ -148,23 +338,33 @@ class VMwareCluster(object): self.datacenter.hostFolder.CreateClusterEx(self.cluster_name, cluster_config_spec) self.module.exit_json(changed=True) except vim.fault.DuplicateName: - self.module.fail_json(msg="A cluster with the name %s already exists" % self.cluster_name) - except vmodl.fault.InvalidArgument: - self.module.fail_json(msg="Cluster configuration specification parameter is invalid") - except vim.fault.InvalidName: - self.module.fail_json(msg="%s is an invalid name for a cluster" % self.cluster_name) - except vmodl.fault.NotSupported: + # To match other vmware_* modules + pass + except vmodl.fault.InvalidArgument as invalid_args: + self.module.fail_json(msg="Cluster configuration specification" + " parameter is invalid : %s" % to_native(invalid_args.msg)) + except vim.fault.InvalidName as invalid_name: + self.module.fail_json(msg="'%s' is an invalid name for a" + " cluster : %s" % (self.cluster_name, + to_native(invalid_name.msg))) + except vmodl.fault.NotSupported as not_supported: # This should never happen - self.module.fail_json(msg="Trying to create a cluster on an incorrect folder object") + self.module.fail_json(msg="Trying to create a cluster on an incorrect" + " folder object : %s" % to_native(not_supported.msg)) except vmodl.RuntimeFault as runtime_fault: - self.module.fail_json(msg=runtime_fault.msg) + self.module.fail_json(msg=to_native(runtime_fault.msg)) except vmodl.MethodFault as method_fault: # This should never happen either - self.module.fail_json(msg=method_fault.msg) + self.module.fail_json(msg=to_native(method_fault.msg)) + except Exception as generic_exc: + self.module.fail_json(msg="Failed to create cluster" + " due to generic exception %s" % to_native(generic_exc)) def state_destroy_cluster(self): - changed = True - result = None + """ + Destroy cluster + """ + changed, result = False, None try: if not self.module.check_mode: @@ -172,79 +372,167 @@ class VMwareCluster(object): changed, result = wait_for_task(task) self.module.exit_json(changed=changed, result=result) except vim.fault.VimFault as vim_fault: - self.module.fail_json(msg=vim_fault.msg) + self.module.fail_json(msg=to_native(vim_fault.msg)) except vmodl.RuntimeFault as runtime_fault: - self.module.fail_json(msg=runtime_fault.msg) + self.module.fail_json(msg=to_native(runtime_fault.msg)) except vmodl.MethodFault as method_fault: - self.module.fail_json(msg=method_fault.msg) + self.module.fail_json(msg=to_native(method_fault.msg)) + except Exception as generic_exc: + self.module.fail_json(msg="Failed to destroy cluster" + " due to generic exception %s" % to_native(generic_exc)) def state_exit_unchanged(self): + """ + Exit without any change + """ self.module.exit_json(changed=False) def state_update_cluster(self): + """ + Update cluster configuration of existing cluster + """ + changed, result = False, None cluster_config_spec = vim.cluster.ConfigSpecEx() - changed = True - result = None - - if self.cluster.configurationEx.dasConfig.enabled != self.enable_ha: + diff = False # Triggers Reconfigure Task only when there is a change + if self.check_ha_config_diff(): cluster_config_spec.dasConfig = self.configure_ha() - if self.cluster.configurationEx.drsConfig.enabled != self.enable_drs: + diff = True + if self.check_drs_config_diff(): cluster_config_spec.drsConfig = self.configure_drs() - if self.cluster.configurationEx.vsanConfigInfo.enabled != self.enable_vsan: + diff = True + if self.check_vsan_config_diff(): cluster_config_spec.vsanConfig = self.configure_vsan() + diff = True try: - if not self.module.check_mode: + if not self.module.check_mode and diff: task = self.cluster.ReconfigureComputeResource_Task(cluster_config_spec, True) changed, result = wait_for_task(task) self.module.exit_json(changed=changed, result=result) except vmodl.RuntimeFault as runtime_fault: - self.module.fail_json(msg=runtime_fault.msg) + self.module.fail_json(msg=to_native(runtime_fault.msg)) except vmodl.MethodFault as method_fault: - self.module.fail_json(msg=method_fault.msg) + self.module.fail_json(msg=to_native(method_fault.msg)) except TaskError as task_e: - self.module.fail_json(msg=str(task_e)) + self.module.fail_json(msg=to_native(task_e)) + except Exception as generic_exc: + self.module.fail_json(msg="Failed to update cluster" + " due to generic exception %s" % to_native(generic_exc)) + + def check_ha_config_diff(self): + """ + Check HA configuration diff + Returns: True if there is diff, else False + + """ + das_config = self.cluster.configurationEx.dasConfig + if das_config.enabled != self.enable_ha or \ + das_config.admissionControlPolicy.failoverLevel != self.params.get('ha_failover_level') or \ + das_config.vmMonitoring != self.params.get('ha_vm_monitoring') or \ + das_config.hostMonitoring != self.params.get('ha_host_monitoring') or \ + das_config.admissionControlPolicy.failoverLevel != self.params.get('ha_failover_level') or \ + das_config.admissionControlEnabled != self.params.get('ha_admission_control_enabled') or \ + das_config.defaultVmSettings.restartPriority != self.params.get('ha_restart_priority') or \ + das_config.defaultVmSettings.vmToolsMonitoringSettings.vmMonitoring != self.params.get('ha_vm_monitoring') or \ + das_config.defaultVmSettings.vmToolsMonitoringSettings.failureInterval != self.params.get('ha_vm_failure_interval') or \ + das_config.defaultVmSettings.vmToolsMonitoringSettings.minUpTime != self.params.get('ha_vm_min_up_time') or \ + das_config.defaultVmSettings.vmToolsMonitoringSettings.maxFailures != self.params.get('ha_vm_max_failures') or \ + das_config.defaultVmSettings.vmToolsMonitoringSettings.maxFailureWindow != self.params.get('ha_vm_max_failure_window'): + return True + return False + + def check_drs_config_diff(self): + """ + Check DRS configuration diff + Returns: True if there is diff, else False + + """ + drs_config = self.cluster.configurationEx.drsConfig + + if drs_config.enabled != self.enable_drs or \ + drs_config.enableVmBehaviorOverrides != self.params.get('drs_enable_vm_behavior_overrides') or \ + drs_config.defaultVmBehavior != self.params.get('drs_default_vm_behavior') or \ + drs_config.vmotionRate != self.params.get('drs_vmotion_rate'): + return True + return False + + def check_vsan_config_diff(self): + """ + Check VSAN configuration diff + Returns: True if there is diff, else False + + """ + vsan_config = self.cluster.configurationEx.vsanConfigInfo + + if vsan_config.enabled != self.enable_vsan or \ + vsan_config.defaultConfig.autoClaimStorage != self.params.get('vsan_auto_claim_storage'): + return True + return False def check_cluster_configuration(self): + """ + Check cluster configuration + Returns: 'Present' if cluster exists, else 'absent' + + """ try: self.datacenter = find_datacenter_by_name(self.content, self.datacenter_name) if self.datacenter is None: - self.module.fail_json(msg="Datacenter %s does not exist, " - "please create first with Ansible Module vmware_datacenter or manually." - % self.datacenter_name) - self.cluster = find_cluster_by_name_datacenter(self.datacenter, self.cluster_name) + self.module.fail_json(msg="Datacenter %s does not exist." % self.datacenter_name) + self.cluster = self.find_cluster_by_name(cluster_name=self.cluster_name) if self.cluster is None: return 'absent' - else: - desired_state = (self.enable_ha, - self.enable_drs, - self.enable_vsan) - current_state = (self.cluster.configurationEx.dasConfig.enabled, - self.cluster.configurationEx.drsConfig.enabled, - self.cluster.configurationEx.vsanConfigInfo.enabled) - - if desired_state != current_state: - return 'update' - else: - return 'present' + return 'present' except vmodl.RuntimeFault as runtime_fault: - self.module.fail_json(msg=runtime_fault.msg) + self.module.fail_json(msg=to_native(runtime_fault.msg)) except vmodl.MethodFault as method_fault: - self.module.fail_json(msg=method_fault.msg) + self.module.fail_json(msg=to_native(method_fault.msg)) + except Exception as generic_exc: + self.module.fail_json(msg="Failed to check configuration" + " due to generic exception %s" % to_native(generic_exc)) def main(): - argument_spec = vmware_argument_spec() argument_spec.update(dict( cluster_name=dict(type='str', required=True), - datacenter_name=dict(type='str', required=True), + datacenter=dict(type='str', required=True, aliases=['datacenter_name']), + state=dict(type='str', + default='present', + choices=['absent', 'present']), + # DRS enable_drs=dict(type='bool', default=False), + drs_enable_vm_behavior_overrides=dict(type='bool', default=True), + drs_default_vm_behavior=dict(type='str', + choices=['fullyAutomated', 'manual', 'partiallyAutomated'], + default='fullyAutomated'), + drs_vmotion_rate=dict(type='int', + choices=range(1, 6), + default=3), + # HA enable_ha=dict(type='bool', default=False), + ha_failover_level=dict(type='int', default=2), + ha_host_monitoring=dict(type='str', + default='enabled', + choices=['enabled', 'disabled']), + # HA VM Monitoring related parameters + ha_vm_monitoring=dict(type='str', + choices=['vmAndAppMonitoring', 'vmMonitoringOnly', 'vmMonitoringDisabled'], + default='vmMonitoringDisabled'), + ha_vm_failure_interval=dict(type='int', default=30), + ha_vm_min_up_time=dict(type='int', default=120), + ha_vm_max_failures=dict(type='int', default=3), + ha_vm_max_failure_window=dict(type='int', default=-1), + + ha_restart_priority=dict(type='str', + choices=['high', 'low', 'medium', 'disabled'], + default='medium'), + ha_admission_control_enabled=dict(type='bool', default=True), + # VSAN enable_vsan=dict(type='bool', default=False), - state=dict(type='str', default='present', choices=['absent', 'present']), + vsan_auto_claim_storage=dict(type='bool', default=False), )) module = AnsibleModule( @@ -252,9 +540,6 @@ def main(): supports_check_mode=True, ) - if not HAS_PYVMOMI: - module.fail_json(msg='pyvmomi is required for this module') - vmware_cluster = VMwareCluster(module) vmware_cluster.process_state()