diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-18 05:52:22 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-18 05:52:22 +0000 |
commit | 38b7c80217c4e72b1d8988eb1e60bb6e77334114 (patch) | |
tree | 356e9fd3762877d07cde52d21e77070aeff7e789 /ansible_collections/infinidat/infinibox/plugins/modules | |
parent | Adding upstream version 7.7.0+dfsg. (diff) | |
download | ansible-38b7c80217c4e72b1d8988eb1e60bb6e77334114.tar.xz ansible-38b7c80217c4e72b1d8988eb1e60bb6e77334114.zip |
Adding upstream version 9.4.0+dfsg.upstream/9.4.0+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'ansible_collections/infinidat/infinibox/plugins/modules')
20 files changed, 4364 insertions, 866 deletions
diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_certificate.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_certificate.py new file mode 100644 index 000000000..bb32b48b1 --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_certificate.py @@ -0,0 +1,199 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +"""This module creates or modifies SSL certificates on Infinibox.""" + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_certificate +version_added: 2.16.0 +short_description: Create (present state) or clear (absent state) SSL certificates on Infinibox +description: + - This module uploads (present state) or clears (absent state) SSL certificates on Infinibox +author: David Ohlemacher (@ohlemacher) +options: + certificate_file_name: + description: + - Name with full path of a certificate file. + type: str + required: false + state: + description: + - Creates/Modifies the systems SSL certificate by uploading one from a file, when using state present. + - For state absent, the current certificate is removed and a new self-signed certificate is automatically generated by the IBOX. + - State stat shows the existing certificate's details. + type: str + required: false + default: present + choices: [ "stat", "present", "absent" ] +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Upload SSL certificate from file + infini_certificate: + certificate_file_name: cert.crt + state: present + user: admin + password: secret + system: ibox001 + +- name: State SSL certificate + infini_certificate: + state: stat + user: admin + password: secret + system: ibox001 + +- name: Clear SSL certificate + infini_certificate: + state: absent + user: admin + password: secret + system: ibox001 +""" + +# RETURN = r''' # ''' + +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + merge_two_dicts, + get_system, + infinibox_argument_spec, +) + +HAS_URLLIB3 = True +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + HAS_URLLIB3 = False + + +def handle_stat(module): + """ Handle the stat state parameter """ + certificate_file_name = module.params['certificate_file_name'] + path = "system/certificates" + system = get_system(module) + try: + cert_result = system.api.get(path=path).get_result()[0] + except APICommandFailed: + msg = f"Cannot stat SSL certificate {certificate_file_name}" + module.fail_json(msg=msg) + result = dict( + changed=False, + msg="SSL certificate stat {certificate_file_name} found" + ) + result = merge_two_dicts(result, cert_result) + module.exit_json(**result) + + +def handle_present(module): + """ Handle the present state parameter """ + certificate_file_name = module.params['certificate_file_name'] + path = "system/certificates" + system = get_system(module) + + with open(certificate_file_name, 'rb') as cert_file: + try: + try: + files = {'file': cert_file} + except FileNotFoundError: + module.fail_json(msg=f"Cannot find SSL certificate file named {certificate_file_name}") + except Exception as err: # pylint: disable=broad-exception-caught + module.fail_json(msg=f"Cannot open SSL certificate file named {certificate_file_name}: {err}") + cert_result = system.api.post(path=path, files=files).get_result() + except APICommandFailed as err: + msg = f"Cannot upload cert: {err}" + module.fail_json(msg=msg) + + cert_serial = cert_result['certificate']['serial_number'] + cert_issued_by_cn = cert_result['certificate']['issued_by']['CN'] + cert_issued_to_cn = cert_result['certificate']['issued_to']['CN'] + result = dict( + changed=True, + msg="System SSL certificate uploaded successfully. " + + f"Certificate S/N {cert_serial} issued by CN {cert_issued_by_cn} to CN {cert_issued_to_cn}" + ) + result = merge_two_dicts(result, cert_result) + module.exit_json(**result) + + +def handle_absent(module): + """ Handle the absent state parameter. Clear existing cert. IBOX will install self signed cert. """ + path = "system/certificates/generate_self_signed?approved=true" + system = get_system(module) + try: + cert_result = system.api.post(path=path).get_result() + except APICommandFailed as err: + msg = f"Cannot clear SSL certificate: {err}" + module.fail_json(msg=msg) + result = dict( + changed=True, + msg="System SSL certificate cleared and a self signed certificate was installed successfully" + ) + result = merge_two_dicts(result, cert_result) + module.exit_json(**result) + + +def execute_state(module): + """Handle states""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + elif state == "absent": + handle_absent(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + certificate_file_name = module.params["certificate_file_name"] + state = module.params["state"] + + if state in ["stat", "absent"]: + pass + if state in ["present"]: + if not certificate_file_name: + msg = "Certificate file name parameter must be provided" + module.fail_json(msg=msg) + + +def main(): + """ Main """ + argument_spec = infinibox_argument_spec() + argument_spec.update( + dict( + certificate_file_name=dict(required=False, default=None), + state=dict(default="present", choices=["stat", "present", "absent"]), + ) + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_URLLIB3: + module.fail_json(msg=missing_required_lib("urllib3")) + + check_options(module) + execute_state(module) + + +if __name__ == "__main__": + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_cluster.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_cluster.py index fe682cf3c..c1972efe2 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_cluster.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_cluster.py @@ -1,10 +1,15 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +""" A module for managing Infinibox clusters """ + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import (absolute_import, division, print_function) + __metaclass__ = type DOCUMENTATION = r''' @@ -51,12 +56,9 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - try: from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, - INFINISDK_IMPORT_ERROR, api_wrapper, infinibox_argument_spec, get_system, @@ -67,7 +69,6 @@ try: except ModuleNotFoundError: from infinibox import ( # Used when hacking HAS_INFINISDK, - INFINISDK_IMPORT_ERROR, api_wrapper, infinibox_argument_spec, get_system, @@ -76,12 +77,6 @@ except ModuleNotFoundError: merge_two_dicts, ) -try: - from infi.dtypes.iqn import make_iscsi_name - HAS_INFI_MOD = True -except ImportError: - HAS_INFI_MOD = False - @api_wrapper def get_host_by_name(system, host_name): @@ -98,67 +93,58 @@ def get_host_by_name(system, host_name): @api_wrapper def create_cluster(module, system): - # print("create cluster") - changed = True + """ Create a cluster """ + changed = False if not module.check_mode: cluster = system.host_clusters.create(name=module.params['name']) cluster_hosts = module.params['cluster_hosts'] - for cluster_host in cluster_hosts: - if cluster_host['host_cluster_state'] == 'present': - host = get_host_by_name(system, cluster_host['host_name']) - cluster.add_host(host) - # print("Added host {0} to cluster {1}".format(host.get_name, cluster.get_name())) - # else: - # print("Skipped adding (absent) host {0} to cluster {1}".format(host.get_name, cluster.get_name())) + if cluster_hosts: + for cluster_host in cluster_hosts: + if cluster_host['host_cluster_state'] == 'present': + host = get_host_by_name(system, cluster_host['host_name']) + cluster.add_host(host) + changed = True return changed @api_wrapper def update_cluster(module, system, cluster): - # print("update cluster") + """ Update a cluster """ changed = False # e.g. of one host dict found in the module.params['cluster_hosts'] list: # {host_name: <'some_name'>, host_cluster_state: <'present' or 'absent'>} module_cluster_hosts = module.params['cluster_hosts'] current_cluster_hosts_names = [host.get_name() for host in cluster.get_field('hosts')] - # print("current_cluster_hosts_names:", current_cluster_hosts_names) - for module_cluster_host in module_cluster_hosts: - module_cluster_host_name = module_cluster_host['host_name'] - # print("module_cluster_host_name:", module_cluster_host_name) - # Need to add host to cluster? - if module_cluster_host_name not in current_cluster_hosts_names: - if module_cluster_host['host_cluster_state'] == 'present': - host = get_host_by_name(system, module_cluster_host_name) - if not host: - msg = 'Cannot find host {0} to add to cluster {1}'.format( - module_cluster_host_name, - cluster.get_name(), - ) - module.fail_json(msg=msg) - cluster.add_host(host) - # print("Added host {0} to cluster {1}".format(host.get_name(), cluster.get_name())) - changed = True - # Need to remove host from cluster? - elif module_cluster_host_name in current_cluster_hosts_names: - if module_cluster_host['host_cluster_state'] == 'absent': - host = get_host_by_name(system, module_cluster_host_name) - if not host: - msg = 'Cannot find host {0} to add to cluster {1}'.format( - module_cluster_host_name, - cluster.get_name(), - ) - module.fail_json(msg=msg) - cluster.remove_host(host) - # print("Removed host {0} from cluster {1}".format(host.get_name(), cluster.get_name())) - changed = True + if module_cluster_hosts: + for module_cluster_host in module_cluster_hosts: + module_cluster_host_name = module_cluster_host['host_name'] + # Need to add host to cluster? + if module_cluster_host_name not in current_cluster_hosts_names: + if module_cluster_host['host_cluster_state'] == 'present': + host = get_host_by_name(system, module_cluster_host_name) + if not host: + msg = f'Cannot find host {module_cluster_host_name} to add to cluster {cluster.get_name()}' + module.fail_json(msg=msg) + cluster.add_host(host) + changed = True + # Need to remove host from cluster? + elif module_cluster_host_name in current_cluster_hosts_names: + if module_cluster_host['host_cluster_state'] == 'absent': + host = get_host_by_name(system, module_cluster_host_name) + if not host: + msg = f'Cannot find host {module_cluster_host_name} to add to cluster {cluster.get_name()}' + module.fail_json(msg=msg) + cluster.remove_host(host) + changed = True return changed @api_wrapper def delete_cluster(module, cluster): + """ Delete a cluster """ if not cluster: - msg = "Cluster {0} not found".format(cluster.get_name()) + msg = f"Cluster {cluster.get_name()} not found" module.fail_json(msg=msg) changed = True if not module.check_mode: @@ -166,13 +152,8 @@ def delete_cluster(module, cluster): return changed -def get_sys_cluster(module): - system = get_system(module) - cluster = get_cluster(module, system) - return (system, cluster) - - def get_cluster_fields(cluster): + """ Find fields for cluster """ fields = cluster.get_fields(from_cache=True, raw_value=True) created_at, created_at_timezone = unixMillisecondsToDate(fields.get('created_at', None)) field_dict = dict( @@ -192,10 +173,12 @@ def get_cluster_fields(cluster): def handle_stat(module): - system, cluster = get_sys_cluster(module) + """ Handle stat state """ + system = get_system(module) + cluster = get_cluster(module, system) cluster_name = module.params["name"] if not cluster: - module.fail_json(msg='Cluster {0} not found'.format(cluster_name)) + module.fail_json(msg=f'Cluster {cluster_name} not found') field_dict = get_cluster_fields(cluster) result = dict( changed=False, @@ -206,34 +189,39 @@ def handle_stat(module): def handle_present(module): - system, cluster = get_sys_cluster(module) + """ Handle present state """ + system = get_system(module) + cluster = get_cluster(module, system) cluster_name = module.params["name"] if not cluster: changed = create_cluster(module, system) - msg = 'Cluster {0} created'.format(cluster_name) + msg = f'Cluster {cluster_name} created' module.exit_json(changed=changed, msg=msg) else: changed = update_cluster(module, system, cluster) if changed: - msg = 'Cluster {0} updated'.format(cluster_name) + msg = f'Cluster {cluster_name} updated' else: - msg = 'Cluster {0} required no changes'.format(cluster_name) + msg = f'Cluster {cluster_name} required no changes' module.exit_json(changed=changed, msg=msg) def handle_absent(module): - system, cluster = get_sys_cluster(module) + """ Handle absent state """ + system = get_system(module) + cluster = get_cluster(module, system) cluster_name = module.params["name"] if not cluster: changed = False - msg = "Cluster {0} already absent".format(cluster_name) + msg = f"Cluster {cluster_name} already absent" else: changed = delete_cluster(module, cluster) - msg = "Cluster {0} removed".format(cluster_name) + msg = f"Cluster {cluster_name} removed" module.exit_json(changed=changed, msg=msg) def execute_state(module): + """ Handle states """ state = module.params['state'] try: if state == 'stat': @@ -243,36 +231,38 @@ def execute_state(module): elif state == 'absent': handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() def check_options(module): + """ Check module parameters for logic errors """ state = module.params['state'] if state == 'present': - if module.params['cluster_hosts'] is None: - module.fail_json(msg='Option cluster_hosts, a list, must be provided') - cluster_hosts = module.params['cluster_hosts'] - for host in cluster_hosts: - try: - # Check host has required keys - valid_keys = ['host_name', 'host_cluster_state'] - for valid_key in valid_keys: - not_used = host[valid_key] - # Check host has no unknown keys - if len(host.keys()) != len(valid_keys): - raise KeyError - except KeyError: - msg = 'With state present, all cluster_hosts ' \ - + 'require host_name and host_cluster_state key:values ' \ - + 'and no others' - module.fail_json(msg=msg) + if cluster_hosts: + for host in cluster_hosts: + try: + # Check host has required keys + valid_keys = ['host_name', 'host_cluster_state'] + for valid_key in valid_keys: + # _ = host[valid_key] + if valid_key not in host.keys(): + raise KeyError + # Check host has no unknown keys + if len(host.keys()) != len(valid_keys): + raise KeyError + except KeyError: + msg = 'With state present, all cluster_hosts ' \ + + 'require host_name and host_cluster_state key:values ' \ + + 'and no others' + module.fail_json(msg=msg) def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( @@ -284,9 +274,6 @@ def main(): module = AnsibleModule(argument_spec, supports_check_mode=True) - if not HAS_INFI_MOD: - module.fail_json(msg=missing_required_lib('infi.dtypes.iqn')) - if not HAS_INFINISDK: module.fail_json(msg=missing_required_lib('infinisdk')) diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_config.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_config.py new file mode 100644 index 000000000..881480008 --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_config.py @@ -0,0 +1,238 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +"""This module Modifies config on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_config +version_added: 2.13.0 +short_description: Modify config on Infinibox +description: + - This module modifies system config on Infinibox. +author: Wei Wang (@wwang) +options: + config_group: + description: + - Config group + type: str + required: true + choices: [ "core", "ip_config", "iscsi", "limits", "mgmt", "ndoe_interfaces", "overriders", "security", "ssh" ] + key: + description: + - Name of the config + type: str + required: true + value: + description: + - Value of the config key + type: str + required: false + state: + description: + - Query or modifies config when. + type: str + required: false + default: present + choices: [ "stat", "present" ] + +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Set compression setting to true + infini_config: + config_group: "mgmt" + key: "pool.compression_enabled_default" + value: false + state: "present" + user: "{{ user }}" + password: "{{ password }}" + system: "{{ system }}" +""" + +# RETURN = r''' # ''' + +# -*- coding: utf-8 -*- +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + HAS_INFINISDK, + api_wrapper, + infinibox_argument_spec, + get_system, +) + +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + pass # Handled by HAS_INFINISDK from module_utils + + +@api_wrapper +def get_config(module, disable_fail=False): + """ + Find and return config setting value + Use disable_fail when we are looking for config + and it may or may not exist and neither case is an error. + """ + system = get_system(module) + config_group = module.params["config_group"] + key = module.params["key"] + result = None + + path = f"config/{config_group}/{key}" + try: + api_response = system.api.get(path=path) + except APICommandFailed as err: + module.fail_json(msg=f"Cannot {config_group} key {key}: {err}") + + if api_response: + result = api_response.get_result() + good_status = api_response.response.status_code == 200 + if not disable_fail and not good_status: + msg = f"Configuration for {config_group} with key {key} failed" + module.fail_json(msg=msg) + elif disable_fail and not good_status: + return None + return result + + +def handle_stat(module): + """Return config stat""" + + config_group = module.params["config_group"] + key = module.params["key"] + value = get_config(module) + + result = { + "changed": False, + "object_type": config_group, + "key": key, + "value": value, + } + module.exit_json(**result) + + +@api_wrapper +def set_config(module): + """ + Find and return config setting value + Use disable_fail when we are looking for config + and it may or may not exist and neither case is an error. + """ + system = get_system(module) + config_group = module.params["config_group"] + key = module.params["key"] + value = module.params["value"] + + path = f"config/{config_group}/{key}" + + if value.lower() == "true": + data = True + elif value.lower() == "false": + data = False + + try: + system.api.put(path=path, data=data) + except APICommandFailed as err: + module.fail_json(msg=f"Cannot set config group {config_group} key {key} to value {value}: {err}") + # Variable 'changed' not returned by design + + +def handle_present(module): + """Make config present""" + changed = False + msg = "Config unchanged" + if not module.check_mode: + old_config = get_config(module, disable_fail=True) + set_config(module) + new_config = get_config(module) + changed = new_config != old_config + if changed: + msg = "Config changed" + else: + msg = "Config unchanged since the value is the same as the existing config" + module.exit_json(changed=changed, msg=msg) + + +def execute_state(module): + """Determine which state function to execute and do so""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + state = module.params["state"] + config_group = module.params["config_group"] + key = module.params["key"] + value = module.params["value"] + vtype = type(value) + + groups = [ + "core", + "ip_config", + "iscsi", + "limits", + "mgmt", + "ndoe_interfaces", + "overriders", + "security", + "ssh", + ] + + if state == "present" and key == "pool.compression_enabled_default": + if not isinstance(value, str): # isvalue.lower() not in values: + module.fail_json( + f"Value must be of type {type(str())}. Invalid value: {value} of {vtype}." + ) + if config_group not in groups: + module.fail_json( + f"Config_group must be one of {groups}" + ) + + +def main(): + """Main module function""" + argument_spec = infinibox_argument_spec() + + argument_spec.update( + { + "config_group": {"required": True, "choices": ["core", "ip_config", "iscsi", "limits", "mgmt", "ndoe_interfaces", "overriders", "security", "ssh"]}, + "key": {"required": True, "no_log": False}, + "value": {"required": False, "default": None}, + "state": {"required": False, "default": "present", "choices": ["stat", "present"]}, + } + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_INFINISDK: + module.fail_json(msg=missing_required_lib("infinisdk")) + + check_options(module) + execute_state(module) + + +if __name__ == '__main__': + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_event.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_event.py new file mode 100644 index 000000000..bc6cbdade --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_event.py @@ -0,0 +1,141 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +"""This module sends events to Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_event +version_added: 2.16.0 +short_description: Create custom events on Infinibox +description: + - This module creates events on Infinibox. +author: David Ohlemacher (@ohlemacher) +options: + description_template: + description: + - The content of the custom event + type: str + required: true + visibility: + description: + - The event's visibility + type: str + required: false + choices: + - CUSTOMER + - INFINIDAT + default: CUSTOMER + level: + description: + - The level of the custom event + type: str + required: true + choices: + - INFO + - WARNING + - ERROR + - CRITICAL + state: + description: + - Creates a custom event when present. Stat is not yet implemented. There is no way to remove events once posted, so abent is also not implemented. + type: str + required: false + default: present + choices: [ "present" ] + +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Create custom info event + infini_event: + description_template: Message content + level: INFO + state: present + user: admin + password: secret + system: ibox001 +""" + +# RETURN = r''' # ''' + +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + HAS_INFINISDK, + infinibox_argument_spec, + get_system, +) + + +def handle_stat(module): + """Handle stat state""" + msg = "handle_stat() is not implemented" + module.exit_json(msg=msg) + + +def handle_present(module): + """Handle present state""" + system = get_system(module) + description_template = module.params["description_template"] + level = module.params["level"] + visibility = module.params["visibility"] + + path = "events/custom" + json_data = { + "description_template": description_template, + "level": level, + "visibility": visibility, + } + system.api.post(path=path, data=json_data) + module.exit_json(changed=True, msg="Event posted") + + +def execute_state(module): + """Handle states""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + else: + module.exit_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def main(): + """ Main """ + argument_spec = infinibox_argument_spec() + argument_spec.update( + dict( + description_template=dict(required=True), + level=dict(required=True, choices=["INFO", "WARNING", "ERROR", "CRITICAL"]), + state=dict(required=False, default="present", choices=["present"]), + visibility=dict(default="CUSTOMER", required=False, choices=["CUSTOMER", "INFINIDAT"]), + ) + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_INFINISDK: + module.exit_json(msg=missing_required_lib("infinisdk")) + + execute_state(module) + + +if __name__ == "__main__": + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_export.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_export.py index f83e9b1f1..409c89924 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_export.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_export.py @@ -1,10 +1,15 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat(info@infinidat.com) +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +"""This module modifies exports on Infinibox.""" + +# Copyright: (c) 2024, Infinidat(info@infinidat.com) # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import (absolute_import, division, print_function) + __metaclass__ = type DOCUMENTATION = r''' @@ -82,13 +87,13 @@ EXAMPLES = r''' client_list: - client: 192.168.0.2 access: RW - no_root_squash: True + no_root_squash: true - client: 192.168.0.100 access: RO - no_root_squash: False + no_root_squash: false - client: 192.168.0.10-192.168.0.20 access: RO - no_root_squash: False + no_root_squash: false system: ibox001 user: admin password: secret @@ -98,8 +103,6 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, api_wrapper, @@ -110,23 +113,22 @@ from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox impo merge_two_dicts, ) -MUNCH_IMP_ERR = None +HAS_MUNCH = True try: from munch import unmunchify - HAS_MUNCH = True except ImportError: HAS_MUNCH = False - MUNCH_IMPORT_ERROR = traceback.format_exc() def transform(d): + """ Create a frozen set from a normal set's items """ return frozenset(d.items()) def create_export(module, export, filesystem, system): - """ Create new filesystem or update existing one""" + """ Create new export """ if export: - raise AssertionError("Export {0} already exists".format(export.get_name())) + raise AssertionError(f"Export {export.get_name()} already exists") changed = False name = module.params['name'] @@ -141,14 +143,13 @@ def create_export(module, export, filesystem, system): @api_wrapper -def update_export(module, export, filesystem, system): - """ Create new filesystem or update existing one""" +def update_export(module, export): + """ Update existing export """ if not export: - raise AssertionError("Export {0} does not exist and cannot be updated".format(export.get_name())) + raise AssertionError(f"Export {export.get_name()} does not exist and cannot be updated") changed = False - name = module.params['name'] client_list = module.params['client_list'] if client_list: @@ -164,21 +165,15 @@ def update_export(module, export, filesystem, system): @api_wrapper def delete_export(module, export): - """ Delete file system""" + """ Delete export """ if not module.check_mode: export.delete() changed = True return changed -def get_sys_exp_fs(module): - system = get_system(module) - filesystem = get_filesystem(module, system) - export = get_export(module, system) - return (system, export, filesystem) - - def get_export_fields(export): + """ Return export fields dict """ fields = export.get_fields() # from_cache=True, raw_value=True) export_id = fields.get('id', None) permissions = fields.get('permissions', None) @@ -192,15 +187,13 @@ def get_export_fields(export): def handle_stat(module): - """ - Gather stats on export and return. Changed is always False. - """ - system, export, filesystem = get_sys_exp_fs(module) + """ Gather stats on export and return. Changed is always False. """ + name = module.params['name'] + filesystem_name = module.params['filesystem'] + system = get_system(module) + export = get_export(module, system) if not export: - module.fail_json(msg='Export "{0}" of file system "{1}" not found'.format( - module.params['name'], - module.params['filesystem'], - )) + module.fail_json(msg=f"Export '{name}' of file system '{filesystem_name}' not found") field_dict = get_export_fields(export) result = dict( @@ -212,30 +205,38 @@ def handle_stat(module): def handle_present(module): - system, export, filesystem = get_sys_exp_fs(module) + """ Handle present state """ + system = get_system(module) + filesystem = get_filesystem(module, system) + export = get_export(module, system) + filesystem_name = module.params['filesystem'] if not filesystem: - module.fail_json(msg='File system {0} not found'.format(module.params['filesystem'])) + module.fail_json(msg=f'File system {filesystem_name} not found') elif not export: changed = create_export(module, export, filesystem, system) module.exit_json(changed=changed, msg="File system export created") else: - changed = update_export(module, export, filesystem, system) + changed = update_export(module, export) module.exit_json(changed=changed, msg="File system export updated") def handle_absent(module): - system, export, filesystem = get_sys_exp_fs(module) + """ Handle absent state """ + system = get_system(module) + export = get_export(module, system) + filesystem_name = module.params['filesystem'] if not export: changed = False - msg = "Export of {0} already absent".format(module.params['filesystem']) + msg = "Export of {filesystem_name} already absent" module.exit_json(changed=changed, msg=msg) else: changed = delete_export(module, export) - msg = "Export of {0} deleted".format(module.params['filesystem']) + msg = f"Export of {filesystem_name} deleted" module.exit_json(changed=changed, msg=msg) def execute_state(module): + """ Execute states """ state = module.params['state'] try: if state == 'stat': @@ -245,13 +246,14 @@ def execute_state(module): elif state == 'absent': handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_export_client.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_export_client.py index d35705787..d1889511f 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_export_client.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_export_client.py @@ -1,10 +1,15 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) +""" Manage Infinibox export clients """ + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position, wrong-import-order + from __future__ import (absolute_import, division, print_function) + __metaclass__ = type DOCUMENTATION = r''' @@ -58,7 +63,7 @@ EXAMPLES = r''' infini_export_client: client: 10.0.0.1 access_mode: RW - no_root_squash: yes + no_root_squash: true export: /data state: present # Default user: admin @@ -69,7 +74,7 @@ EXAMPLES = r''' infini_export_client: client: "{{ item }}" access_mode: RO - no_root_squash: no + no_root_squash: false export: /data user: admin password: secret @@ -94,7 +99,7 @@ from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox impo merge_two_dicts, ) -MUNCH_IMP_ERR = None +MUNCH_IMPORT_ERROR = None try: from munch import Munch, unmunchify HAS_MUNCH = True @@ -145,7 +150,7 @@ def update_client(module, export): @api_wrapper def delete_client(module, export): - """Update export client list""" + """delete export client from client list""" if export is None and module.params['state'] == 'absent': module.exit_json(changed=False) @@ -168,13 +173,8 @@ def delete_client(module, export): return changed -def get_sys_exp(module): - system = get_system(module) - export = get_export(module, system) - return (system, export) - - def get_export_client_fields(export, client_name): + """ Get export client fields """ fields = export.get_fields() # from_cache=True, raw_value=True) permissions = fields.get('permissions', None) for munched_perm in permissions: @@ -185,13 +185,15 @@ def get_export_client_fields(export, client_name): no_root_squash=perm['no_root_squash'], ) return field_dict - raise AssertionError("No client {0} match to exports found".format(client_name)) + raise AssertionError(f"No client {client_name} match to exports found") def handle_stat(module): - system, export = get_sys_exp(module) + """ Execute the stat state """ + system = get_system(module) + export = get_export(module, system) if not export: - module.fail_json(msg='Export {0} not found'.format(module.params['export'])) + module.fail_json(msg=f"Export {module.params['export']} not found") client_name = module.params['client'] field_dict = get_export_client_fields(export, client_name) result = dict( @@ -203,9 +205,11 @@ def handle_stat(module): def handle_present(module): - system, export = get_sys_exp(module) + """ Execute the present state """ + system = get_system(module) + export = get_export(module, system) if not export: - msg = 'Export {0} not found'.format(module.params['export']) + msg = f"Export {module.params['export']} not found" module.fail_json(msg=msg) changed = update_client(module, export) @@ -214,7 +218,9 @@ def handle_present(module): def handle_absent(module): - system, export = get_sys_exp(module) + """ Execute the absent state """ + system = get_system(module) + export = get_export(module, system) if not export: changed = False msg = "Export client already absent" @@ -226,6 +232,7 @@ def handle_absent(module): def execute_state(module): + """ Execute a state """ state = module.params['state'] try: if state == 'stat': @@ -235,13 +242,14 @@ def execute_state(module): elif state == 'absent': handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_fibre_channel_switch.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_fibre_channel_switch.py new file mode 100644 index 000000000..f64808af9 --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_fibre_channel_switch.py @@ -0,0 +1,180 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +""" Manage switch names on Infinibox """ + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_fibre_channel_switch +version_added: 2.16.0 +short_description: Manage Infinibox FC switch names +description: + - This module renames FC switch names (rename state) or shows information about FC switches (stat state) +author: David Ohlemacher (@ohlemacher) +options: + switch_name: + description: + - Current name of an existing fibre channel switch. + type: str + required: true + new_switch_name: + description: + - New name for an existing fibre channel switch. + type: str + required: false + state: + description: + - Rename an FC switch name, when using state rename. + - States present and absent are not implemented. + - State stat shows the existing FC switch details. + type: str + required: false + default: rename + choices: [ "stat", "rename" ] +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Rename fibre channel switch + infini_fibre_channel: + switch_name: VSAN 100 + state: rename + user: admin + password: secret + system: ibox001 + +- name: Get information about fibre channel switch + infini_fibre_channel: + switch_name: VSAN 2000 + state: stat + user: admin + password: secret + system: ibox001 +""" + +# RETURN = r''' # ''' + +from ansible.module_utils.basic import AnsibleModule + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + merge_two_dicts, + get_system, + infinibox_argument_spec, +) + +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + pass # Handled by HAS_INFINISDK from module_utils + + +def find_switch_by_name(module): + """ Find switch by name """ + switch = module.params['switch_name'] + path = f"fc/switches?name={switch}" + system = get_system(module) + try: + switch_result = system.api.get(path=path).get_result() + if not switch_result: + msg = f"Cannot find switch {switch}" + module.exit_json(msg=msg) + except APICommandFailed as err: + msg = f"Cannot find switch {switch}: {err}" + module.exit_json(msg=msg) + return switch_result[0] + + +def handle_stat(module): + """ Handle stat state """ + switch_name = module.params['switch_name'] + switch_result = find_switch_by_name(module) + result = dict( + changed=False, + msg=f"Switch stat {switch_name} found" + ) + result = merge_two_dicts(result, switch_result) + module.exit_json(**result) + + +def handle_rename(module): + """ Handle rename state """ + switch_name = module.params['switch_name'] + new_switch_name = module.params['new_switch_name'] + + switch_result = find_switch_by_name(module) + switch_id = switch_result['id'] + + path = f"fc/switches/{switch_id}" + data = { + "name": new_switch_name, + } + try: + system = get_system(module) + rename_result = system.api.put(path=path, data=data).get_result() + except APICommandFailed as err: + msg = f"Cannot rename fc switch {switch_name}: {err}" + module.exit_json(msg=msg) + + result = dict( + changed=True, + msg=f"FC switch renamed from {switch_name} to {new_switch_name}" + ) + result = merge_two_dicts(result, rename_result) + module.exit_json(**result) + + +def execute_state(module): + """Handle states""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "rename": + handle_rename(module) + else: + module.exit_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + new_switch_name = module.params["new_switch_name"] + state = module.params["state"] + + if state in ["rename"]: + if not new_switch_name: + msg = "New switch name parameter must be provided" + module.exit_json(msg=msg) + + +def main(): + """ Main """ + argument_spec = infinibox_argument_spec() + argument_spec.update( + dict( + switch_name=dict(required=True, type="str"), + new_switch_name=dict(required=False, type="str"), + state=dict(default="rename", choices=["stat", "rename"]), + ) + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + check_options(module) + execute_state(module) + + +if __name__ == "__main__": + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_fs.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_fs.py index f9cd2bd56..d4d75a07a 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_fs.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_fs.py @@ -1,27 +1,77 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +"""This module manages file systems on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) -from __future__ import (absolute_import, division, print_function) +from __future__ import absolute_import, division, print_function __metaclass__ = type -DOCUMENTATION = r''' +DOCUMENTATION = r""" --- module: infini_fs -version_added: '2.3.0' +version_added: 2.3.0 short_description: Create, Delete or Modify filesystems on Infinibox description: - This module creates, deletes or modifies filesystems on Infinibox. author: David Ohlemacher (@ohlemacher) options: + fs_type: + description: + - Specifies the file system type, regular or snapshot. + type: str + required: false + default: master + choices: [ "master", "snapshot" ] name: description: - File system name. + required: false + type: str + parent_fs_name: + description: + - Specify a fs name. This is the fs parent for creating a snapshot. Required if fs_type is snapshot. + type: str + required: false + pool: + description: + - Pool that will host file system. required: true type: str + restore_fs_from_snapshot: + description: + - Specify true to restore a file system (parent_fs_name) from an existing snapshot specified by the name field. + - State must be set to present and fs_type must be 'snapshot'. + type: bool + required: false + default: false + serial: + description: + - Serial number matching an existing file system. + required: false + type: str + size: + description: + - File system size in MB, GB or TB units. See examples. + required: false + type: str + snapshot_lock_expires_at: + description: + - This will cause a snapshot to be locked at the specified date-time. + Uses python's datetime format YYYY-mm-dd HH:MM:SS.ffffff, e.g. 2020-02-13 16:21:59.699700 + type: str + required: false + snapshot_lock_only: + description: + - This will lock an existing snapshot but will suppress refreshing the snapshot. + type: bool + required: false + default: false state: description: - Creates/Modifies file system when present or removes when absent. @@ -35,63 +85,96 @@ options: required: false default: true type: bool - pool: + write_protected: description: - - Pool that will host file system. - required: true + - Specifies if the file system should be write protected. Default will be True for snapshots, False for master file systems. type: str - size: - description: - - File system size in MB, GB or TB units. See examples. required: false - type: str + default: "Default" + choices: ["Default", "True", "False"] extends_documentation_fragment: - infinibox requirements: - capacity -''' +""" -EXAMPLES = r''' +EXAMPLES = r""" - name: Create new file system named foo under pool named bar infini_fs: name: foo - size: 1TB + size: 1GB pool: bar thin_provision: true state: present user: admin password: secret system: ibox001 -''' +- name: Create snapshot named foo_snap from fs named foo + infini_fs: + name: foo_snap + pool: bar + fs_type: snapshot + parent_fs_name: foo + state: present + user: admin + password: secret + system: ibox001 +- name: Stat snapshot, also a fs, named foo_snap + infini_fs: + name: foo_snap + pool: bar + state: present + user: admin + password: secret + system: ibox001 +- name: Remove snapshot, also a fs, named foo_snap + infini_fs: + name: foo_snap + state: absent + user: admin + password: secret + system: ibox001 +""" # RETURN = r''' # ''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - +HAS_INFINISDK = True try: from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( - HAS_INFINISDK, api_wrapper, - infinibox_argument_spec, + check_snapshot_lock_options, + get_filesystem, + get_fs_by_sn, get_pool, get_system, - get_filesystem + infinibox_argument_spec, + manage_snapshot_locks, ) except ModuleNotFoundError: from infinibox import ( # Used when hacking - HAS_INFINISDK, api_wrapper, - infinibox_argument_spec, + check_snapshot_lock_options, + get_filesystem, get_pool, get_system, - get_filesystem + infinibox_argument_spec, + manage_snapshot_locks, ) +except ImportError: + HAS_INFINISDK = False + +try: + from infinisdk.core.exceptions import APICommandFailed + from infinisdk.core.exceptions import ObjectNotFound +except ImportError: + HAS_INFINISDK = False CAPACITY_IMP_ERR = None try: from capacity import KiB, Capacity + HAS_CAPACITY = True except ImportError: HAS_CAPACITY = False @@ -99,152 +182,414 @@ except ImportError: @api_wrapper def create_filesystem(module, system): - """Create Filesystem""" - changed = True + """ Create Filesystem """ + changed = False if not module.check_mode: - if module.params['thin_provision']: - provisioning = 'THIN' + if module.params["thin_provision"]: + provisioning = "THIN" else: - provisioning = 'THICK' + provisioning = "THICK" + filesystem = system.filesystems.create( - name=module.params['name'], - pool=get_pool(module, system), + name=module.params["name"], provtype=provisioning, + pool=get_pool(module, system), ) - if module.params['size']: - size = Capacity(module.params['size']).roundup(64 * KiB) + if module.params["size"]: + size = Capacity(module.params["size"]).roundup(64 * KiB) filesystem.update_size(size) + + is_write_prot = filesystem.is_write_protected() + desired_is_write_prot = module.params["write_protected"] + if is_write_prot != desired_is_write_prot: + filesystem.update_field("write_protected", desired_is_write_prot) + changed = True return changed @api_wrapper def update_filesystem(module, filesystem): - """Update Filesystem""" + """ Update Filesystem """ changed = False - if module.params['size']: - size = Capacity(module.params['size']).roundup(64 * KiB) + + if module.check_mode: + return changed + + if module.params["size"]: + size = Capacity(module.params["size"]).roundup(64 * KiB) if filesystem.get_size() != size: - if not module.check_mode: - filesystem.update_size(size) + filesystem.update_size(size) changed = True + if module.params["thin_provision"] is not None: provisioning = str(filesystem.get_provisioning()) - if provisioning == 'THICK' and module.params['thin_provision']: - if not module.check_mode: - filesystem.update_provisioning('THIN') + if provisioning == "THICK" and module.params["thin_provision"]: + filesystem.update_provisioning("THIN") changed = True - if provisioning == 'THIN' and not module.params['thin_provision']: - if not module.check_mode: - filesystem.update_provisioning('THICK') + if provisioning == "THIN" and not module.params["thin_provision"]: + filesystem.update_provisioning("THICK") changed = True + + is_write_prot = filesystem.is_write_protected() + desired_is_write_prot = module.params["write_protected"] + if is_write_prot != desired_is_write_prot: + filesystem.update_field("write_protected", desired_is_write_prot) + changed = True + return changed @api_wrapper def delete_filesystem(module, filesystem): - """ Delete Filesystem""" + """ Delete Filesystem """ + changed = False if not module.check_mode: filesystem.delete() - module.exit_json(changed=True) + changed = True + return changed -def get_sys_pool_fs(module): - system = get_system(module) - pool = get_pool(module, system) - filesystem = get_filesystem(module, system) - return (system, pool, filesystem) +@api_wrapper +def create_fs_snapshot(module, system): + """ Create Snapshot from parent fs """ + snapshot_name = module.params["name"] + parent_fs_name = module.params["parent_fs_name"] + changed = False + if not module.check_mode: + try: + parent_fs = system.filesystems.get(name=parent_fs_name) + except ObjectNotFound: + msg = f"Cannot create snapshot {snapshot_name}. Parent file system {parent_fs_name} not found" + module.fail_json(msg=msg) + if not parent_fs: + msg = f"Cannot find new snapshot's parent file system named {parent_fs_name}" + module.fail_json(msg=msg) + if not module.check_mode: + if module.params["snapshot_lock_only"]: + msg = "Snapshot does not exist. Cannot comply with 'snapshot_lock_only: true'." + module.fail_json(msg=msg) + check_snapshot_lock_options(module) + snapshot = parent_fs.create_snapshot(name=snapshot_name) + + is_write_prot = snapshot.is_write_protected() + desired_is_write_prot = module.params["write_protected"] + if is_write_prot != desired_is_write_prot: + snapshot.update_field("write_protected", desired_is_write_prot) + + manage_snapshot_locks(module, snapshot) + changed = True + return changed + + +@api_wrapper +def update_fs_snapshot(module, snapshot): + """ Update/refresh fs snapshot. May also lock it. """ + refresh_changed = False + lock_changed = False + if not module.check_mode: + if not module.params["snapshot_lock_only"]: + snap_is_locked = snapshot.get_lock_state() == "LOCKED" + if not snap_is_locked: + if not module.check_mode: + snapshot.refresh_snapshot() + refresh_changed = True + else: + msg = "File system snapshot is locked and may not be refreshed" + module.fail_json(msg=msg) + + check_snapshot_lock_options(module) + lock_changed = manage_snapshot_locks(module, snapshot) + + if module.params["write_protected"] is not None: + is_write_prot = snapshot.is_write_protected() + desired_is_write_prot = module.params["write_protected"] + if is_write_prot != desired_is_write_prot: + snapshot.update_field("write_protected", desired_is_write_prot) + + return refresh_changed or lock_changed + + +@api_wrapper +def find_fs_id(module, system, fs_name): + """ Find the ID of this fs """ + fs_url = f"filesystems?name={fs_name}&fields=id" + fs = system.api.get(path=fs_url) + + result = fs.get_json()["result"] + if len(result) != 1: + module.fail_json(f"Cannot find a file ststem with name '{fs_name}'") + + fs_id = result[0]["id"] + return fs_id + + +@api_wrapper +def restore_fs_from_snapshot(module, system): + """ Use snapshot to restore a file system """ + changed = False + is_restoring = module.params["restore_fs_from_snapshot"] + fs_type = module.params["fs_type"] + snap_name = module.params["name"] + snap_id = find_fs_id(module, system, snap_name) + parent_fs_name = module.params["parent_fs_name"] + parent_fs_id = find_fs_id(module, system, parent_fs_name) + + # Check params + if not is_restoring: + raise AssertionError("A programming error occurred. is_restoring is not True") + if fs_type != "snapshot": + module.exit_json(msg="Cannot restore a parent file system from snapshot unless the file system type is 'snapshot'") + if not parent_fs_name: + module.exit_json(msg="Cannot restore a parent file system from snapshot unless the parent file system name is specified") + + if not module.check_mode: + restore_url = f"filesystems/{parent_fs_id}/restore?approved=true" + restore_data = { + "source_id": snap_id, + } + try: + system.api.post(path=restore_url, data=restore_data) + changed = True + except APICommandFailed as err: + module.fail_json(msg=f"Cannot restore file system {parent_fs_name} from snapshot {snap_name}: {str(err)}") + return changed def handle_stat(module): - system, pool, filesystem = get_sys_pool_fs(module) - if not pool: - module.fail_json(msg='Pool {0} not found'.format(module.params['pool'])) + """ Handle the stat state """ + system = get_system(module) + pool = get_pool(module, system) + if module.params["name"]: + filesystem = get_filesystem(module, system) + else: + filesystem = get_fs_by_sn(module, system) + fs_type = module.params["fs_type"] + + if fs_type == "master": + if not pool: + module.fail_json(msg=f"Pool {module.params['pool']} not found") if not filesystem: - module.fail_json(msg='File system {0} not found'.format(module.params['name'])) + module.fail_json(msg=f"File system {module.params['name']} not found") fields = filesystem.get_fields() # from_cache=True, raw_value=True) + + created_at = str(fields.get("created_at", None)) + filesystem_id = fields.get("id", None) + filesystem_type = fields.get("type", None) + has_children = fields.get("has_children", None) + lock_expires_at = str(filesystem.get_lock_expires_at()) + lock_state = filesystem.get_lock_state() + mapped = str(fields.get("mapped", None)) name = fields.get("name", None) - used = fields.get('used_size', None) - filesystem_id = fields.get('id', None) - provisioning = fields.get('provisioning', None) + parent_id = fields.get("parent_id", None) + provisioning = fields.get("provisioning", None) + serial = fields.get("serial", None) + size = str(filesystem.get_size()) + updated_at = str(fields.get("updated_at", None)) + used = str(fields.get("used_size", None)) + write_protected = fields.get("write_protected", None) + if filesystem_type == "SNAPSHOT": + msg = "File system snapshot stat found" + else: + msg = "File system stat found" result = dict( changed=False, - name=name, - size=str(filesystem.get_size()), - used=str(used), + created_at=created_at, filesystem_id=filesystem_id, + filesystem_type=filesystem_type, + has_children=has_children, + lock_state=lock_state, + lock_expires_at=lock_expires_at, + mapped=mapped, + msg=msg, + name=name, + parent_id=parent_id, provisioning=provisioning, - msg='File system stat found' + serial=serial, + size=size, + updated_at=updated_at, + used=used, + write_protected=write_protected, ) module.exit_json(**result) def handle_present(module): - system, pool, filesystem = get_sys_pool_fs(module) - if not pool: - module.fail_json(msg='Pool {0} not found'.format(module.params['pool'])) - if not filesystem: - changed = create_filesystem(module, system) - module.exit_json(changed=changed, msg="File system created") + """ Handle the present state """ + system = get_system(module) + pool = get_pool(module, system) + if module.params["name"]: + filesystem = get_filesystem(module, system) else: - changed = update_filesystem(module, filesystem) - module.exit_json(changed=changed, msg="File system updated") + filesystem = get_fs_by_sn(module, system) + fs_type = module.params["fs_type"] + is_restoring = module.params["restore_fs_from_snapshot"] + if fs_type == "master": + if not pool: + module.fail_json(msg=f"Pool {module.params['pool']} not found") + if not filesystem: + changed = create_filesystem(module, system) + module.exit_json(changed=changed, msg="File system created") + else: + changed = update_filesystem(module, filesystem) + module.exit_json(changed=changed, msg="File system updated") + elif fs_type == "snapshot": + snapshot = filesystem + if is_restoring: + # Restore fs from snapshot + changed = restore_fs_from_snapshot(module, system) + snap_fs_name = module.params["name"] + parent_fs_name = module.params["parent_fs_name"] + msg = f"File system {parent_fs_name} restored from snapshot {snap_fs_name}" + module.exit_json(changed=changed, msg=msg) + else: + if not snapshot: + changed = create_fs_snapshot(module, system) + module.exit_json(changed=changed, msg="File system snapshot created") + else: + changed = update_fs_snapshot(module, filesystem) + module.exit_json(changed=changed, msg="File system snapshot updated") def handle_absent(module): - system, pool, filesystem = get_sys_pool_fs(module) + """ Handle the absent state """ + system = get_system(module) + pool = get_pool(module, system) + if module.params["name"]: + filesystem = get_filesystem(module, system) + else: + filesystem = get_fs_by_sn(module, system) + + if filesystem and filesystem.get_lock_state() == "LOCKED": + msg = "Cannot delete snapshot. Locked." + module.fail_json(changed=False, msg=msg) + if not pool or not filesystem: module.exit_json(changed=False, msg="File system already absent") - else: + + existing_fs_type = filesystem.get_type() + + if existing_fs_type == "MASTER": changed = delete_filesystem(module, filesystem) module.exit_json(changed=changed, msg="File system removed") + elif existing_fs_type == "SNAPSHOT": + snapshot = filesystem + changed = delete_filesystem(module, snapshot) + module.exit_json(changed=changed, msg="Snapshot removed") + else: + module.fail_json(msg="A programming error has occured") def execute_state(module): - state = module.params['state'] + """ Execute states """ + state = module.params["state"] try: - if state == 'stat': + if state == "stat": handle_stat(module) - elif state == 'present': + elif state == "present": handle_present(module) - elif state == 'absent': + elif state == "absent": handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") finally: system = get_system(module) system.logout() +def check_options(module): + """Verify module options are sane""" + name = module.params["name"] + serial = module.params["serial"] + state = module.params["state"] + size = module.params["size"] + pool = module.params["pool"] + fs_type = module.params["fs_type"] + parent_fs_name = module.params["parent_fs_name"] + + if state == "stat": + if not name and not serial: + msg = "Name or serial parameter must be provided" + module.fail_json(msg=msg) + if state in ["present", "absent"]: + if not name: + msg = "Name parameter must be provided" + module.fail_json(msg=msg) + + if state == "present": + if fs_type == "master": + if parent_fs_name: + msg = "parent_fs_name should not be specified " + msg += "if fs_type is 'master'. Used for snapshots only." + module.fail_json(msg=msg) + if not size: + msg = "Size is required to create a master file system" + module.fail_json(msg=msg) + if not pool: + msg = "For state 'present', pool is required" + module.fail_json(msg=msg) + elif fs_type == "snapshot": + if size: + msg = "Size should not be specified " + msg += "for fs_type snapshot" + module.fail_json(msg=msg) + if not parent_fs_name: + msg = "For state 'present' and fs_type 'snapshot', " + msg += "parent_fs_name is required" + module.fail_json(msg=msg) + else: + msg = "A programming error has occurred" + module.fail_json(msg=msg) + + def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name=dict(required=True), - state=dict(default='present', choices=['stat', 'present', 'absent']), + fs_type=dict(choices=["master", "snapshot"], default="master"), + name=dict(required=False, default=None), + parent_fs_name=dict(default=None, required=False), pool=dict(required=True), + restore_fs_from_snapshot=dict(default=False, type="bool"), + serial=dict(required=False, default=None), size=dict(), - thin_provision=dict(type=bool, default=True), + snapshot_lock_expires_at=dict(), + snapshot_lock_only=dict(required=False, type="bool", default=False), + state=dict(default="present", choices=["stat", "present", "absent"]), + thin_provision=dict(default=True, type="bool"), + write_protected=dict(choices=["True", "False", "Default"], default="Default"), ) ) module = AnsibleModule(argument_spec, supports_check_mode=True) + if module.params["write_protected"] == "Default": + if module.params["fs_type"] == "master": # Use default for master fs + module.params["write_protected"] = False + else: # Use default for snapshot + module.params["write_protected"] = True + else: + module.params["write_protected"] = module.params["write_protected"] == "True" + if not HAS_INFINISDK: - module.fail_json(msg=missing_required_lib('infinisdk')) + module.fail_json(msg=missing_required_lib("infinisdk")) if not HAS_CAPACITY: - module.fail_json(msg=missing_required_lib('capacity')) + module.fail_json(msg=missing_required_lib("capacity")) - if module.params['size']: + if module.params["size"]: try: - Capacity(module.params['size']) - except Exception: - module.fail_json(msg='size (Physical Capacity) should be defined in MB, GB, TB or PB units') + Capacity(module.params["size"]) + except Exception: # pylint: disable=broad-exception-caught + module.fail_json( + msg="size (Physical Capacity) should be defined in MB, GB, TB or PB units" + ) + check_options(module) execute_state(module) -if __name__ == '__main__': +if __name__ == "__main__": main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_host.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_host.py index 68d78546e..91eeab2ee 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_host.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_host.py @@ -1,7 +1,11 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-list-literal,use-dict-literal,line-too-long,wrong-import-position,multiple-statements + +""" Manage hosts on Infinibox """ + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import (absolute_import, division, print_function) @@ -11,7 +15,7 @@ __metaclass__ = type DOCUMENTATION = r''' --- module: infini_host -version_added: '2.3.0' +version_added: 2.3.0 short_description: Create, Delete or Modify Hosts on Infinibox description: - This module creates, deletes or modifies hosts on Infinibox. @@ -20,10 +24,12 @@ options: name: description: - Host Name + type: str required: true state: description: - Creates/Modifies Host when present or removes when absent + type: str required: false default: present choices: [ "stat", "present", "absent" ] @@ -44,9 +50,6 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - -from infi.dtypes.iqn import make_iscsi_name from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, api_wrapper, @@ -60,22 +63,16 @@ from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox impo @api_wrapper def create_host(module, system): - + """ Create a host """ changed = True - if not module.check_mode: - host = system.hosts.create(name=module.params['name']) - return changed - - -@api_wrapper -def update_host(module, host): - changed = False + system.hosts.create(name=module.params['name']) return changed @api_wrapper def delete_host(module, host): + """ Delete a host """ changed = True if not module.check_mode: # May raise APICommandFailed if mapped, etc. @@ -83,13 +80,8 @@ def delete_host(module, host): return changed -def get_sys_host(module): - system = get_system(module) - host = get_host(module, system) - return (system, host) - - def get_host_fields(host): + """ Get host fields """ fields = host.get_fields(from_cache=True, raw_value=True) created_at, created_at_timezone = unixMillisecondsToDate(fields.get('created_at', None)) field_dict = dict( @@ -117,45 +109,52 @@ def get_host_fields(host): def handle_stat(module): - system, host = get_sys_host(module) + """ Handle the stat state """ + system = get_system(module) + host = get_host(module, system) host_name = module.params["name"] if not host: - module.fail_json(msg='Host {0} not found'.format(host_name)) + module.fail_json(msg=f'Host {host_name} not found') field_dict = get_host_fields(host) result = dict( changed=False, - msg='Host stat found' + msg=f'Host {host_name} stat found' ) result = merge_two_dicts(result, field_dict) module.exit_json(**result) def handle_present(module): - system, host = get_sys_host(module) + """ Handle the present state """ + system = get_system(module) + host = get_host(module, system) host_name = module.params["name"] if not host: changed = create_host(module, system) - msg = 'Host {0} created'.format(host_name) + msg = f'Host {host_name} created' module.exit_json(changed=changed, msg=msg) else: - changed = update_host(module, host) - msg = 'Host {0} updated'.format(host_name) + changed = False + msg = f'Host {host_name} exists and does not need to be updated' module.exit_json(changed=changed, msg=msg) def handle_absent(module): - system, host = get_sys_host(module) + """ Handle the absent state """ + system = get_system(module) + host = get_host(module, system) host_name = module.params["name"] if not host: - msg = "Host {0} already absent".format(host_name) + msg = f"Host {host_name} already absent" module.exit_json(changed=False, msg=msg) else: changed = delete_host(module, host) - msg = "Host {0} removed".format(host_name) + msg = f"Host {host_name} removed" module.exit_json(changed=changed, msg=msg) def execute_state(module): + """ Execute a state """ state = module.params['state'] try: if state == 'stat': @@ -165,13 +164,14 @@ def execute_state(module): elif state == 'absent': handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_map.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_map.py index e3757e021..1111930a3 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_map.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_map.py @@ -1,7 +1,11 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +"""This module creates, deletes or modifies mappings on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import (absolute_import, division, print_function) @@ -11,7 +15,7 @@ __metaclass__ = type DOCUMENTATION = r''' --- module: infini_map -version_added: '2.9.0' +version_added: 2.9.0 short_description: Create and Delete mapping of a volume to a host or cluster on Infinibox description: - This module creates or deletes mappings of volumes to hosts or clusters @@ -24,10 +28,12 @@ options: host: description: - Host Name + type: str required: false cluster: description: - Cluster Name + type: str required: false state: description: @@ -40,10 +46,12 @@ options: volume: description: - Volume name to map to the host. + type: str required: true lun: description: - Volume lun. + type: int extends_documentation_fragment: - infinibox ''' @@ -86,15 +94,8 @@ EXAMPLES = r''' password: secret ''' - # RETURN = r''' # ''' -import traceback -# import sh - -# rescan_scsi = sh.Command("rescan-scsi-bus.sh") -# rescan_scsi_remove = rescan_scsi.bake("--remove") - from ansible.module_utils.basic import AnsibleModule, missing_required_lib from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( @@ -102,7 +103,6 @@ from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox impo api_wrapper, get_cluster, get_host, - get_pool, get_system, get_volume, infinibox_argument_spec, @@ -110,39 +110,31 @@ from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox impo ) try: - from infinisdk.core.exceptions import APICommandFailed, ObjectNotFound + from infinisdk.core.exceptions import APICommandFailed except ImportError: pass # Handled by HAS_INFINISDK from module_utils def vol_is_mapped_to_host(volume, host): - volume_fields = volume.get_fields() - volume_id = volume_fields.get('id') + """ Return a bool showing if a vol is mapped to a host """ host_luns = host.get_luns() - # print('volume id: {0}'.format(volume_id)) - # print('host luns: {0}'.format(str(host_luns))) for lun in host_luns: if lun.volume == volume: - # print('found mapped volume: {0}'.format(volume)) return True return False def vol_is_mapped_to_cluster(volume, cluster): - volume_fields = volume.get_fields() - volume_id = volume_fields.get('id') + """ Return a bool showing if a vol is mapped to a cluster """ cluster_luns = cluster.get_luns() - # print('volume id: {0}'.format(volume_id)) - # print('host luns: {0}'.format(str(host_luns))) - for lun in cluster_luns: if lun.volume == volume: - # print('found mapped volume: {0}'.format(volume)) return True return False def find_host_lun_use(module, host, volume): + """ Return a dict showing if a host lun matches a volume. """ check_result = {'lun_used': False, 'lun_volume_matches': False} desired_lun = module.params['lun'] @@ -158,12 +150,13 @@ def find_host_lun_use(module, host, volume): def find_cluster_lun_use(module, cluster, volume): + """ Return a dict showing if a cluster lun matches a volume. """ check_result = {'lun_used': False, 'lun_volume_matches': False} desired_lun = module.params['lun'] if desired_lun: for cluster_lun in cluster.get_luns(): - if desired_lun == cluster.lun: + if desired_lun == cluster_lun: if cluster.volume == volume: check_result = {'lun_used': True, 'lun_volume_matches': True} else: @@ -173,6 +166,7 @@ def find_cluster_lun_use(module, cluster, volume): def find_host_lun(host, volume): + """ Find a hosts lun """ found_lun = None luns = host.get_luns() @@ -183,6 +177,7 @@ def find_host_lun(host, volume): def find_cluster_lun(cluster, volume): + """ Find a cluster's LUN """ found_lun = None luns = cluster.get_luns() @@ -194,12 +189,8 @@ def find_cluster_lun(cluster, volume): @api_wrapper def create_mapping(module, system): - """ - Create mapping of volume to host or cluster. If already mapped, exit_json with changed False. - """ + """ Create mapping of volume to host or cluster. If already mapped, exit_json with changed False. """ - host_name = module.params['host'] - cluster_name = module.params['cluster'] host = get_host(module, system) cluster = get_cluster(module, system) @@ -221,20 +212,18 @@ def create_mapping(module, system): @api_wrapper def create_mapping_to_cluster(module, system): - """ - Create mapping of volume to cluster. If already mapped, exit_json with changed False. - """ + """ Create mapping of volume to cluster. If already mapped, exit_json with changed False. """ changed = False cluster = get_cluster(module, system) volume = get_volume(module, system) + volume_name = module.params['volume'] + cluster_name = module.params['cluster'] + lun_name = module.params['lun'] lun_use = find_cluster_lun_use(module, cluster, volume) if lun_use['lun_used']: - msg = "Cannot create mapping of volume '{0}' to cluster '{1}' using lun '{2}'. Lun in use.".format( - volume.get_name(), - cluster.get_name(), - module.params['lun']) + msg = f"Cannot create mapping of volume '{volume_name}' to cluster '{cluster_name}' using lun '{lun_name}'. Lun in use." module.fail_json(msg=msg) try: @@ -244,30 +233,26 @@ def create_mapping_to_cluster(module, system): changed = True except APICommandFailed as err: if "is already mapped" not in str(err): - module.fail_json('Cannot map volume {0} to cluster {1}: {2}. Already mapped.'.format( - module.params['volume'], - module.params['cluster'], - str(err))) + msg = f"Cannot map volume '{volume_name}' to cluster '{cluster_name}': {str(err)}. Already mapped." + module.fail_json(msg=msg) return changed @api_wrapper def create_mapping_to_host(module, system): - """ - Create mapping of volume to host. If already mapped, exit_json with changed False. - """ + """ Create mapping of volume to host. If already mapped, exit_json with changed False. """ changed = False host = system.hosts.get(name=module.params['host']) volume = get_volume(module, system) + volume_name = module.params['volume'] + host_name = module.params['host'] + lun_name = module.params['lun'] lun_use = find_host_lun_use(module, host, volume) if lun_use['lun_used']: - msg = "Cannot create mapping of volume '{0}' to host '{1}' using lun '{2}'. Lun in use.".format( - volume.get_name(), - host.get_name(), - module.params['lun']) + msg = f"Cannot create mapping of volume '{volume_name}' to host '{host_name}' using lun '{lun_name}'. Lun in use." module.fail_json(msg=msg) try: @@ -277,35 +262,29 @@ def create_mapping_to_host(module, system): changed = True except APICommandFailed as err: if "is already mapped" not in str(err): - module.fail_json('Cannot map volume {0} to host {1}: {2}. Already mapped.'.format( - module.params['volume'], - module.params['host'], - str(err))) + msg = f"Cannot map volume '{host_name}' to host '{host_name}': {str(err)}. Already mapped." + module.fail_json(msg=msg) return changed @api_wrapper def update_mapping_to_host(module, system): + """ Update a mapping to a host """ host = get_host(module, system) volume = get_volume(module, system) + volume_name = module.params['volume'] + host_name = module.params['host'] desired_lun = module.params['lun'] if not vol_is_mapped_to_host(volume, host): - msg = "Volume {0} is not mapped to host {1}".format( - volume.get_name(), - host.get_name(), - ) + msg = f"Volume '{volume_name}' is not mapped to host '{host_name}'" module.fail_json(msg=msg) if desired_lun: found_lun = find_host_lun(host, volume) if found_lun != desired_lun: - msg = "Cannot change the lun from '{0}' to '{1}' for existing mapping of volume '{2}' to host '{3}'".format( - found_lun, - desired_lun, - volume.get_name(), - host.get_name()) + msg = f"Cannot change the lun from '{found_lun}' to '{desired_lun}' for existing mapping of volume '{volume_name}' to host '{host_name}'" module.fail_json(msg=msg) changed = False @@ -314,25 +293,21 @@ def update_mapping_to_host(module, system): @api_wrapper def update_mapping_to_cluster(module, system): + """ Update a mapping to a cluster """ cluster = get_cluster(module, system) volume = get_volume(module, system) desired_lun = module.params['lun'] + volume_name = module.params['volume'] + cluster_name = module.params['cluster'] if not vol_is_mapped_to_cluster(volume, cluster): - msg = "Volume {0} is not mapped to cluster {1}".format( - volume.get_name(), - cluster.get_name(), - ) + msg = f"Volume {volume_name} is not mapped to cluster {cluster_name}" module.fail_json(msg=msg) if desired_lun: found_lun = find_cluster_lun(cluster, volume) if found_lun != desired_lun: - msg = "Cannot change the lun from '{0}' to '{1}' for existing mapping of volume '{2}' to cluster '{3}'".format( - found_lun, - desired_lun, - volume.get_name(), - cluster.get_name()) + msg = f"Cannot change the lun from '{found_lun}' to '{desired_lun}' for existing mapping of volume '{volume_name}' to cluster '{cluster_name}'" module.fail_json(msg=msg) changed = False @@ -341,6 +316,7 @@ def update_mapping_to_cluster(module, system): @api_wrapper def delete_mapping(module, system): + """ Delete a mapping """ host = get_host(module, system) cluster = get_cluster(module, system) if host: @@ -373,34 +349,22 @@ def delete_mapping_to_host(module, system): if not module.check_mode: volume = get_volume(module, system) host = get_host(module, system) + volume_name = module.params['volume'] + host_name = module.params['host'] if volume and host: try: existing_lun = find_host_lun(host, volume) host.unmap_volume(volume) changed = True - msg = "Volume '{0}' was unmapped from host '{1}' freeing lun '{2}'".format( - module.params['volume'], - module.params['host'], - existing_lun, - ) - + msg = f"Volume '{volume_name}' was unmapped from host '{host_name}' freeing lun '{existing_lun}'" except KeyError as err: if 'has no logical units' not in str(err): - module.fail_json('Cannot unmap volume {0} from host {1}: {2}'.format( - module.params['volume'], - module.params['host'], - str(err))) + module.fail_json(f"Cannot unmap volume '{volume_name}' from host '{host_name}': {str(err)}") else: - msg = "Volume {0} was not mapped to host {1} and so unmapping was not executed".format( - module.params['volume'], - module.params['host'], - ) + msg = f"Volume '{volume_name}' was not mapped to host '{host_name}' and so unmapping was not executed" else: - msg = "Either volume {0} or host {1} does not exist. Unmapping was not executed".format( - module.params['volume'], - module.params['host'], - ) + msg = f"Either volume '{volume_name}' or host '{host_name}' does not exist. Unmapping was not executed" else: # check_mode changed = True @@ -421,55 +385,31 @@ def delete_mapping_to_cluster(module, system): if not module.check_mode: volume = get_volume(module, system) cluster = get_cluster(module, system) + volume_name = module.params['volume'] + cluster_name = module.params['cluster'] if volume and cluster: try: existing_lun = find_cluster_lun(cluster, volume) cluster.unmap_volume(volume) changed = True - msg = "Volume '{0}' was unmapped from cluster '{1}' freeing lun '{2}'".format( - module.params['volume'], - module.params['cluster'], - existing_lun, - ) + msg = f"Volume '{volume_name}' was unmapped from cluster '{cluster_name}' freeing lun '{existing_lun}'" except KeyError as err: if 'has no logical units' not in str(err): - module.fail_json('Cannot unmap volume {0} from cluster {1}: {2}'.format( - module.params['volume'], - module.params['cluster'], - str(err))) + msg = f"Cannot unmap volume '{volume_name}' from cluster '{cluster_name}': {str(err)}" + module.fail_json(msg=msg) else: - msg = "Volume {0} was not mapped to cluster {1} and so unmapping was not executed".format( - module.params['volume'], - module.params['cluster'], - ) + msg = f"Volume '{volume_name}' was not mapped to cluster '{cluster_name}' and so unmapping was not executed" else: - msg = "Either volume {0} or cluster {1} does not exist. Unmapping was not executed".format( - module.params['volume'], - module.params['cluster'], - ) + msg = f"Either volume '{volume_name}' or cluster '{cluster_name}' does not exist. Unmapping was not executed" else: # check_mode changed = True module.exit_json(msg=msg, changed=changed) -def get_sys_vol_host_cluster(module): - system = get_system(module) - volume = get_volume(module, system) - host = get_host(module, system) - cluster = get_cluster(module, system) - return (system, volume, host, cluster) - - -def get_sys_vol_cluster(module): - system = get_system(module) - volume = get_volume(module, system) - cluster = get_cluster(module, system) - return (system, volume, cluster) - - def get_mapping_fields(volume, host_or_cluster): + """ Get mapping fields """ luns = host_or_cluster.get_luns() for lun in luns: if volume.get_name() == lun.volume.get_name(): @@ -481,50 +421,56 @@ def get_mapping_fields(volume, host_or_cluster): def handle_stat(module): - system, volume, host, cluster = get_sys_vol_host_cluster(module) + """ Return mapping stat """ + system = get_system(module) + volume = get_volume(module, system) + host = get_host(module, system) + cluster = get_cluster(module, system) volume_name = module.params['volume'] host_name = module.params['host'] - if not host_name: - host_name = "not specified" - cluster_name = module.params['cluster'] - if not cluster_name: - cluster_name = "not specified" if not volume: - module.fail_json(msg='Volume {0} not found'.format(volume_name)) + module.fail_json(msg=f"Volume '{volume_name}' not found") + if not host and not cluster: - module.fail_json(msg='Neither host [{0}] nor cluster [{1}] found'.format(host_name, cluster_name)) + msg = f"Neither host '{host_name}' nor cluster '{cluster_name}' found" + module.fail_json(msg=msg) + if (not host or not vol_is_mapped_to_host(volume, host)) \ and (not cluster or not vol_is_mapped_to_cluster(volume, cluster)): - msg = 'Volume {0} is mapped to neither host {1} nor cluster {2}'.format(volume_name, host_name, cluster_name) + if host_name: + msg = f"Volume '{volume_name}' is not mapped to host '{host_name}'" + elif cluster_name: + msg = f"Volume '{volume_name}' is not mapped to cluster '{cluster_name}'" module.fail_json(msg=msg) - if host: + + if host and host_name: found_lun = find_host_lun(host, volume) field_dict = get_mapping_fields(volume, host) if found_lun is not None: - msg = 'Volume {0} is mapped to host {1} using lun {2}'.format(volume_name, host_name, found_lun), + msg = f"Volume '{volume_name}' is mapped to host '{host_name}' using lun '{found_lun}'" result = dict( changed=False, volume_lun=found_lun, msg=msg, ) else: - msg = 'Volume {0} is not mapped to host {1}'.format(volume_name, host_name) + msg = f"Volume '{volume_name}' is not mapped to host '{host_name}'" module.fail_json(msg=msg) - elif cluster: + elif cluster and cluster_name: found_lun = find_cluster_lun(cluster, volume) field_dict = get_mapping_fields(volume, cluster) if found_lun is not None: - msg = 'Volume {0} is mapped to cluster {1} using lun {2}'.format(volume_name, cluster_name, found_lun) + msg = f"Volume '{volume_name}' is mapped to cluster '{cluster_name}' using lun '{found_lun}'" result = dict( changed=False, volume_lun=found_lun, msg=msg, ) else: - msg = 'Volume {0} is not mapped to cluster {1}'.format(volume_name, cluster_name) + msg = f"Volume '{volume_name}' is not mapped to cluster '{cluster_name}'" module.fail_json(msg=msg) else: msg = 'A programming error has occurred in handle_stat()' @@ -534,50 +480,38 @@ def handle_stat(module): def handle_present(module): - system, volume, host, cluster = get_sys_vol_host_cluster(module) + """ Create or update mapping """ + system = get_system(module) + volume = get_volume(module, system) + host = get_host(module, system) + cluster = get_cluster(module, system) volume_name = module.params['volume'] host_name = module.params['host'] cluster_name = module.params['cluster'] if not volume: - module.fail_json(changed=False, msg='Volume {0} not found'.format(volume_name)) + module.fail_json(changed=False, msg=f"Volume '{volume_name}' not found") if not host and not cluster: if not host_name: host_name = "not specified" if not cluster_name: cluster_name = "not specified" - module.fail_json(changed=False, msg='Neither host [{0}] nor cluster [{1}] found'.format(host_name, cluster_name)) + module.fail_json(changed=False, msg=f"Neither host '{host_name}' nor cluster '{cluster_name}' found") if host: if not vol_is_mapped_to_host(volume, host): changed = create_mapping(module, system) - # TODO: Why is find_host_lun() returning None after creating the mapping? - # host.get_luns() returns an empty list, why? - # existing_lun = find_host_lun(host, volume) - # msg = "Volume '{0}' map to host '{1}' created using lun '{2}'".format( - # volume.get_name(), - # host.get_name(), - # existing_lun, - # ) - msg = "Volume '{0}' map to host '{1}' created".format(volume_name, host_name) + msg = f"Volume '{volume_name}' map to host '{host_name}' created" else: changed = update_mapping_to_host(module, system) existing_lun = find_host_lun(host, volume) - msg = "Volume '{0}' map to host '{1}' already exists using lun '{2}'".format(volume_name, host_name, existing_lun) + msg = f"Volume '{volume_name}' map to host '{host_name}' already exists using lun '{existing_lun}'" elif cluster: if not vol_is_mapped_to_cluster(volume, cluster): changed = create_mapping(module, system) - # TODO: Why is find_host_lun() returning None after creating the mapping? - # host.get_luns() returns an empty list, why? - # existing_lun = find_host_lun(host, volume) - # msg = "Volume '{0}' map to host '{1}' created using lun '{2}'".format( - # volume.get_name(), - # host.get_name(), - # existing_lun, - # ) - msg = "Volume '{0}' map to cluster '{1}' created".format(volume_name, cluster_name) + msg = f"Volume '{volume_name}' map to cluster '{cluster_name}' created" else: changed = update_mapping_to_cluster(module, system) existing_lun = find_cluster_lun(cluster, volume) - msg = "Volume '{0}' map to cluster '{1}' already exists using lun '{2}'".format(volume_name, cluster_name, existing_lun) + msg = f"Volume '{volume_name}' map to cluster '{cluster_name}' already exists using lun '{existing_lun}'" result = dict( changed=changed, @@ -587,18 +521,23 @@ def handle_present(module): def handle_absent(module): - system, volume, host, cluster = get_sys_vol_host_cluster(module) + """ Remove mapping """ + system = get_system(module) + volume = get_volume(module, system) + host = get_host(module, system) + cluster = get_cluster(module, system) volume_name = module.params['volume'] host_name = module.params['host'] cluster_name = module.params['cluster'] if not volume or (not host and not cluster): - module.exit_json(changed=False, msg='Mapping of volume {0} to host {1} or cluster {2} already absent'.format(volume_name, host_name, cluster_name)) + module.exit_json(changed=False, msg=f'Mapping of volume {volume_name} to host {host_name} or cluster {cluster_name} already absent') else: changed = delete_mapping(module, system) module.exit_json(changed=changed, msg="Mapping removed") def execute_state(module): + """Determine which state function to execute and do so""" state = module.params['state'] try: if state == 'stat': @@ -608,14 +547,14 @@ def execute_state(module): elif state == 'absent': handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() def check_parameters(module): - volume_name = module.params['volume'] + """Verify module options are sane""" host_name = module.params['host'] cluster_name = module.params['cluster'] if host_name and cluster_name: @@ -628,17 +567,15 @@ def check_parameters(module): def main(): - """ - Gather auguments and manage mapping of vols to hosts. - """ + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( - host=dict(required=False, default=""), - cluster=dict(required=False, default=""), + host=dict(required=False, default=None), + cluster=dict(required=False, default=None), state=dict(default='present', choices=['stat', 'present', 'absent']), volume=dict(required=True), - lun=dict(type=int), + lun=dict(type="int"), ) ) diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_metadata.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_metadata.py new file mode 100644 index 000000000..15374a52e --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_metadata.py @@ -0,0 +1,674 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +"""This module creates, deletes or modifies metadata on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_metadata +version_added: 2.13.0 +short_description: Create, Delete or Modify metadata on Infinibox +description: + - This module creates, deletes or modifies metadata on Infinibox. + - Deleting metadata by object, without specifying a key, is not implemented for any object_type (e.g. DELETE api/rest/metadata/system). + - This would delete all metadata belonging to the object. Instead delete each key explicitely using its key name. +author: David Ohlemacher (@ohlemacher) +options: + object_type: + description: + - Type of object + type: str + required: true + choices: ["cluster", "fs", "fs-snap", "host", "pool", "system", "vol", "vol-snap"] + object_name: + description: + - Name of the object. Not used if object_type is system + type: str + required: false + key: + description: + - Name of the metadata key + type: str + required: true + value: + description: + - Value of the metadata key + type: str + required: false + state: + description: + - Creates/Modifies metadata when present or removes when absent. + type: str + required: false + default: present + choices: [ "stat", "present", "absent" ] + +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Create new metadata key foo with value bar + infini_metadata: + name: foo + key: bar + state: present + user: admin + password: secret + system: ibox001 +- name: Stat metadata key named foo + infini_metadata: + name: foo + state: stat + user: admin + password: secret + system: ibox001 +- name: Remove metadata keyn named foo + infini_vol: + name: foo_snap + state: absent + user: admin + password: secret + system: ibox001 +""" + +# RETURN = r''' # ''' + +import json + +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + HAS_INFINISDK, + api_wrapper, + get_cluster, + get_filesystem, + get_host, + get_pool, + get_system, + get_volume, + infinibox_argument_spec, +) + +HAS_INFINISDK = True +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + HAS_INFINISDK = False + +HAS_CAPACITY = False + + +@api_wrapper +def get_metadata_vol(module, disable_fail): + """ Get metadata about a volume """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + vol = get_volume(module, system) + if vol: + path = f"metadata/{vol.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"Volume {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"Volume with object name {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata_fs(module, disable_fail): + """ Get metadata about a fs """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + fs = get_filesystem(module, system) + if fs: + path = f"metadata/{fs.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"File system {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"File system named {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata_host(module, disable_fail): + """ Get metadata about a host """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + host = get_host(module, system) + if host: + path = f"metadata/{host.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"Host {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"Host named {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata_cluster(module, disable_fail): + """ Get metadata about a cluster """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + cluster = get_cluster(module, system) + if cluster: + path = f"metadata/{cluster.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"Cluster {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"Cluster named {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata_fssnap(module, disable_fail): + """ Get metadata about a fs snapshot """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + fssnap = get_filesystem(module, system) + if fssnap: + path = f"metadata/{fssnap.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"File system snapshot {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"File system snapshot named {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata_pool(module, disable_fail): + """ Get metadata about a pool """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + pool = get_pool(module, system) + if pool: + path = f"metadata/{pool.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"Pool {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"Pool named {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata_volsnap(module, disable_fail): + """ Get metadata for a volume snapshot """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + metadata = None + + volsnap = get_volume(module, system) + if volsnap: + path = f"metadata/{volsnap.id}/{key}" + try: + metadata = system.api.get(path=path) + except APICommandFailed: + if not disable_fail: + module.fail_json( + f"Cannot find {object_type} metadata key. " + f"Volume snapshot {object_name} key {key} not found" + ) + elif not disable_fail: + msg = f"Volume snapshot named {object_name} not found. Cannot stat its metadata." + module.fail_json(msg=msg) + + return metadata + + +@api_wrapper +def get_metadata(module, disable_fail=False): + """ + Find and return metadata + Use disable_fail when we are looking for metadata + and it may or may not exist and neither case is an error. + """ + system = get_system(module) + object_type = module.params["object_type"] + object_name = module.params["object_name"] + key = module.params["key"] + + if object_type == "system": + path = f"metadata/{object_type}?key={key}" + metadata = system.api.get(path=path) + elif object_type == "fs": + metadata = get_metadata_fs(module, disable_fail) + elif object_type == "vol": + metadata = get_metadata_vol(module, disable_fail) + elif object_type == "host": + metadata = get_metadata_host(module, disable_fail) + elif object_type == "cluster": + metadata = get_metadata_cluster(module, disable_fail) + elif object_type == "fs-snap": + metadata = get_metadata_fs(module, disable_fail) + elif object_type == "pool": + metadata = get_metadata_pool(module, disable_fail) + elif object_type == "vol-snap": + metadata = get_metadata_volsnap(module, disable_fail) + + else: + msg = f"Metadata for {object_type} not supported. Cannot stat." + module.fail_json(msg=msg) + + if metadata: + result = metadata.get_result() + if not disable_fail and not result: + msg = f"Metadata for {object_type} with key {key} not found. Cannot stat." + module.fail_json(msg=msg) + return result + + if disable_fail: + return None + + msg = f"Metadata for {object_type} named {object_name} not found. Cannot stat." + module.fail_json(msg=msg) + return None # Quiet pylint + + +@api_wrapper +def put_metadata(module): # pylint: disable=too-many-statements + """Create metadata key with a value. The changed variable is found elsewhere.""" + system = get_system(module) + + object_type = module.params["object_type"] + key = module.params["key"] + value = module.params["value"] + + # Could check metadata value size < 32k + + if object_type == "system": + path = "metadata/system" + elif object_type == "vol": + vol = get_volume(module, system) + if not vol: + object_name = module.params["object_name"] + msg = f"Volume {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{vol.id}" + elif object_type == "fs": + fs = get_filesystem(module, system) + if not fs: + object_name = module.params["object_name"] + msg = f"File system {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{fs.id}" + elif object_type == "host": + host = get_host(module, system) + if not host: + object_name = module.params["object_name"] + msg = f"Cluster {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{host.id}" + elif object_type == "cluster": + cluster = get_cluster(module, system) + if not cluster: + object_name = module.params["object_name"] + msg = f"Cluster {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{cluster.id}" + elif object_type == "fs-snap": + fssnap = get_filesystem(module, system) + if not fssnap: + object_name = module.params["object_name"] + msg = f"File system snapshot {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{fssnap.id}" + elif object_type == "pool": + pool = get_pool(module, system) + if not pool: + object_name = module.params["object_name"] + msg = f"Pool {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{pool.id}" + elif object_type == "vol-snap": + volsnap = get_volume(module, system) + if not volsnap: + object_name = module.params["object_name"] + msg = f"Volume snapshot {object_name} not found. Cannot add metadata key {key}." + module.fail_json(msg=msg) + path = f"metadata/{volsnap.id}" + + # Create json data + data = { + key: value + } + + # Put + system.api.put(path=path, data=data) + # Variable 'changed' not returned by design + + +@api_wrapper +def delete_metadata(module): # pylint: disable=too-many-return-statements + """ + Remove metadata key. + Not implemented by design: Deleting all of the system's metadata + using 'DELETE api/rest/metadata/system'. + """ + system = get_system(module) + changed = False + object_type = module.params["object_type"] + key = module.params["key"] + if object_type == "system": + path = f"metadata/system/{key}" + elif object_type == "vol": + vol = get_volume(module, system) + if not vol: + changed = False + return changed # No vol therefore no metadata to delete + path = f"metadata/{vol.id}/{key}" + elif object_type == "fs": + fs = get_filesystem(module, system) + if not fs: + changed = False + return changed # No fs therefore no metadata to delete + path = f"metadata/{fs.id}/{key}" + elif object_type == "host": + host = get_host(module, system) + if not host: + changed = False + return changed # No host therefore no metadata to delete + path = f"metadata/{host.id}/{key}" + elif object_type == "cluster": + cluster = get_cluster(module, system) + if not cluster: + changed = False + return changed # No cluster therefore no metadata to delete + path = f"metadata/{cluster.id}/{key}" + elif object_type == "fs-snap": + fssnap = get_filesystem(module, system) + if not fssnap: + changed = False + return changed # No fssnap therefore no metadata to delete + path = f"metadata/{fssnap.id}/{key}" + elif object_type == "pool": + pool = get_pool(module, system) + if not pool: + changed = False + return changed # No pool therefore no metadata to delete + path = f"metadata/{pool.id}/{key}" + elif object_type == "vol-snap": + volsnap = get_volume(module, system) + if not volsnap: + changed = False + return changed # No volsnap therefore no metadata to delete + path = f"metadata/{volsnap.id}/{key}" + else: + module.fail_json(f"Object type {object_type} not supported") + + try: + system.api.delete(path=path) + changed = True + except APICommandFailed as err: + if err.status_code != 404: + raise + return changed + + +def handle_stat(module): + """Return metadata stat""" + object_type = module.params["object_type"] + key = module.params["key"] + metadata = get_metadata(module) + if object_type == "system": + metadata_id = metadata[0]["id"] + object_id = metadata[0]["object_id"] + value = metadata[0]["value"] + else: + metadata_id = metadata["id"] + object_id = metadata["object_id"] + value = metadata["value"] + + result = { + "msg": "Metadata found", + "changed": False, + "object_type": object_type, + "key": key, + "id": metadata_id, + "object_id": object_id, + "value": value, + } + module.exit_json(**result) + + +def handle_present(module): + """Make metadata present""" + changed = False + msg = "Metadata unchanged" + if not module.check_mode: + old_metadata = get_metadata(module, disable_fail=True) + put_metadata(module) + new_metadata = get_metadata(module) + changed = new_metadata != old_metadata + if changed: + msg = "Metadata changed" + else: + msg = "Metadata unchanged since the value is the same as the existing metadata" + module.exit_json(changed=changed, msg=msg) + + +def handle_absent(module): + """Make metadata absent""" + msg = "Metadata unchanged" + changed = False + if not module.check_mode: + changed = delete_metadata(module) + if changed: + msg = "Metadata removed" + else: + msg = "Metadata did not exist so no removal was necessary" + module.exit_json(changed=changed, msg=msg) + + +def execute_state(module): + """Determine which state function to execute and do so""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + elif state == "absent": + handle_absent(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + state = module.params["state"] + object_type = module.params["object_type"] + object_name = module.params["object_name"] + + # Check object_type + object_types = [ + "cluster", + "fs", + "fs-snap", + "host", + "pool", + "system", + "vol", + "vol-snap", + ] + if object_type not in object_types: + module.fail_json( + f"Cannot create {object_type} metadata. Object type must be one of {object_types}" + ) + + # Check object_name + if object_type == "system": + if object_name: + module.fail_json("An object_name for object_type system must not be provided.") + else: + if not object_name: + module.fail_json( + f"The name of the {object_type} must be provided as object_name." + ) + + key = module.params["key"] + if not key: + module.fail_json(f"Cannot create a {object_type} metadata key without providing a key name") + + if state == "stat": + pass + elif state == "present": + # Check value + key = module.params["key"] + value = module.params["value"] + if not value: + module.fail_json( + f"Cannot create a {object_type} metadata key {key} without providing a value" + ) + # Check system object_type + if object_type == "system": + if key == "ui-dataset-default-provisioning": + values = ["THICK", "THIN"] + if value not in values: + module.fail_json( + f"Cannot create {object_type} metadata for key {key}. " + f"Value must be one of {values}. Invalid value: {value}." + ) + + # Convert bool string to bool + if key in [ + "ui-dataset-base2-units", + "ui-feedback-dialog", + "ui-feedback-form", + ]: + try: + module.params["value"] = json.loads(value.lower()) + except json.decoder.JSONDecodeError: + module.fail_json( + f"Cannot create {object_type} metadata for key {key}. " + f"Value must be able to be decoded as a boolean. Invalid value: {value}." + ) + + # Convert integer string to int + if key in [ + "ui-bulk-volume-zero-padding", + "ui-table-export-limit" + ]: + try: + module.params["value"] = json.loads(value.lower()) + except json.decoder.JSONDecodeError: + module.fail_json( + f"Cannot create {object_type} metadata for key {key}. " + f"Value must be of type integer. Invalid value: {value}." + ) + + elif state == "absent": + pass + else: + module.fail_json(f"Invalid state '{state}' provided") + + +def main(): + """ Main """ + argument_spec = infinibox_argument_spec() + + argument_spec.update( + { + "object_type": {"required": True, "choices": ["cluster", "fs", "fs-snap", "host", "pool", "system", "vol", "vol-snap"]}, + "object_name": {"required": False, "default": None}, + "key": {"required": True, "no_log": False}, + "value": {"required": False, "default": None}, + "state": {"default": "present", "choices": ["stat", "present", "absent"]}, + } + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_INFINISDK: + module.fail_json(msg=missing_required_lib("infinisdk")) + + check_options(module) + execute_state(module) + + +if __name__ == "__main__": + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_network_space.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_network_space.py index 013d86e5e..f9f02e11d 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_network_space.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_network_space.py @@ -1,9 +1,13 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-list-literal,use-dict-literal,line-too-long,wrong-import-position + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) +"""This module creates, deletes or modifies network spaces on Infinibox.""" + from __future__ import (absolute_import, division, print_function) __metaclass__ = type @@ -20,25 +24,35 @@ options: name: description: - Network space name + type: str required: true state: description: - Creates/Modifies network spaces when present. Removes when absent. Shows status when stat. + type: str required: false default: present choices: [ "stat", "present", "absent" ] interfaces: description: - - A list of interfaces for the space. + - A list of interface IDs for the space. required: false type: list - elements: str + elements: int + default: [] + network_config: + description: + - A network description. + type: dict + default: {} + required: false service: description: - Choose a service. + type: str required: false - default: "replication" - choices: ["replication", "NAS", "iSCSI"] + default: "RMR_SERVICE" + choices: ["RMR_SERVICE", "NAS_SERVICE", "ISCSI_SERVICE"] mtu: description: - Set an MTU. If not specified, defaults to 1500 bytes. @@ -54,6 +68,11 @@ options: - Network mask. required: false type: int + default_gateway: + description: + - Default gateway. + type: str + required: false ips: description: - List of IPs. @@ -69,7 +88,12 @@ options: - It does not affect sync-replication or active-active traffic. required: false type: int - + async_only: + description: + - Run asynchronously only. + required: false + type: bool + default: false extends_documentation_fragment: - infinibox ''' @@ -80,20 +104,20 @@ EXAMPLES = r''' name: iSCSI state: present interfaces: - - 1680 - - 1679 - - 1678 + - 1680 + - 1679 + - 1678 service: ISCSI_SERVICE netmask: 19 network: 172.31.32.0 default_gateway: 172.31.63.254 ips: - - 172.31.32.145 - - 172.31.32.146 - - 172.31.32.147 - - 172.31.32.148 - - 172.31.32.149 - - 172.31.32.150 + - 172.31.32.145 + - 172.31.32.146 + - 172.31.32.147 + - 172.31.32.148 + - 172.31.32.149 + - 172.31.32.150 user: admin password: secret system: ibox001 @@ -103,28 +127,24 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, api_wrapper, infinibox_argument_spec, get_system, - unixMillisecondsToDate, merge_two_dicts, get_net_space, ) try: from infinisdk.core.exceptions import APICommandFailed - from infinisdk.core.exceptions import ObjectNotFound - from infi.dtypes.iqn import make_iscsi_name except ImportError: pass # Handled by HAS_INFINISDK from module_utils @api_wrapper def create_empty_network_space(module, system): + """ Create an empty network space """ # Create network space network_space_name = module.params["name"] service = module.params["service"] @@ -137,9 +157,7 @@ def create_empty_network_space(module, system): } interfaces = module.params["interfaces"] - # print("Creating network space {0}".format(network_space_name)) - product_id = system.api.get('system/product_id') - # print("api: {0}".format(product_id.get_result())) + # product_id = system.api.get('system/product_id') net_create_url = "network/spaces" net_create_data = { @@ -153,11 +171,13 @@ def create_empty_network_space(module, system): if mtu: net_create_data["mtu"] = mtu - net_create = system.api.post( - path=net_create_url, - data=net_create_data - ) - # print("net_create: {0}".format(net_create)) + try: + system.api.post( + path=net_create_url, + data=net_create_data + ) + except APICommandFailed as err: + module.fail_json(msg=f"Cannot create empty network space {network_space_name}: {err}") @api_wrapper @@ -166,36 +186,33 @@ def find_network_space_id(module, system): Find the ID of this network space """ network_space_name = module.params["name"] - net_id_url = "network/spaces?name={0}&fields=id".format(network_space_name) + net_id_url = f"network/spaces?name={network_space_name}&fields=id" net_id = system.api.get( path=net_id_url ) result = net_id.get_json()['result'][0] space_id = result['id'] - # print("Network space has ID {0}".format(space_id)) return space_id @api_wrapper def add_ips_to_network_space(module, system, space_id): + """ Add IPs to space. Ignore address conflict errors. """ network_space_name = module.params["name"] - # print("Adding IPs to network space {0}".format(network_space_name)) - ips = module.params["ips"] for ip in ips: - ip_url = "network/spaces/{0}/ips".format(space_id) + ip_url = f"network/spaces/{space_id}/ips" ip_data = ip - ip_add = system.api.post( - path=ip_url, - data=ip_data - ) - # print("add_ips json: {0}".format(ip_add.get_json())) - result = ip_add.get_json()['result'] - # print("add ip result: {0}".format(result)) + try: + system.api.post(path=ip_url, data=ip_data) + except APICommandFailed as err: + if err.error_code != "NET_SPACE_ADDRESS_CONFLICT": # Ignore + module.fail_json(msg=f"Cannot add IP {ip} to network space {network_space_name}: {err}") @api_wrapper def create_network_space(module, system): + """ Create a network space """ if not module.check_mode: # Create space create_empty_network_space(module, system) @@ -214,54 +231,43 @@ def create_network_space(module, system): def update_network_space(module, system): """ Update network space. - TODO - This is incomplete and will not update the space. - It will instead return changed=False and a message. - To implement this we will need to find the existing space. - For each field that we support updating, we need to compare existing - to new values and if different update. We will need to iterate - over the settings or we will receive: - Status: 400 - Code: NOT_SUPPORTED_MULTIPLE_UPDATE + Update fields individually. If grouped the API will generate + a NOT_SUPPORTED_MULTIPLE_UPDATE error. """ - changed = False - msg = "Update is not supported yet" - module.exit_json(changed=changed, msg=msg) - - # TODO Everything below is incomplete - # Update network space - network_space_name = module.params["name"] - service = module.params["service"] - network_config = { - "netmask": module.params["netmask"], - "network": module.params["network"], - "default_gateway": module.params["default_gateway"], - } - interfaces = module.params["interfaces"] - - # print("Updating network space {0}".format(network_space_name)) - - # Find space's ID space_id = find_network_space_id(module, system) - - net_url = "network/spaces/{0}".format(space_id) - net_data = { - "name": network_space_name, - "service": service, - "network_config": network_config, - "interfaces": interfaces, - } - - # Find existing space - net_existing = system.api.get(path=net_url) - - net_update = system.api.put( - path=net_url, - data=net_data - ) - # print("net_update: {0}".format(net_update)) + datas = [ + {"interfaces": module.params["interfaces"]}, + {"mtu": module.params["mtu"]}, + {"network_config": + { + "default_gateway": module.params["default_gateway"], + "netmask": module.params["netmask"], + "network": module.params["network"], + } + }, + {"rate_limit": module.params["rate_limit"]}, + {"properties": + { + "is_async_only": module.params["async_only"], + } + }, + ] + for data in datas: + try: + system.api.put( + path=f"network/spaces/{space_id}", + data=data + ) + except APICommandFailed as err: + msg = f"Cannot update network space: {err}" + module.fail_json(msg=msg) + add_ips_to_network_space(module, system, space_id) + changed = True + return changed -def get_network_space_fields(module, network_space): +def get_network_space_fields(network_space): + """ Get the network space fields and return as a dict """ fields = network_space.get_fields(from_cache=True, raw_value=True) field_dict = dict( @@ -282,16 +288,18 @@ def get_network_space_fields(module, network_space): def handle_stat(module): + """ Return details about the space """ network_space_name = module.params["name"] system = get_system(module) net_space = get_net_space(module, system) if not net_space: - module.fail_json(msg="Network space {0} not found".format(network_space_name)) + module.fail_json(msg=f"Network space {network_space_name} not found") - field_dict = get_network_space_fields(module, net_space) + field_dict = get_network_space_fields(net_space) result = dict( - changed=False, msg="Network space {0} stat found".format(network_space_name) + changed=False, + msg=f"Network space {network_space_name} stat found" ) result = merge_two_dicts(result, field_dict) module.exit_json(**result) @@ -305,14 +313,41 @@ def handle_present(module): system = get_system(module) net_space = get_net_space(module, system) if net_space: - changed = update_network_space(module, net_space) - msg = "Host {0} updated".format(network_space_name) + changed = update_network_space(module, system) + msg = f"Network space named {network_space_name} updated" else: changed = create_network_space(module, system) - msg = "Network space {0} created".format(network_space_name) + msg = f"Network space named {network_space_name} created" module.exit_json(changed=changed, msg=msg) +def disable_and_delete_ip(module, network_space, ip): + """ + Disable and delete a network space IP + """ + if not ip: + return # Nothing to do + addr = ip['ip_address'] + network_space_name = module.params["name"] + ip_type = ip['type'] + mgmt = "" + if ip_type == "MANAGEMENT": + mgmt = "management " # Trailing space by design + + try: + try: + network_space.disable_ip_address(addr) + except APICommandFailed as err: + if err.error_code == "IP_ADDRESS_ALREADY_DISABLED": + pass + else: + module.fail_json(msg=f"Disabling of network space {network_space_name} IP {mgmt}{addr} API command failed") + + network_space.remove_ip_address(addr) + except Exception as err: # pylint: disable=broad-exception-caught + module.fail_json(msg=f"Disabling or removal of network space {network_space_name} IP {mgmt}{addr} failed: {err}") + + def handle_absent(module): """ Remove a namespace. First, may disable and remove the namespace's IPs. @@ -322,55 +357,34 @@ def handle_absent(module): network_space = get_net_space(module, system) if not network_space: changed = False - msg = "Network space {0} already absent".format(network_space_name) + msg = f"Network space {network_space_name} already absent" else: # Find IPs from space ips = list(network_space.get_ips()) # Disable and delete IPs from space if not module.check_mode: + management_ip = None # Must be disabled and deleted last for ip in ips: - addr = ip["ip_address"] - - # print("Disabling IP {0}".format(addr)) - try: - network_space.disable_ip_address(addr) - except APICommandFailed as err: - if err.error_code == "IP_ADDRESS_ALREADY_DISABLED": - pass - # print("Already disabled IP {0}".format(addr)) - else: - # print("Failed to disable IP {0}".format(addr)) - module.fail_json( - msg="Disabling of network space {0} IP {1} failed".format( - network_space_name, addr - ) - ) - - # print("Removing IP {0}".format(addr)) - try: - network_space.remove_ip_address(addr) - except Exception as err: - module.fail_json( - msg="Removal of network space {0} IP {1} failed: {2}".format( - network_space_name, addr, err - ) - ) + if ip['type'] == 'MANAGEMENT': + management_ip = ip + continue + disable_and_delete_ip(module, network_space, ip) + disable_and_delete_ip(module, network_space, management_ip) # Delete space network_space.delete() changed = True - msg = "Network space {0} removed".format(network_space_name) + msg = f"Network space {network_space_name} removed" else: changed = False - msg = "Network space {0} not altered due to checkmode".format( - network_space_name - ) + msg = f"Network space {network_space_name} not altered due to checkmode" module.exit_json(changed=changed, msg=msg) def execute_state(module): + """ Execute a state """ state = module.params["state"] try: if state == "stat": @@ -381,7 +395,7 @@ def execute_state(module): handle_absent(module) else: module.fail_json( - msg="Internal handler error. Invalid state: {0}".format(state) + msg=f"Internal handler error. Invalid state: {state}" ) finally: system = get_system(module) @@ -389,6 +403,7 @@ def execute_state(module): def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( @@ -397,18 +412,19 @@ def main(): default="present", required=False, choices=["stat", "present", "absent"] ), service=dict( - default="replication", + default="RMR_SERVICE", required=False, - choices=["replication", "NAS_SERVICE", "ISCSI_SERVICE"], + choices=["RMR_SERVICE", "NAS_SERVICE", "ISCSI_SERVICE"], ), - mtu=dict(default=None, required=False, type=int), + mtu=dict(default=None, required=False, type="int"), network=dict(default=None, required=False), - netmask=dict(default=None, required=False, type=int), + netmask=dict(default=None, required=False, type="int"), default_gateway=dict(default=None, required=False), interfaces=dict(default=list(), required=False, type="list", elements="int"), - network_config=dict(default=dict(), required=False, type=dict), + network_config=dict(default=dict(), required=False, type="dict"), ips=dict(default=list(), required=False, type="list", elements="str"), - rate_limit=dict(default=None, required=False, type=int), + rate_limit=dict(default=None, required=False, type="int"), + async_only=dict(default=False, required=False, type="bool"), ) ) diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_notification_rule.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_notification_rule.py new file mode 100644 index 000000000..1916cdb49 --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_notification_rule.py @@ -0,0 +1,360 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +"""This module creates, deletes or modifies metadata on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_notification_rule +version_added: 2.13.0 +short_description: Config notification rules +description: + - This module config notification rules on Infinibox +author: Wei Wang (@wwang) +options: + name: + description: + - Name of the rule + type: str + required: true + event_level: + description: + - Event levels + type: list + elements: str + required: false + default: [] + include_events: + description: + - Included events + type: list + elements: str + required: false + default: [] + exclude_events: + description: + - Exclued events + type: list + elements: str + required: false + default: [] + recipients: + description: + - Email list of the recipients + - Recipients and target are exclusive to each other, i.e. only recipients or target + should be used, don't use both at the same time. + type: list + elements: str + required: false + default: [] + target: + description: + - Notification target + - Recipients and target are exclusive to each other, i.e. only recipients or target + should be used, don't use both at the same time. + type: str + required: false + state: + description: + - Query or modifies config. + type: str + required: false + default: present + choices: [ "stat", "present", "absent" ] + +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Create a new notification rule to a target + infini_notification_rule: + name: "test-rule-to-target" # this need to be uniq + event_level: + - ERROR + - CRITICAL + include_events: + - ACTIVATION_PAUSED + exclude_events: + - ACTIVE_DIRECTORY_ALL_DOMAIN_CONTROLLERS_DOWN + - ACTIVE_DIRECTORY_LEFT + target: testgraylog1 + state: "present" + user: "{{ user }}" + password: "{{ password }}" + system: "{{ system }}" +""" + +# RETURN = r''' # ''' + +# -*- coding: utf-8 -*- +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +HAS_ARROW = False + +try: + from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + HAS_INFINISDK, + api_wrapper, + infinibox_argument_spec, + get_system, + ) +except ModuleNotFoundError: + from infinibox import ( # Used when hacking + HAS_INFINISDK, + api_wrapper, + infinibox_argument_spec, + get_system, + ) + + +@api_wrapper +def find_target_id(module, system): + """ Find the ID of the target by name """ + target = module.params["target"] + path = f"notifications/targets?name={target}&fields=id" + api_result = system.api.get( + path=path + ) + if len(api_result.get_json()['result']) > 0: + result = api_result.get_json()['result'][0] + target_id = result['id'] + else: + target_id = None + return target_id + + +@api_wrapper +def get_rules(module): + """ Get all rules """ + system = get_system(module) + path = "notifications/rules" + rules = system.api.get(path=path) + return rules + + +@api_wrapper +def find_rule_id(module, system): + """ Find the ID of the rule by name """ + rule_name = module.params["name"] + path = f"notifications/rules?name={rule_name}&fields=id" + api_result = system.api.get( + path=path + ) + if len(api_result.get_json()['result']) > 0: + result = api_result.get_json()['result'][0] + rule_id = result['id'] + else: + rule_id = None + return rule_id + + +@api_wrapper +def delete_rule(module): + """ Delete a notification rule """ + system = get_system(module) + rule_id = find_rule_id(module, system) + path = f"notifications/rules/{rule_id}?approved=true" + system.api.delete(path=path) + + +@api_wrapper +def create_rule(module): + """ Create a new notifition rule """ + system = get_system(module) + name = module.params["name"] + event_level = module.params["event_level"] + include_events = module.params["include_events"] + exclude_events = module.params["exclude_events"] + recipients = module.params["recipients"] + target = module.params["target"] + path = "notifications/rules" + + json_data = { + "name": name, + "event_level": event_level, + "include_events": include_events, + "exclude_events": exclude_events, + } + + if recipients: + target_parameters = { + "recipients": recipients + } + target_id = 3 # Target ID for sending to recipients + json_data["target_parameters"] = target_parameters + elif target: + target_id = find_target_id(module, system) + else: + msg = "Neither recipients nor target parameters specified" + module.fail_json(msg=msg) + + json_data["target_id"] = target_id + + system.api.post(path=path, data=json_data) + + +@api_wrapper +def update_rule(module): + """ + Update an existing rule. + """ + system = get_system(module) + name = module.params["name"] + event_level = module.params["event_level"] + include_events = module.params["include_events"] + exclude_events = module.params["exclude_events"] + recipients = module.params["recipients"] + target = module.params["target"] + + json_data = { + "name": name, + "event_level": event_level, + "include_events": include_events, + "exclude_events": exclude_events, + } + + if recipients: + target_parameters = { + "recipients": recipients + } + target_id = 3 # Target ID for sending to recipients + json_data["target_parameters"] = target_parameters + elif target: + target_id = find_target_id(module, system) + else: + msg = "Neither recipients nor target parameters specified" + module.fail_json(msg=msg) + + json_data["target_id"] = target_id + rule_id = find_rule_id(module, system) + path = f"notifications/rules/{rule_id}" + system.api.put(path=path, data=json_data) + + +def handle_present(module): + """ Create or update a rule """ + system = get_system(module) + name = module.params["name"] + changed = False + if not module.check_mode: + rule_id = find_rule_id(module, system) + if not rule_id: + create_rule(module) + changed = True + msg = f"Rule named {name} created" + else: + update_rule(module) + msg = f"Rule named {name} updated" + changed = True + + module.exit_json(changed=changed, msg=msg) + + +def handle_stat(module): + """ Return rule stat """ + result = None + system = get_system(module) + name = module.params['name'] + rule_id = find_rule_id(module, system) + if rule_id: + path = f"notifications/rules/{rule_id}" + api_result = system.api.get(path=path) + result = api_result.get_json()['result'] + result["rule_id"] = result.pop("id") # Rename id to rule_id + result["msg"] = f"Stat for notification rule named {name}" + result["changed"] = False + module.exit_json(**result) + msg = f"Notification rule {name} not found" + module.fail_json(msg=msg) + + +def handle_absent(module): + """ Remove rule """ + changed = False + name = module.params["name"] + system = get_system(module) + + rule_id = find_rule_id(module, system) + if not rule_id: + msg = f"Rule named {name} not found. Deletion not required." + changed = False + else: + msg = f"Rule named {name} has been deleted" + changed = True + if not module.check_mode: + delete_rule(module) + + module.exit_json(changed=changed, msg=msg) + + +def execute_state(module): + """Determine which state function to execute and do so""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + elif state == "absent": + handle_absent(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + recipients = module.params['recipients'] + target = module.params['target'] + if recipients and target: + msg = "Cannot specify both recipients and target parameters" + module.fail_json(msg=msg) + if recipients: + for recipient in recipients: + if len(recipient) == 1: + msg = f"{recipient} is an invalid email address. Recipients '{recipients}' must be provided as a list, e.g. '[ \"user@example.com\" ]'" + module.fail_json(msg=msg) + if '@' not in recipient: + msg = f"{recipient} is an invalid email address" + module.fail_json(msg=msg) + + +def main(): + """Main module function""" + argument_spec = infinibox_argument_spec() + + argument_spec.update( + { + "name": {"required": True}, + "event_level": {"required": False, "default": [], "type": "list", "elements": "str"}, + "include_events": {"required": False, "default": [], "type": "list", "elements": "str"}, + "exclude_events": {"required": False, "default": [], "type": "list", "elements": "str"}, + "recipients": {"required": False, "default": [], "type": "list", "elements": "str"}, + "target": {"required": False, "type": "str", "default": None}, + "state": {"default": "present", "choices": ["stat", "present", "absent"]}, + } + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_INFINISDK: + module.fail_json(msg=missing_required_lib("infinisdk")) + + check_options(module) + execute_state(module) + + +if __name__ == '__main__': + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_notification_target.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_notification_target.py new file mode 100644 index 000000000..a9e707289 --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_notification_target.py @@ -0,0 +1,361 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +"""This module creates, deletes or modifies metadata on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_notification_target +version_added: 2.13.0 +short_description: Config notification target +description: + - This module configures syslog notification targets on an Infinibox +author: Wei Wang (@wwang) +options: + name: + description: + - Name of the syslog target + type: str + required: true + host: + description: + - Host name or IP address of the target + type: str + required: false + port: + description: + - Port of the target + type: int + required: false + default: 514 + transport: + description: + - TCP or UDP + type: str + required: false + choices: + - UDP + - TCP + default: UDP + protocol: + description: + - Protocol used for this target. Currently, the only valid value is SYSLOG. + type: str + required: false + choices: + - SYSLOG + default: SYSLOG + facility: + description: + - Facility + choices: + - LOCAL0 + - LOCAL1 + - LOCAL2 + - LOCAL3 + - LOCAL4 + - LOCAL5 + - LOCAL6 + - LOCAL7 + type: str + required: false + default: LOCAL7 + visibility: + description: + - Visibility + type: str + choices: + - CUSTOMER + - INFINIDAT + required: false + default: CUSTOMER + post_test: + description: + - Run a test after new target is created + type: bool + required: false + default: true + state: + description: + - Query or modifies target + type: str + required: false + default: present + choices: [ "stat", "present", "absent" ] + +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Create notification targets + infini_notification_target: + state: present + name: testgraylog1 + protocol: SYSLOG + host: 172.31.77.214 + port: 8067 + facility: LOCAL7 + transport: TCP + visibility: CUSTOMER + post_test: true + user: "{{ user }}" + password: "{{ password }}" + system: "{{ system }}" +- name: Create a new notification rule to a target + infini_notification_rule: + name: "test-rule-to-target" # this need to be uniq + event_level: + - ERROR + - CRITICAL + include_events: + - ACTIVATION_PAUSED + exclude_events: + - ACTIVE_DIRECTORY_ALL_DOMAIN_CONTROLLERS_DOWN + - ACTIVE_DIRECTORY_LEFT + target: testgraylog1 + state: "present" + user: "{{ user }}" + password: "{{ password }}" + system: "{{ system }}" +""" + +# RETURN = r''' # ''' + +# -*- coding: utf-8 -*- +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + HAS_INFINISDK, + api_wrapper, + infinibox_argument_spec, + get_system, + merge_two_dicts, +) + +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + pass # Handled by HAS_INFINISDK from module_utils + + +@api_wrapper +def get_target(module): + """ + Find and return config setting value + Use disable_fail when we are looking for config + and it may or may not exist and neither case is an error. + """ + name = module.params['name'] + path = f"notifications/targets?name={name}" + system = get_system(module) + + try: + target = system.api.get(path=path) + except APICommandFailed as err: + msg = f"Cannot find notification target {name}: {err}" + module.fail_json(msg=msg) + + if not target: + msg = f"Users repository {name} not found. Cannot stat." + module.fail_json(msg=msg) + result = target.get_result() + return result + + +def handle_stat(module): + """Return config stat""" + name = module.params['name'] + try: + result = get_target(module)[0] + except IndexError: + module.fail_json(f"Cannot stat notification target {name}. Target not found.") + result2 = { + "changed": False, + "msg": f"Found notification target {name}", + } + result = merge_two_dicts(result, result2) + module.exit_json(**result) + + +@api_wrapper +def find_target_id(module, system): + """ Find the ID of the target by name """ + target_name = module.params["name"] + + try: + path = f"notifications/targets?name={target_name}&fields=id" + api_result = system.api.get(path=path) + except APICommandFailed as err: + msg = f"Cannot find ID for notification target {target_name}: {err}" + module.fail_json(msg=msg) + + if len(api_result.get_json()['result']) > 0: + result = api_result.get_json()['result'][0] + target_id = result['id'] + else: + target_id = None + return target_id + + +@api_wrapper +def delete_target(module): + """ Delete a notification target """ + system = get_system(module) + name = module.params["name"] + target_id = find_target_id(module, system) + + try: + path = f"notifications/targets/{target_id}?approved=true" + system.api.delete(path=path) + except APICommandFailed as err: + msg = f"Cannot delete notification target {name}: {err}" + module.fail_json(msg=msg) + + +@api_wrapper +def create_target(module): + """ Create a new notifition target """ + system = get_system(module) + name = module.params["name"] + protocol = module.params["protocol"] + host = module.params["host"] + port = module.params["port"] + facility = module.params["facility"] + transport = module.params["transport"] + post_test = module.params["post_test"] + visibility = module.params["visibility"] + + path = "notifications/targets" + + json_data = { + "name": name, + "protocol": protocol, + "host": host, + "port": port, + "facility": facility, + "transport": transport, + "visibility": visibility + } + + try: + system.api.post(path=path, data=json_data) + except APICommandFailed as err: + msg = f"Cannot create notification target {name}: {err}" + module.fail_json(msg=msg) + + if post_test: + target_id = find_target_id(module, system) + path = f"notifications/targets/{target_id}/test" + json_data = {} + try: + system.api.post(path=path, data=json_data) + except APICommandFailed as err: + msg = f"Cannot test notification target {name}: {err}" + module.fail_json(msg=msg) + + +@api_wrapper +def update_target(module): + """ Update an existing target. """ + delete_target(module) + create_target(module) + + +def handle_present(module): + """Make config present""" + system = get_system(module) + name = module.params["name"] + changed = False + if not module.check_mode: + target_id = find_target_id(module, system) + if not target_id: + create_target(module) + msg = f"Target {name} created" + else: + update_target(module) + msg = f"Target {name} deleted and recreated" + changed = True + module.exit_json(changed=changed, msg=msg) + + +def handle_absent(module): + """Make config present""" + changed = False + name = module.params["name"] + system = get_system(module) + target_id = find_target_id(module, system) + + if not target_id: + msg = f"Target {name} already does not exist" + changed = False + else: + msg = f"Target {name} has been deleted" + if not module.check_mode: + changed = True + delete_target(module) + + module.exit_json(changed=changed, msg=msg) + + +def execute_state(module): + """ Determine which state function to execute and do so """ + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + elif state == "absent": + handle_absent(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """ Verify module options are sane """ + if module.params['protocol'] != "SYSLOG": + module.fail_json(msg="The only supported protocol is SYSLOG") + + +def main(): + """ Main """ + argument_spec = infinibox_argument_spec() + + argument_spec.update( + { + "name": {"required": True}, + "host": {"required": False}, + "port": {"required": False, "type": "int", "default": 514}, + "transport": {"required": False, "default": "UDP", "choices": ["UDP", "TCP"]}, + "protocol": {"required": False, "default": "SYSLOG", "choices": ["SYSLOG"]}, + "facility": {"required": False, "default": "LOCAL7", "choices": ["LOCAL0", "LOCAL1", "LOCAL2", "LOCAL3", "LOCAL4", "LOCAL5", "LOCAL6", "LOCAL7"]}, + "visibility": {"required": False, "default": "CUSTOMER", "choices": ["CUSTOMER", "INFINIDAT"]}, + "post_test": {"required": False, "default": True, "type": "bool"}, + "state": {"default": "present", "choices": ["stat", "present", "absent"]}, + } + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_INFINISDK: + module.fail_json(msg=missing_required_lib("infinisdk")) + + check_options(module) + execute_state(module) + + +if __name__ == '__main__': + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_pool.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_pool.py index d02657a19..43daa71be 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_pool.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_pool.py @@ -1,7 +1,11 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +"""This module creates, deletes or modifies pools on Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import (absolute_import, division, print_function) @@ -77,7 +81,7 @@ EXAMPLES = r''' - name: Disable SSD Cache on pool infini_pool: name: foo - ssd_cache: no + ssd_cache: false user: admin password: secret system: ibox001 @@ -85,7 +89,7 @@ EXAMPLES = r''' - name: Disable Compression on pool infini_pool: name: foo - compression: no + compression: false user: admin password: secret system: ibox001 @@ -95,7 +99,6 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, api_wrapper, @@ -114,7 +117,7 @@ except ImportError: @api_wrapper def create_pool(module, system): - """Create Pool""" + """ Create Pool """ name = module.params['name'] size = module.params['size'] vsize = module.params['vsize'] @@ -141,8 +144,8 @@ def create_pool(module, system): @api_wrapper -def update_pool(module, system, pool): - """Update Pool""" +def update_pool(module, pool): + """ Update Pool """ changed = False size = module.params['size'] @@ -184,23 +187,21 @@ def update_pool(module, system, pool): @api_wrapper def delete_pool(module, pool): - """Delete Pool""" + """ Delete Pool """ if not module.check_mode: pool.delete() msg = 'Pool deleted' module.exit_json(changed=True, msg=msg) -def get_sys_pool(module): +def handle_stat(module): + """ Show details about a pool """ system = get_system(module) pool = get_pool(module, system) - return (system, pool) - -def handle_stat(module): - system, pool = get_sys_pool(module) + name = module.params['name'] if not pool: - module.fail_json(msg='Pool {0} not found'.format(module.params['name'])) + module.fail_json(msg=f'Pool {name} not found') fields = pool.get_fields() # print('fields: {0}'.format(fields)) free_physical_capacity = fields.get('free_physical_capacity', None) @@ -216,17 +217,21 @@ def handle_stat(module): def handle_present(module): - system, pool = get_sys_pool(module) + """ Create pool """ + system = get_system(module) + pool = get_pool(module, system) if not pool: create_pool(module, system) module.exit_json(changed=True, msg="Pool created") else: - changed = update_pool(module, system, pool) + changed = update_pool(module, pool) module.exit_json(changed=changed, msg="Pool updated") def handle_absent(module): - system, pool = get_sys_pool(module) + """ Remove pool """ + system = get_system(module) + pool = get_pool(module, system) if not pool: module.exit_json(changed=False, msg="Pool already absent") else: @@ -235,6 +240,7 @@ def handle_absent(module): def execute_state(module): + """Determine which state function to execute and do so""" state = module.params['state'] try: if state == 'stat': @@ -244,13 +250,14 @@ def execute_state(module): elif state == 'absent': handle_absent(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( @@ -274,13 +281,13 @@ def main(): if module.params['size']: try: Capacity(module.params['size']) - except Exception: + except Exception: # pylint: disable=broad-exception-caught module.fail_json(msg='size (Physical Capacity) should be defined in MB, GB, TB or PB units') if module.params['vsize']: try: Capacity(module.params['vsize']) - except Exception: + except Exception: # pylint: disable=broad-exception-caught module.fail_json(msg='vsize (Virtual Capacity) should be defined in MB, GB, TB or PB units') execute_state(module) diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_port.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_port.py index 303127260..6502045a9 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_port.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_port.py @@ -1,7 +1,11 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-list-literal,use-dict-literal,line-too-long,wrong-import-position,multiple-statements + +"""This module manages ports on an Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import (absolute_import, division, print_function) @@ -11,7 +15,7 @@ __metaclass__ = type DOCUMENTATION = r''' --- module: infini_port -version_added: '2.9.0' +version_added: 2.9.0 short_description: Add and Delete fiber channel and iSCSI ports to a host on Infinibox description: - This module adds or deletes fiber channel or iSCSI ports to hosts on @@ -21,6 +25,7 @@ options: host: description: - Host Name + type: str required: true state: description: @@ -57,7 +62,7 @@ EXAMPLES = r''' - "00:00:00:00:00:00:00" - "11:11:11:11:11:11:11" iqns: - - "iqn.yyyy-mm.reverse-domain:unique-string" + - "iqn.yyyy-mm.reverse-domain:unique-string" system: ibox01 user: admin password: secret @@ -67,8 +72,6 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, api_wrapper, @@ -135,6 +138,7 @@ def delete_ports(module, system): def get_sys_host(module): + """ Get parameters """ system = get_system(module) host = get_host(module, system) return (system, host) @@ -168,12 +172,12 @@ def find_host_initiators_data(module, system, host, initiator_type): Only include desired initiator keys for each initiator. Return the filtered and edited host initiator list. """ - request = "initiators?page=1&page_size=1000&host_id={0}".format(host.id) + request = f"initiators?page=1&page_size=1000&host_id={host.id}" # print("\nrequest:", request, "initiator_type:", initiator_type) get_initiators_result = system.api.get(request, check_version=False) result_code = get_initiators_result.status_code if result_code != 200: - msg = "get initiators REST call failed. code: {0}".format(result_code) + msg = f"get initiators REST call failed. code: {result_code}" module.fail_json(msg=msg) # Only return initiators of the desired type. @@ -208,7 +212,7 @@ def find_host_initiators_data(module, system, host, initiator_type): return host_initiators_by_type -def get_port_fields(module, system, host): +def get_port_fields(module, system, host): # pylint: disable=too-many-locals """ Return a dict with desired fields from FC and ISCSI ports associated with the host. """ @@ -297,13 +301,12 @@ def handle_stat(module): Return json with status. """ system, host = get_sys_host(module) - host_name = module.params["host"] if not host: - module.fail_json(msg="Host {0} not found".format(host_name)) + module.fail_json(msg=f"Host {host_name} not found") field_dict = get_port_fields(module, system, host) - result = dict(changed=False, msg="Host {0} ports found".format(host_name),) + result = dict(changed=False, msg=f"Host {host_name} ports found") result = merge_two_dicts(result, field_dict) module.exit_json(**result) @@ -313,16 +316,15 @@ def handle_present(module): Handle present state. Fail if host is None. """ system, host = get_sys_host(module) - host_name = module.params["host"] if not host: - module.fail_json(msg="Host {0} not found".format(host_name)) + module.fail_json(msg=f"Host {host_name} not found") changed = update_ports(module, system) if changed: - msg = "Mapping created for host {0}".format(host.get_name()) + msg = f"Mapping created for host {host_name}" else: - msg = "No mapping changes were required for host {0}".format(host.get_name()) + msg = f"No mapping changes were required for host {host_name}" result = dict(changed=changed, msg=msg,) module.exit_json(**result) @@ -333,18 +335,17 @@ def handle_absent(module): Handle absent state. Fail if host is None. """ system, host = get_sys_host(module) + host_name = module.params["host"] if not host: module.exit_json( - changed=False, msg="Host {0} not found".format(module.params["host"]) + changed=False, msg=f"Host {host_name} not found" ) changed = delete_ports(module, system) if changed: - msg = "Mapping removed from host {0}".format(host.get_name()) + msg = f"Mapping removed from host {host_name}" else: - msg = "No mapping changes were required. Mapping already removed from host {0}".format( - host.get_name() - ) + msg = f"No mapping changes were required. Mapping already removed from host {host_name}" result = dict(changed=changed, msg=msg,) module.exit_json(**result) @@ -364,26 +365,21 @@ def execute_state(module): handle_absent(module) else: module.fail_json( - msg="Internal handler error. Invalid state: {0}".format(state) + msg=f"Internal handler error. Invalid state: {state}" ) finally: system = get_system(module) system.logout() -def check_options(module): - pass - - def main(): """ Gather auguments and manage mapping of vols to hosts. """ argument_spec = infinibox_argument_spec() - null_list = list() argument_spec.update( dict( - host=dict(required=True, type=str), + host=dict(required=True, type="str"), state=dict(default="present", choices=["stat", "present", "absent"]), wwns=dict(type="list", elements="str", default=list()), iqns=dict(type="list", elements="str", default=list()), @@ -395,7 +391,6 @@ def main(): if not HAS_INFINISDK: module.fail_json(msg=missing_required_lib("infinisdk")) - check_options(module) execute_state(module) diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_sso.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_sso.py new file mode 100644 index 000000000..66219e08b --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_sso.py @@ -0,0 +1,299 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +""" Manage SSO """ + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_sso +version_added: 2.16.0 +short_description: Configures or queries SSO on Infinibox +description: + - This module configures (present state) or gets information about (absent state) SSO on Infinibox +author: David Ohlemacher (@ohlemacher) +options: + name: + description: + - Sets a name to reference the SSO by. + required: true + type: str + issuer: + description: + - URI of the SSO issuer. + required: false + type: str + sign_on_url: + description: + - URL for sign on. + type: str + required: false + signed_assertion: + description: + - Signed assertion + type: bool + required: false + default: false + signed_response: + description: + - Signed response + required: false + type: bool + default: false + signing_certificate: + description: + - Signing certificate content. + type: str + required: false + enabled: + description: + - Determines if the SSO is enabled. + required: false + default: true + type: bool + state: + description: + - Creates/Modifies the SSO, when using state present. + - For state absent, the SSO is removed. + - State stat shows the existing SSO's details. + type: str + required: false + default: present + choices: [ "stat", "present", "absent" ] +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Configure SSO + infini_sso: + name: OKTA + enabled: true + issuer: "http://www.okta.com/eykRra384o32rrTs" + sign_on_url: "https://infinidat.okta.com/app/infinidat_psus/exkra32oyyU6KCUCk2p7/sso/saml" + state: present + user: admin + password: secret + system: ibox001 + +- name: Stat SSO + infini_sso: + name: OKTA + state: stat + user: admin + password: secret + system: ibox001 + +- name: Clear SSO configuration + infini_sso: + state: absent + user: admin + password: secret + system: ibox001 +""" + +# RETURN = r''' # ''' + +from ansible.module_utils.basic import AnsibleModule + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + api_wrapper, + merge_two_dicts, + get_system, + infinibox_argument_spec, +) + +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + pass # Handled by HAS_INFINISDK from module_utils + + +@api_wrapper +def find_sso(module, name): + """ Find a SSO using its name """ + path = f"config/sso/idps?name={name}" + + try: + system = get_system(module) + sso_result = system.api.get(path=path).get_result() + except APICommandFailed as err: + msg = f"Cannot find SSO identity provider {name}: {err}" + module.fail_json(msg=msg) + + return sso_result + + +def handle_stat(module): + """ Handle the stat state """ + name = module.params["name"] + sso_result = find_sso(module, name) + if not sso_result: + msg = f"SSO identity provider {name} not found. Cannot stat." + module.fail_json(msg=msg) + + result = dict( + changed=False, + msg=f"SSO identity provider {name} stat found" + ) + + result = merge_two_dicts(result, sso_result[0]) + result['signing_certificate'] = "redacted" + module.exit_json(**result) + + +def handle_present(module): # pylint: disable=too-many-locals + """ Handle the present state """ + enabled = module.params['enabled'] + issuer = module.params['issuer'] + sign_on_url = module.params['sign_on_url'] + signed_assertion = module.params['signed_assertion'] + signed_response = module.params['signed_response'] + signing_certificate = module.params['signing_certificate'] + name = module.params['name'] + + existing_sso = find_sso(module, name) + if existing_sso: + existing_sso_id = existing_sso[0]['id'] + delete_sso(module, existing_sso_id) + + path = "config/sso/idps" + data = { + "enabled": enabled, + "issuer": issuer, + "name": name, + "sign_on_url": sign_on_url, + "signed_assertion": signed_assertion, + "signed_response": signed_response, + "signing_certificate": signing_certificate, + } + + try: + system = get_system(module) + sso_result = system.api.post(path=path, data=data).get_result() + except APICommandFailed as err: + msg = f"Cannot configure SSO identity provider named {name}: {err}" + module.fail_json(msg=msg) + + if not existing_sso: + msg = f"SSO identity provider named {name} successfully configured" + else: + msg = f"SSO identity provider named {name} successfully removed and recreated with updated parameters" + result = dict( + changed=True, + msg=msg, + ) + result = merge_two_dicts(result, sso_result) + result['signing_certificate'] = "redacted" + + module.exit_json(**result) + + +def delete_sso(module, sso_id): + """ Delete a SSO. Reference its ID. """ + path = f"config/sso/idps/{sso_id}" + name = module.params["name"] + try: + system = get_system(module) + sso_result = system.api.delete(path=path).get_result() + except APICommandFailed as err: + msg = f"Cannot delete SSO identity provider {name}: {err}" + module.fail_json(msg=msg) + return sso_result + + +def handle_absent(module): + """ Handle the absent state """ + name = module.params["name"] + found_sso = find_sso(module, name) + if not found_sso: + result = dict( + changed=False, + msg=f"SSO {name} already not found" + ) + module.exit_json(**result) + + sso_id = found_sso[0]['id'] + sso_result = delete_sso(module, sso_id) + + if not sso_result: + msg = f"SSO identity provider named {name} with ID {sso_id} not found. Cannot delete." + module.fail_json(msg=msg) + + result = dict( + changed=True, + msg=f"SSO identity provider named {name} deleted" + ) + + result = merge_two_dicts(result, sso_result) + result['signing_certificate'] = "redacted" + module.exit_json(**result) + + +def execute_state(module): + """Handle states""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + elif state == "absent": + handle_absent(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + signing_certificate = module.params["signing_certificate"] + sign_on_url = module.params["sign_on_url"] + state = module.params["state"] + is_failed = False + msg = "" + if state in ["present"]: + if not sign_on_url: + msg += "A sign_on_url parameter must be provided. " + is_failed = True + if not signing_certificate: + msg += "A signing_certificate parameter must be provided. " + is_failed = True + if is_failed: + module.fail_json(msg=msg) + + +def main(): + """ Main """ + argument_spec = infinibox_argument_spec() + argument_spec.update( + dict( + enabled=dict(required=False, type="bool", default=True), + issuer=dict(required=False, default=None), + name=dict(required=True), + sign_on_url=dict(required=False, default=None), + signed_assertion=dict(required=False, type="bool", default=False), + signed_response=dict(required=False, type="bool", default=False), + signing_certificate=dict(required=False, default=None, no_log=True), + state=dict(default="present", choices=["stat", "present", "absent"]), + ) + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + check_options(module) + execute_state(module) + + +if __name__ == "__main__": + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_user.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_user.py index 01bcd0a5f..1915e5d87 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_user.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_user.py @@ -1,16 +1,21 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) +""" Manage Infinibox users """ + +# pylint: disable=invalid-name,use-dict-literal,line-too-long,wrong-import-position + from __future__ import (absolute_import, division, print_function) + __metaclass__ = type DOCUMENTATION = r''' --- module: infini_user -version_added: '2.9.0' +version_added: 2.9.0 short_description: Create, Delete and Modify a User on Infinibox description: - This module creates, deletes or modifies a user on Infinibox. @@ -21,7 +26,7 @@ options: - The new user's Name. Once a user is created, the user_name may not be changed from this module. It may be changed from the UI or from infinishell. - required: true + required: false type: str user_email: description: @@ -52,12 +57,42 @@ options: type: str state: description: - - Creates/Modifies user when present or removes when absent + - Creates/Modifies user when present or removes when absent. + - Use state 'login' to test user credentials. + - Use state 'reset' to reset a user password. required: false default: present - choices: [ "stat", "reset_password", "present", "absent" ] + choices: [ "stat", "reset_password", "present", "absent", "login" ] type: str + user_ldap_group_name: + description: + - Name of the LDAP user group + required: false + type: str + user_ldap_group_dn: + description: + - DN of the LDAP user group + required: false + type: str + user_ldap_group_ldap: + description: + - Name of the LDAP + required: false + type: str + user_ldap_group_role: + description: + - Role for the LDAP user group + choices: [ "admin", "pool_admin", "read_only" ] + required: false + type: str + user_ldap_group_pools: + description: + - A list of existing pools managed by the LDAP user group + default: [] + required: false + type: list + elements: str extends_documentation_fragment: - infinibox ''' @@ -81,27 +116,57 @@ EXAMPLES = r''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( - HAS_INFINISDK, api_wrapper, infinibox_argument_spec, get_system, get_user, - get_pool, - unixMillisecondsToDate, merge_two_dicts, ) + +HAS_INFINISDK = True try: - from infi.dtypes.iqn import make_iscsi_name + from infinisdk.core.exceptions import APICommandFailed except ImportError: - pass # Handled by HAS_INFINISDK from module_utils + HAS_INFINISDK = False + + +@api_wrapper +def find_user_ldap_group_id(module): + """ + Find the ID of the LDAP user group by name + """ + ldap_id = None + ldap_name = module.params["user_ldap_group_name"] + path = f"users?name={ldap_name}&type=eq%3ALdap" + system = get_system(module) + api_result = system.api.get(path=path) + if len(api_result.get_json()['result']) > 0: + result = api_result.get_json()['result'][0] + ldap_id = result['id'] + return ldap_id + + +@api_wrapper +def find_ldap_id(module): + """ + Find the ID of the LDAP by name + """ + ldap_id = None + ldap_name = module.params["user_ldap_group_ldap"] + path = f"config/ldap?name={ldap_name}&fields=id" + system = get_system(module) + api_result = system.api.get(path=path) + if len(api_result.get_json()['result']) > 0: + result = api_result.get_json()['result'][0] + ldap_id = result['id'] + return ldap_id @api_wrapper def create_user(module, system): + """ Create user """ if not module.check_mode: user = system.users.create(name=module.params['user_name'], password=module.params['user_password'], @@ -120,6 +185,42 @@ def create_user(module, system): return changed +@api_wrapper +def create_ldap_user_group(module): + """ Create ldap user group """ + ldap_group_name = module.params['user_ldap_group_name'] + ldap_name = module.params['user_ldap_group_ldap'] + ldap_id = find_ldap_id(module) + ldap_pools = module.params['user_ldap_group_pools'] + if not ldap_id: + msg = f'Cannot create LDAP group {ldap_group_name}. Cannot find ID for LDAP name {ldap_name}' + module.fail_json(msg=msg) + path = "users" + system = get_system(module) + data = { + "name": ldap_group_name, + "dn": module.params['user_ldap_group_dn'], + "ldap_id": ldap_id, + "role": module.params['user_ldap_group_role'], + "type": "Ldap", + } + try: + system.api.post(path=path, data=data) + except APICommandFailed as err: + if err.status_code in [409]: + msg = f'Cannot create user_ldap_group_name {ldap_group_name}: {err.message}' + module.fail_json(msg) + changed = True + + user = get_user(module, system, ldap_group_name) + for pool_name in ldap_pools: + # Pylint is not finding Infinibox.pools but Python does. + pool = system.pools.get(name=pool_name) # pylint: disable=no-member + add_user_to_pool_owners(user, pool) + + return changed + + def add_user_to_pool_owners(user, pool): """ Find the current list of pool owners and add user using pool.set_owners(). @@ -127,27 +228,20 @@ def add_user_to_pool_owners(user, pool): get owners, add user, then set owners. Further, we need to know if the owners changed. Use sets of owners to compare. """ - # print("add_user_to_pool_owners(): start") changed = False pool_fields = pool.get_fields(from_cache=True, raw_value=True) pool_owners = pool_fields.get('owners', []) - # print('pool_owners:', pool_owners, 'pool_owners type:', type(pool_owners)) - # print('user:', user) - # print('pool:', pool) pool_owners_set = set(pool_owners) - # print('pool_owners_set:', pool_owners_set) new_pool_owners_set = pool_owners_set.copy() new_pool_owners_set.add(user.id) - # print('new_pool_owners_set:', new_pool_owners_set) if pool_owners_set != new_pool_owners_set: pool.set_owners([user]) changed = True - # print("changed:", changed) - # print("add_user_to_pool_owners(): end") return changed def remove_user_from_pool_owners(user, pool): + """ Remove user from pool owners """ changed = False pool_fields = pool.get_fields(from_cache=True, raw_value=True) pool_owners = pool_fields.get('owners', []) @@ -162,9 +256,9 @@ def remove_user_from_pool_owners(user, pool): @api_wrapper def update_user(module, system, user): - # print("update_user()") + """ Update user """ if user is None: - raise AssertionError("Cannot update user {0}. User not found.".format(module.params["user_name"])) + raise AssertionError(f"Cannot update user {module.params['user_name']}. User not found.") changed = False fields = user.get_fields(from_cache=True, raw_value=True) @@ -182,23 +276,34 @@ def update_user(module, system, user): try: pool_name = module.params['user_pool'] pool = system.pools.get(name=pool_name) - except Exception as err: - module.fail_json(msg='Cannot find pool {0}: {1}'.format(pool_name, err)) + except Exception as err: # pylint: disable=broad-exception-caught + module.fail_json(msg=f'Cannot find pool {pool_name}: {err}') if add_user_to_pool_owners(user, pool): changed = True return changed +def update_ldap_user_group(module): + """ Update ldap user group by deleting and creating the LDAP user""" + changed = delete_ldap_user_group(module) + if not changed: + module.fail_json(msg='Cannot delete LDAP user {ldap_group_name}. Cannot find ID for LDAP group.') + create_ldap_user_group(module) + changed = True + return changed + + @api_wrapper -def reset_user_password(module, system, user): - # print("update_user()") +def reset_user_password(module, user): + """ Reset user's password """ if user is None: - raise AssertionError("Cannot change user {0} password. User not found.".format(module.params["user_name"])) + module.fail_json(msg=f'Cannot change user {module.params["user_name"]} password. User not found.') user.update_password(module.params['user_password']) @api_wrapper def delete_user(module, user): + """ Delete a user """ if not user: return False @@ -209,82 +314,196 @@ def delete_user(module, user): return changed -def get_sys_user(module): +@api_wrapper +def delete_ldap_user_group(module): + """ Delete a ldap user group """ + changed = False + ldap_group_name = module.params['user_ldap_group_name'] + ldap_group_id = find_user_ldap_group_id(module) + if not ldap_group_id: + changed = False + return changed + path = f"users/{ldap_group_id}?approved=yes" system = get_system(module) - user = get_user(module, system) - # print("get_sys_user(): user:", user) - return (system, user) + try: + system.api.delete(path=path) + changed = True + except APICommandFailed as err: + if err.status_code in [404]: + changed = False + else: + msg = f'An error occurred deleting user_ldap_group_name {ldap_group_name}: {err}' + module.fail_json(msg) + return changed + + +def get_user_ldap_group(module): + """ + Find the LDAP user group by name + """ + result = None + user_ldap_group_name = module.params["user_ldap_group_name"] + path = f"users?name={user_ldap_group_name}&type=eq%3ALdap" + system = get_system(module) + api_result = system.api.get(path=path) + if len(api_result.get_json()['result']) > 0: + result = api_result.get_json()['result'][0] + return result def get_user_fields(user): + """ Get user's fields """ pools = user.get_owned_pools() pool_names = [pool.get_field('name') for pool in pools] fields = user.get_fields(from_cache=True, raw_value=True) - field_dict = dict( - id=user.id, - enabled=fields.get('enabled', None), - role=fields.get('role', None), - email=fields.get('email', None), - pools=pool_names, - ) + field_dict = { + "dn": fields.get('dn', None), + "email": fields.get('email', None), + "enabled": fields.get('enabled', None), + "id": user.id, + "ldap_id": fields.get('ldap_id', None), + "pools": pool_names, + "role": fields.get('role', None), + "roles": fields.get('roles', []), + "type": fields.get('type', None), + } return field_dict def handle_stat(module): - system, user = get_sys_user(module) - user_name = module.params["user_name"] - if not user: - module.fail_json(msg='User {0} not found'.format(user_name)) - field_dict = get_user_fields(user) - result = dict( - changed=False, - msg='User stat found' - ) + """ Handle stat for user or LDAP group user """ + user_name = module.params['user_name'] + user_ldap_group_name = module.params['user_ldap_group_name'] + if user_name: + system = get_system(module) + user = get_user(module, system) + user_name = module.params["user_name"] + if not user: + module.fail_json(msg=f'User {user_name} not found') + field_dict = get_user_fields(user) + msg = f'User {user_name} stat found' + elif user_ldap_group_name: + user = get_user_ldap_group(module) + if not user: + module.fail_json(msg=f'user_ldap_group_name {user_ldap_group_name} not found') + field_dict = get_user_fields(user) + msg = f'User LDAP group {user_ldap_group_name} stat found' + else: + msg = 'Neither user_name nor user_ldap_group_name were provided for state stat' + module.fail_json(msg) + + result = { + "changed": False, + "msg": msg, + } result = merge_two_dicts(result, field_dict) module.exit_json(**result) def handle_present(module): - system, user = get_sys_user(module) + """ Handle making user present """ user_name = module.params["user_name"] - if not user: - changed = create_user(module, system) - msg = 'User {0} created'.format(user_name) - else: - changed = update_user(module, system, user) - if changed: - msg = 'User {0} updated'.format(user_name) + user_ldap_group_name = module.params["user_ldap_group_name"] + changed = False + msg = 'Message not set' + + if user_name: + system = get_system(module) + user = get_user(module, system) + if not user: + changed = create_user(module, system) + msg = f'User {user_name} created' + else: + changed = update_user(module, system, user) + if changed: + msg = f'User {user_name} updated' + else: + msg = f'User {user_name} update required no changes' + elif user_ldap_group_name: + ldap_user = get_user_ldap_group(module) + if not ldap_user: + changed = create_ldap_user_group(module) + msg = f'LDAP user group {user_ldap_group_name} created' else: - msg = 'User {0} update required no changes'.format(user_name) + changed = update_ldap_user_group(module) + if changed: + msg = f'LDAP user group {user_ldap_group_name} updated by deleting and recreating with updated parameters' + else: + msg = f'LDAP user group {user_ldap_group_name} update not required - no changes' + else: + msg = 'Neither user_name nor user_ldap_group_name were provided' + module.fail_json(msg) + module.exit_json(changed=changed, msg=msg) def handle_absent(module): - system, user = get_sys_user(module) - user_name = module.params["user_name"] - if not user: - changed = False - msg = "User {0} already absent".format(user_name) + """ Handle making user absent """ + user_name = module.params['user_name'] + user_ldap_group_name = module.params['user_ldap_group_name'] + if user_name: + system = get_system(module) + user = get_user(module, system) + user_name = module.params["user_name"] + if not user: + changed = False + msg = f"User {user_name} already absent" + else: + changed = delete_user(module, user) + msg = f"User {user_name} removed" + module.exit_json(changed=changed, msg=msg) + elif user_ldap_group_name: + changed = delete_ldap_user_group(module) + if changed: + msg = f"LDAP group user {user_ldap_group_name} removed" + else: + msg = f"LDAP group user {user_ldap_group_name} already absent" + module.exit_json(changed=changed, msg=msg) else: - changed = delete_user(module, user) - msg = "User {0} removed".format(user_name) - module.exit_json(changed=changed, msg=msg) + msg = 'Neither user_name nor user_ldap_group_name were provided for state absent' + module.fail_json(msg) def handle_reset_password(module): - system, user = get_sys_user(module) + """ Reset user password """ + system = get_system(module) + user = get_user(module, system) user_name = module.params["user_name"] if not user: - msg = 'Cannot change password. User {0} not found'.format(user_name) + msg = f'Cannot change password. User {user_name} not found' module.fail_json(msg=msg) else: - reset_user_password(module, system, user) - msg = 'User {0} password changed'.format(user_name) + reset_user_password(module, user) + msg = f'User {user_name} password changed' module.exit_json(changed=True, msg=msg) +def handle_login(module): + """ Test user credentials by logging in """ + system = get_system(module) + user_name = module.params["user_name"] + user_password = module.params['user_password'] + path = "users/login" + data = { + "username": user_name, + "password": user_password, + } + try: + login = system.api.post(path=path, data=data) + except APICommandFailed: + msg = f'User {user_name} failed to login' + module.fail_json(msg=msg) + if login.status_code == 200: + msg = f'User {user_name} successfully logged in' + module.exit_json(changed=False, msg=msg) + else: + msg = f'User {user_name} failed to login with status code: {login.status_code}' + module.fail_json(msg=msg) + + def execute_state(module): + """ Find state and handle it """ state = module.params['state'] try: if state == 'stat': @@ -295,47 +514,85 @@ def execute_state(module): handle_absent(module) elif state == 'reset_password': handle_reset_password(module) + elif state == 'login': + handle_login(module) else: - module.fail_json(msg='Internal handler error. Invalid state: {0}'.format(state)) + module.fail_json(msg=f'Internal handler error. Invalid state: {state}') finally: system = get_system(module) system.logout() -def check_options(module): +def check_options(module): # pylint: disable=too-many-branches + """ Check option logic """ state = module.params['state'] + user_name = module.params['user_name'] user_role = module.params['user_role'] user_pool = module.params['user_pool'] + user_ldap_group_name = module.params['user_ldap_group_name'] + user_ldap_group_role = module.params['user_ldap_group_role'] if state == 'present': if user_role == 'pool_admin' and not user_pool: module.fail_json(msg='user_role "pool_admin" requires a user_pool to be provided') if user_role != 'pool_admin' and user_pool: module.fail_json(msg='Only user_role "pool_admin" should have a user_pool provided') - valid_keys = ['user_email', 'user_password', 'user_role', 'user_enabled'] - for valid_key in valid_keys: - # Check required keys provided - try: - not_used = module.params[valid_key] - except KeyError: - msg = 'For state "present", options {0} are required'.format(", ".join(valid_keys)) - module.fail_json(msg=msg) - elif state == 'reset_password': - if not module.params['user_password']: - msg = 'For state "reset_password", user_password is required' + if not user_name and not user_ldap_group_name: + msg = 'For state "present", option user_name or user_ldap_group_name is required' + module.fail_json(msg=msg) + + if user_name and user_ldap_group_name: + msg = 'For state "present", option user_name and user_ldap_group_name cannot both be provided' + module.fail_json(msg=msg) + + if user_name: + required_user_params = [ + 'user_email', 'user_password', 'user_role', + ] + for required_param in required_user_params: + param = module.params[required_param] + if param is None: + msg = f"For state 'present', option {required_param} is required with option user_name" + module.fail_json(msg=msg) + + if user_ldap_group_name: + required_user_ldap_params = [ + 'user_ldap_group_dn', 'user_ldap_group_ldap', 'user_ldap_group_role', + ] + for required_param in required_user_ldap_params: + param = module.params[required_param] + if not param: + msg = f'For state "present", option {required_param} is required with option user_ldap_group_name' + module.fail_json(msg=msg) + if user_ldap_group_role == 'pool_admin': + user_ldap_group_pools = module.params['user_ldap_group_pools'] + if not user_ldap_group_pools: + msg = "For state 'present' and user_ldap_group_role 'pool_admin', user_ldap_group_pool must specify one or more pools" + module.fail_json(msg=msg) + + elif state in ['reset_password', 'login']: + if not module.params['user_name'] or not module.params['user_password']: + msg = f"For state '{state}', user_name and user_password are both required" + module.fail_json(msg=msg) def main(): + """ main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( - user_name=dict(required=True), - user_email=dict(required=False), - user_password=dict(required=False, no_log=True), - user_role=dict(required=False, choices=['admin', 'pool_admin', 'read_only']), + user_name=dict(required=False), + user_email=dict(required=False, default=None), + user_password=dict(required=False, no_log=True, default=None), + user_role=dict(required=False, choices=['admin', 'pool_admin', 'read_only'], default=None), user_enabled=dict(required=False, type='bool', default=True), - user_pool=dict(required=False), - state=dict(default='present', choices=['stat', 'reset_password', 'present', 'absent']), + user_pool=dict(required=False, default=None), + user_ldap_group_name=dict(required=False, default=None), + user_ldap_group_dn=dict(required=False, default=None), + user_ldap_group_ldap=dict(required=False, default=None), + user_ldap_group_role=dict(required=False, choices=['admin', 'pool_admin', 'read_only'], default=None), + user_ldap_group_pools=dict(required=False, type='list', elements='str', default=[]), + state=dict(default='present', choices=['stat', 'reset_password', 'present', 'absent', 'login']), ) ) diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_users_repository.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_users_repository.py new file mode 100644 index 000000000..bbc2a8d12 --- /dev/null +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_users_repository.py @@ -0,0 +1,534 @@ +#!/usr/bin/python +# -*- coding: utf-8 -*- + +# pylint: disable=invalid-name,use-list-literal,use-dict-literal,line-too-long,wrong-import-position,multiple-statements + +"""This module creates, deletes or modifies repositories of users that can log on to an Infinibox.""" + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +DOCUMENTATION = r""" +--- +module: infini_users_repository +version_added: 2.13.0 +short_description: Create, Delete or Modify respositories of users that can log on to an Infinibox +description: + - This module creates, deletes or modifies respositories of users that can log on to an Infinibox. +author: David Ohlemacher (@ohlemacher) +options: + ad_auto_discover_servers: + description: + - AD auto discover servers + type: bool + choices: [true, false] + required: false + default: true + ad_domain_name: + description: + - AD domain name + type: str + required: false + bind_password: + description: + - The bind user password + type: str + required: false + bind_username: + description: + - The bind username + type: str + required: false + servers: + description: + - A list of LDAP servers. For an empty list, use []. + required: false + type: list + elements: str + default: [] + name: + description: + - Name of repository + type: str + required: true + ldap_port: + description: + - LDAP or AD port to use + type: int + required: false + default: 636 + ldap_servers: + description: + - List of LDAP or AD servers + type: list + elements: str + required: false + default: [] + repository_type: + description: + - The type of repository + choices: ["ActiveDirectory", "LDAP"] + type: str + required: False + schema_group_memberof_attribute: + description: + - Schema group memberof attribute + type: str + required: false + schema_group_name_attribute: + description: + - Schema group name attribute + type: str + required: false + schema_groups_basedn: + description: + - Schema groups base DN + type: str + required: false + schema_group_class: + description: + - Schema group class + type: str + required: false + schema_users_basedn: + description: + - Schema user base DN + type: str + required: false + schema_user_class: + description: + - Schema user class + type: str + required: false + schema_username_attribute: + description: + - Schema username attribute + type: str + required: false + state: + description: + - Creates/Modifies users repositories when present or removes when absent. + - When getting the stats for a users repository, the module will test + connectivity to the repository and report the result in 'test_ok' as true or false. + required: false + type: str + default: present + choices: [ "stat", "present", "absent" ] + use_ldaps: + description: + - Use SSL (LDAPS) + type: bool + choices: ["true", "false"] + default: true + +extends_documentation_fragment: + - infinibox +""" + +EXAMPLES = r""" +- name: Create AD + infini_users_repository: + name: PSUS_ANSIBLE_ad + bind_password: tuFrAxahuYe4 + bind_username: conldap + ad_domain_name: infinidat.com + repository_type: ActiveDirectory + schema_group_class: group + schema_group_memberof_attribute: memberof + schema_group_name_attribute: cn + schema_groups_basedn: + schema_user_class: user + schema_username_attribute: sAMAccountName + state: present + system: 172.20.67.167 + user: dohlemacher + password: 123456 + +- name: Stat AD + infini_users_repository: + name: PSUS_ANSIBLE_ad + state: stat + user: admin + password: secret + system: ibox001 + +- name: Remove AD + infini_users_repository: + name: PSUS_ANSIBLE_ad + state: absent + user: admin + password: secret + system: ibox001 +""" + +# RETURN = r''' # ''' + +from ansible.module_utils.basic import AnsibleModule, missing_required_lib + +from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( + api_wrapper, + get_system, + infinibox_argument_spec, +) + +HAS_INFINISDK = True +try: + from infinisdk.core.exceptions import APICommandFailed +except ImportError: + HAS_INFINISDK = False + + +@api_wrapper +def get_users_repository(module, disable_fail=False): + """ + Find and return users repository information + Use disable_fail when we are looking for an user repository + and it may or may not exist and neither case is an error. + """ + system = get_system(module) + name = module.params["name"] + + path = f"config/ldap?name={name}" + repo = system.api.get(path=path) + + if repo: + result = repo.get_result() + if not disable_fail and not result: + msg = f"Users repository {name} not found. Cannot stat." + module.fail_json(msg=msg) + return result + + if not disable_fail: + msg = f"Users repository {name} not found. Cannot stat." + module.fail_json(msg=msg) + + return None + + +@api_wrapper +def test_users_repository(module, repository_id, disable_fail=False): + """ + Find and return users repository information + Use disable_fail when we are looking for an user repository + and it may or may not exist and neither case is an error. + """ + system = get_system(module) + name = module.params['name'] + try: + path = f"config/ldap/{repository_id}/test" + result = system.api.post(path=path) + except APICommandFailed as err: + if disable_fail: + return False + msg = f"Users repository {name} testing failed: {str(err)}" + module.fail_json(msg=msg) + if result.response.status_code in [200]: + return True + return False + + +def create_post_data(module): + """Create data dict for post rest calls""" + name = module.params["name"] + repo_type = module.params["repository_type"] + # search_order + schema_definition = { + "group_class": module.params["schema_group_class"], + "group_memberof_attribute": module.params["schema_group_memberof_attribute"], + "group_name_attribute": module.params["schema_group_name_attribute"], + "groups_basedn": module.params["schema_groups_basedn"], + "user_class": module.params["schema_user_class"], + "username_attribute": module.params["schema_username_attribute"], + "users_basedn": module.params["schema_users_basedn"], + } + + # Create json data + data = { + "bind_password": module.params["bind_password"], + "bind_username": module.params["bind_username"], + "ldap_port": module.params["ldap_port"], + "name": name, + "repository_type": repo_type, + "schema_definition": schema_definition, + "use_ldaps": module.params["use_ldaps"], + } + + # Add type specific fields to data dict + if repo_type == "ActiveDirectory": + data["domain_name"] = module.params["ad_domain_name"] + data["servers"] = [] + else: # LDAP + data["domain_name"]: None + data["servers"] = module.params["ldap_servers"] + return data + + +@api_wrapper +def post_users_repository(module): + """ + Create or update users LDAP or AD repo. The changed variable is found elsewhere. + Variable 'changed' not returned by design + """ + system = get_system(module) + name = module.params["name"] + data = create_post_data(module) + path = "config/ldap" + try: + system.api.post(path=path, data=data) + except APICommandFailed as err: + if err.error_code == "LDAP_NAME_CONFLICT": + msg = f"Users repository {name} conflicts." + module.fail_json(msg=msg) + elif err.error_code == "LDAP_BAD_CREDENTIALS": + msg = f"Cannot create users repository {name} due to incorrect LDAP credentials: {err}" + module.fail_json(msg=msg) + else: + msg = f"Cannot create users repository {name}: {err}" + module.fail_json(msg=msg) + + +@api_wrapper +def delete_users_repository(module): + """Delete repo.""" + system = get_system(module) + name = module.params['name'] + changed = False + if not module.check_mode: + repo = get_users_repository(module, disable_fail=True) + if repo and len(repo) == 1: + path = f"config/ldap/{repo[0]['id']}" + try: + system.api.delete(path=path) + changed = True + except APICommandFailed as err: + if err.status_code != 404: + msg = f"Deletion of users repository {name} failed: {str(err)}" + module.fail_json(msg=msg) + return changed + + +def handle_stat(module): + """Return users repository stat""" + name = module.params['name'] + repos = get_users_repository(module) + + if len(repos) != 1: + msg = f"Users repository {name} not found in repository list {repos}. Cannot stat." + module.fail_json(msg=msg) + + result = repos[0] + repository_id = result.pop("id") + result["msg"] = f"Stats for user repository {name}" + result["repository_id"] = repository_id # Rename id to repository_id + result["test_ok"] = test_users_repository(module, repository_id=repository_id, disable_fail=True) + result["changed"] = False + module.exit_json(**result) + + +@api_wrapper +def is_existing_users_repo_equal_to_desired(module): # pylint: disable=too-many-return-statements,multiple-statements + """ Compare two user user repositories. Return a bool. """ + newdata = create_post_data(module) + olddata = get_users_repository(module, disable_fail=True)[0] + if not olddata: + return False + if olddata['bind_username'] != newdata['bind_username']: + return False + if olddata['repository_type'] != newdata['repository_type']: + return False + if olddata['domain_name'] != newdata['domain_name']: + return False + if olddata['ldap_port'] != newdata['ldap_port']: + return False + if olddata['name'] != newdata['name']: + return False + if olddata['schema_definition'] != newdata['schema_definition']: + return False + if olddata['servers'] != newdata['servers']: + return False + if olddata['use_ldaps'] != newdata['use_ldaps']: + return False + return True + + +def handle_present(module): + """Make users repository present""" + name = module.params['name'] + changed = False + msg = "" + if not module.check_mode: + old_users_repo = None + old_users_repo_result = get_users_repository(module, disable_fail=True) + if old_users_repo_result: + old_users_repo = old_users_repo_result[0] + if is_existing_users_repo_equal_to_desired(module): + msg = f"Users repository {name} already exists. No changes required." + module.exit_json(changed=changed, msg=msg) + else: + msg = f"Users repository {name} is being recreated with new settings. " + delete_users_repository(module) + old_users_repo = None + changed = True + + post_users_repository(module) + + new_users_repo = get_users_repository(module) + changed = new_users_repo != old_users_repo + if changed: + if old_users_repo: + msg = f"{msg}Users repository {name} updated" + else: + msg = f"{msg}Users repository {name} created" + else: + msg = f"Users repository {name} unchanged since the value is the same as the existing users repository" + else: + msg = f"Users repository {name} unchanged due to check_mode" + module.exit_json(changed=changed, msg=msg) + + +def handle_absent(module): + """Make users repository absent""" + name = module.params['name'] + msg = f"Users repository {name} unchanged" + changed = False + if not module.check_mode: + changed = delete_users_repository(module) + if changed: + msg = f"Users repository {name} removed" + else: + msg = f"Users repository {name} did not exist so removal was unnecessary" + module.exit_json(changed=changed, msg=msg) + + +def execute_state(module): + """Determine which state function to execute and do so""" + state = module.params["state"] + try: + if state == "stat": + handle_stat(module) + elif state == "present": + handle_present(module) + elif state == "absent": + handle_absent(module) + else: + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") + finally: + system = get_system(module) + system.logout() + + +def check_options(module): + """Verify module options are sane""" + # ad_domain_name = module.params["ad_domain_name"] + # bind_password = module.params["bind_password"] + # bind_username = module.params["bind_username"] + # ad_domain_name = module.params["ad_domain_name"] + # ldap_servers = module.params["ldap_servers"] + name = module.params["name"] + # ldap_port = module.params["ldap_port"] + repository_type = module.params["repository_type"] + # schema_group_memberof_attribute = module.params["schema_group_memberof_attribute"] + # schema_group_name_attribute = module.params["schema_group_name_attribute"] + # schema_groups_basedn = module.params["schema_groups_basedn"] + # schema_user_class = module.params["schema_user_class"] + # schema_username_attribute = module.params["schema_username_attribute"] + # schema_users_basedn = module.params["schema_users_basedn"] + state = module.params["state"] + + if state == "stat": + pass + elif state == "present": + if repository_type: + common_params = ["bind_password", "bind_username", "schema_group_class", + "schema_group_memberof_attribute", "schema_group_name_attribute", + "schema_user_class", "schema_username_attribute",] + if repository_type == "LDAP": # Creating an LDAP + req_params = common_params + missing_params = [param for param in req_params if not is_set_in_params(module, param)] + if missing_params: + msg = f"Cannot create a new LDAP repository named {name} without providing required parameters: {missing_params}" + module.fail_json(msg=msg) + + disallowed_params = ["ad_domain_name", "ad_auto_discover_servers"] + error_params = [param for param in disallowed_params if is_set_in_params(module, param)] + if error_params: + msg = f"Cannot create a new LDAP repository named {name} when providing disallowed parameters: {error_params}" + module.fail_json(msg=msg) + elif repository_type == "ActiveDirectory": + req_params = common_params + missing_params = [param for param in req_params if not is_set_in_params(module, param)] + if missing_params: + msg = f"Cannot create a new LDAP repository named {name} without providing required parameters: {missing_params}" + module.fail_json(msg=msg) + + disallowed_params = ["ldap_servers"] + error_params = [param for param in disallowed_params if is_set_in_params(module, param)] + if error_params: + msg = f"Cannot create a new LDAP repository named {name} when providing disallowed parameters: {error_params}" + module.fail_json(msg=msg) + else: + msg = f"Unsupported respository type: {repository_type}" + module.fail_json(msg=msg) + else: + msg = "Cannot create a new users repository without providing a repository_type" + module.fail_json(msg=msg) + elif state == "absent": + pass + else: + module.fail_json(f"Invalid state '{state}' provided") + + +def is_set_in_params(module, key): + """A utility function to test if a module param key is set to a truthy value. + Useful in list comprehensions.""" + is_set = False + try: + if module.params[key]: + is_set = True + except KeyError: + pass + return is_set + + +def main(): + """Main module function""" + argument_spec = infinibox_argument_spec() + + argument_spec.update( + { + "ad_auto_discover_servers": {"required": False, "choices": [True, False], "type": "bool", "default": True}, + "ad_domain_name": {"required": False, "default": None}, + "bind_password": {"required": False, "default": None, "no_log": True}, + "bind_username": {"required": False, "default": None}, + "ldap_servers": {"required": False, "default": [], "type": "list", "elements": "str"}, + "name": {"required": True}, + "ldap_port": {"required": False, "type": "int", "default": 636}, + "repository_type": {"required": False, "choices": ["LDAP", "ActiveDirectory"], "default": None}, + "schema_group_class": {"required": False, "default": None}, + "schema_group_memberof_attribute": {"required": False, "default": None}, + "schema_group_name_attribute": {"required": False, "default": None}, + "schema_groups_basedn": {"required": False, "default": None}, + "schema_user_class": {"required": False, "default": None}, + "schema_username_attribute": {"required": False, "default": None}, + "schema_users_basedn": {"required": False, "default": None}, + "servers": {"required": False, "default": [], "type": "list", "elements": "str"}, + "state": {"default": "present", "choices": ["stat", "present", "absent"]}, + "use_ldaps": {"required": False, "choices": [True, False], "type": "bool", "default": True}, + } + ) + + module = AnsibleModule(argument_spec, supports_check_mode=True) + + if not HAS_INFINISDK: + module.fail_json(msg=missing_required_lib("infinisdk")) + + check_options(module) + execute_state(module) + + +if __name__ == "__main__": + main() diff --git a/ansible_collections/infinidat/infinibox/plugins/modules/infini_vol.py b/ansible_collections/infinidat/infinibox/plugins/modules/infini_vol.py index 0c4a579bc..df5b0e756 100644 --- a/ansible_collections/infinidat/infinibox/plugins/modules/infini_vol.py +++ b/ansible_collections/infinidat/infinibox/plugins/modules/infini_vol.py @@ -1,14 +1,18 @@ #!/usr/bin/python # -*- coding: utf-8 -*- -# Copyright: (c) 2022, Infinidat <info@infinidat.com> +# pylint: disable=invalid-name,use-dict-literal,too-many-branches,too-many-locals,line-too-long,wrong-import-position + +""" A module for managing Infinibox volumes """ + +# Copyright: (c) 2024, Infinidat <info@infinidat.com> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) -from __future__ import (absolute_import, division, print_function) +from __future__ import absolute_import, division, print_function __metaclass__ = type -DOCUMENTATION = r''' +DOCUMENTATION = r""" --- module: infini_vol version_added: '2.3.0' @@ -19,24 +23,34 @@ author: David Ohlemacher (@ohlemacher) options: name: description: - - Volume Name - required: true + - Volume name. + type: str + required: false + serial: + description: + - Volume serial number. + type: str + required: false parent_volume_name: description: - Specify a volume name. This is the volume parent for creating a snapshot. Required if volume_type is snapshot. + type: str required: false pool: description: - Pool that master volume will reside within. Required for creating a master volume, but not a snapshot. + type: str required: false size: description: - Volume size in MB, GB or TB units. Required for creating a master volume, but not a snapshot + type: str required: false snapshot_lock_expires_at: description: - This will cause a snapshot to be locked at the specified date-time. Uses python's datetime format YYYY-mm-dd HH:MM:SS.ffffff, e.g. 2020-02-13 16:21:59.699700 + type: str required: false snapshot_lock_only: description: @@ -47,6 +61,7 @@ options: state: description: - Creates/Modifies master volume or snapshot when present or removes when absent. + type: str required: false default: present choices: [ "stat", "present", "absent" ] @@ -59,12 +74,14 @@ options: write_protected: description: - Specifies if the volume should be write protected. Default will be True for snapshots, False for regular volumes. + type: str required: false default: "Default" choices: ["Default", "True", "False"] volume_type: description: - Specifies the volume type, regular volume or snapshot. + type: str required: false default: master choices: [ "master", "snapshot" ] @@ -72,21 +89,23 @@ options: description: - Specify true to restore a volume (parent_volume_name) from an existing snapshot specified by the name field. - State must be set to present and volume_type must be 'snapshot'. + type: bool required: false default: false + extends_documentation_fragment: - infinibox requirements: - capacity -''' +""" -EXAMPLES = r''' +EXAMPLES = r""" - name: Create new volume named foo under pool named bar infini_vol: name: foo # volume_type: master # Default size: 1TB - thin_provision: yes + thin_provision: true pool: bar state: present user: admin @@ -115,25 +134,30 @@ EXAMPLES = r''' user: admin password: secret system: ibox001 -''' +""" # RETURN = r''' # ''' from ansible.module_utils.basic import AnsibleModule, missing_required_lib -import traceback - from ansible_collections.infinidat.infinibox.plugins.module_utils.infinibox import ( HAS_INFINISDK, api_wrapper, - infinibox_argument_spec, - ObjectNotFound, + check_snapshot_lock_options, get_pool, get_system, + get_vol_by_sn, get_volume, - get_vol_sn, + infinibox_argument_spec, + manage_snapshot_locks, ) +HAS_INFINISDK = True +try: + from infinisdk.core.exceptions import APICommandFailed + from infinisdk.core.exceptions import ObjectNotFound +except ImportError: + HAS_INFINISDK = False HAS_CAPACITY = True try: @@ -141,19 +165,10 @@ try: except ImportError: HAS_CAPACITY = False -HAS_ARROW = True -try: - import arrow -except ImportError: - HAS_ARROW = False - -except Exception: - HAS_INFINISDK = False - @api_wrapper def create_volume(module, system): - """Create Volume""" + """ Create Volume """ changed = False if not module.check_mode: if module.params["thin_provision"]: @@ -179,24 +194,21 @@ def create_volume(module, system): @api_wrapper def find_vol_id(module, system, vol_name): - """ - Find the ID of this vol - """ - vol_url = "volumes?name={0}&fields=id".format(vol_name) + """ Find the ID of this vol """ + vol_url = f"volumes?name={vol_name}&fields=id" vol = system.api.get(path=vol_url) result = vol.get_json()["result"] if len(result) != 1: - module.fail_json("Cannot find a volume with name '{0}'".format(vol_name)) + module.fail_json(f"Cannot find a volume with name '{vol_name}'") vol_id = result[0]["id"] - # print("Volume {} has ID {}".format(vol_name, vol_id)) return vol_id @api_wrapper def restore_volume_from_snapshot(module, system): - """Use snapshot to restore a volume""" + """ Use snapshot to restore a volume """ changed = False is_restoring = module.params["restore_volume_from_snapshot"] volume_type = module.params["volume_type"] @@ -209,62 +221,62 @@ def restore_volume_from_snapshot(module, system): if not is_restoring: raise AssertionError("A programming error occurred. is_restoring is not True") if volume_type != "snapshot": - module.fail_json( - msg="Cannot restore a parent volume from snapshot unless the volume " - "type is 'snapshot'" - ) + module.exit_json(msg="Cannot restore a parent volume from snapshot unless the volume type is 'snapshot'") if not parent_volume_name: - module.fail_json( - msg="Cannot restore a parent volume from snapshot unless the parent " - "volume name is specified" - ) + module.exit_json(msg="Cannot restore a parent volume from snapshot unless the parent volume name is specified") if not module.check_mode: - restore_url = "volumes/{0}/restore?approved=true".format(parent_volume_id) + restore_url = f"volumes/{parent_volume_id}/restore?approved=true" restore_data = { "source_id": snap_id, } - restore = system.api.post(path=restore_url, data=restore_data) - changed = True + try: + system.api.post(path=restore_url, data=restore_data) + changed = True + except APICommandFailed as err: + module.fail_json(msg=f"Cannot restore volume {parent_volume_name} from {snap_name}: {err}") return changed @api_wrapper def update_volume(module, volume): - """Update Volume""" + """ Update Volume """ changed = False + + if module.check_mode: + return changed + if module.params["size"]: size = Capacity(module.params["size"]).roundup(64 * KiB) if volume.get_size() != size: - if not module.check_mode: - volume.update_size(size) + volume.update_size(size) changed = True if module.params["thin_provision"] is not None: - type = str(volume.get_provisioning()) - if type == "THICK" and module.params["thin_provision"]: - if not module.check_mode: - volume.update_provisioning("THIN") + provisioning = str(volume.get_provisioning()) + if provisioning == "THICK" and module.params["thin_provision"]: + volume.update_provisioning("THIN") changed = True - if type == "THIN" and not module.params["thin_provision"]: - if not module.check_mode: - volume.update_provisioning("THICK") + if provisioning == "THIN" and not module.params["thin_provision"]: + volume.update_provisioning("THICK") changed = True if module.params["write_protected"] is not None: is_write_prot = volume.is_write_protected() desired_is_write_prot = module.params["write_protected"] if is_write_prot != desired_is_write_prot: volume.update_field("write_protected", desired_is_write_prot) + changed = True return changed @api_wrapper def delete_volume(module, volume): - """ Delete Volume. Volume could be a snapshot.""" + """ Delete Volume. Volume could be a snapshot. """ + changed = False if not module.check_mode: volume.delete() - changed = True - return True + changed = True + return changed @api_wrapper @@ -274,15 +286,11 @@ def create_snapshot(module, system): parent_volume_name = module.params["parent_volume_name"] try: parent_volume = system.volumes.get(name=parent_volume_name) - except ObjectNotFound as err: - msg = "Cannot create snapshot {0}. Parent volume {1} not found".format( - snapshot_name, parent_volume_name - ) + except ObjectNotFound: + msg = f"Cannot create snapshot {snapshot_name}. Parent volume {parent_volume_name} not found" module.fail_json(msg=msg) if not parent_volume: - msg = "Cannot find new snapshot's parent volume named {0}".format( - parent_volume_name - ) + msg = f"Cannot find new snapshot's parent volume named {parent_volume_name}" module.fail_json(msg=msg) if not module.check_mode: if module.params["snapshot_lock_only"]: @@ -304,9 +312,7 @@ def create_snapshot(module, system): @api_wrapper def update_snapshot(module, snapshot): - """ - Update/refresh snapshot. May also lock it. - """ + """ Update/refresh snapshot. May also lock it. """ refresh_changed = False if not module.params["snapshot_lock_only"]: snap_is_locked = snapshot.get_lock_state() == "LOCKED" @@ -331,88 +337,18 @@ def update_snapshot(module, snapshot): return refresh_changed or lock_changed -def get_sys_pool_vol_parname(module): +def handle_stat(module): + """ Handle the stat state """ system = get_system(module) - pool = get_pool(module, system) - if module.params["name"]: + if module.params['name']: volume = get_volume(module, system) else: - volume = get_vol_sn(module, system) - parname = module.params["parent_volume_name"] - return (system, pool, volume, parname) - - -def check_snapshot_lock_options(module): - """ - Check if specified options are feasible for a snapshot. - - Prevent very long lock times. - max_delta_minutes limits locks to 30 days (43200 minutes). - - This functionality is broken out from manage_snapshot_locks() to allow - it to be called by create_snapshot() before the snapshot is actually - created. - """ - snapshot_lock_expires_at = module.params["snapshot_lock_expires_at"] - - if snapshot_lock_expires_at: # Then user has specified wish to lock snap - lock_expires_at = arrow.get(snapshot_lock_expires_at) - - # Check for lock in the past - now = arrow.utcnow() - if lock_expires_at <= now: - msg = "Cannot lock snapshot with a snapshot_lock_expires_at " - msg += "of '{0}' from the past".format(snapshot_lock_expires_at) - module.fail_json(msg=msg) - - # Check for lock later than max lock, i.e. too far in future. - max_delta_minutes = 43200 # 30 days in minutes - max_lock_expires_at = now.shift(minutes=max_delta_minutes) - if lock_expires_at >= max_lock_expires_at: - msg = "snapshot_lock_expires_at exceeds {0} days in the future".format( - max_delta_minutes // 24 // 60 - ) - module.fail_json(msg=msg) - - -def manage_snapshot_locks(module, snapshot): - """ - Manage the locking of a snapshot. Check for bad lock times. - See check_snapshot_lock_options() which has additional checks. - """ - name = module.params["name"] - snapshot_lock_expires_at = module.params["snapshot_lock_expires_at"] - snap_is_locked = snapshot.get_lock_state() == "LOCKED" - current_lock_expires_at = snapshot.get_lock_expires_at() - changed = False - - check_snapshot_lock_options(module) - - if snapshot_lock_expires_at: # Then user has specified wish to lock snap - lock_expires_at = arrow.get(snapshot_lock_expires_at) - if snap_is_locked and lock_expires_at < current_lock_expires_at: - # Lock earlier than current lock - msg = "snapshot_lock_expires_at '{0}' preceeds the current lock time of '{1}'".format( - lock_expires_at, current_lock_expires_at - ) - module.fail_json(msg=msg) - elif snap_is_locked and lock_expires_at == current_lock_expires_at: - # Lock already set to correct time - pass - else: - # Set lock - if not module.check_mode: - snapshot.update_lock_expires_at(lock_expires_at) - changed = True - return changed - - -def handle_stat(module): - system, pool, volume, parname = get_sys_pool_vol_parname(module) + volume = get_vol_by_sn(module, system) if not volume: - msg = "Volume {0} not found. Cannot stat.".format(module.params["name"]) + msg = f"Volume {module.params['name']} not found. Cannot stat." module.fail_json(msg=msg) fields = volume.get_fields() # from_cache=True, raw_value=True) + created_at = str(fields.get("created_at", None)) has_children = fields.get("has_children", None) lock_expires_at = str(volume.get_lock_expires_at()) @@ -429,7 +365,7 @@ def handle_stat(module): volume_type = fields.get("type", None) write_protected = fields.get("write_protected", None) if volume_type == "SNAPSHOT": - msg = "Snapshot stat found" + msg = "Volume snapshot stat found" else: msg = "Volume stat found" @@ -456,7 +392,12 @@ def handle_stat(module): def handle_present(module): - system, pool, volume, parname = get_sys_pool_vol_parname(module) + """ Handle the present state """ + system = get_system(module) + if module.params["name"]: + volume = get_volume(module, system) + else: + volume = get_vol_by_sn(module, system) volume_type = module.params["volume_type"] is_restoring = module.params["restore_volume_from_snapshot"] if volume_type == "master": @@ -465,7 +406,11 @@ def handle_present(module): module.exit_json(changed=changed, msg="Volume created") else: changed = update_volume(module, volume) - module.exit_json(changed=changed, msg="Volume updated") + if changed: + msg = "Volume updated" + else: + msg = "Volume present. No changes were required" + module.exit_json(changed=changed, msg=msg) elif volume_type == "snapshot": snapshot = volume if is_restoring: @@ -484,7 +429,12 @@ def handle_present(module): def handle_absent(module): - system, pool, volume, parname = get_sys_pool_vol_parname(module) + """ Handle the absent state """ + system = get_system(module) + if module.params["name"]: + volume = get_volume(module, system) + else: + volume = get_vol_by_sn(module, system) volume_type = module.params["volume_type"] if volume and volume.get_lock_state() == "LOCKED": @@ -498,10 +448,10 @@ def handle_absent(module): changed = delete_volume(module, volume) module.exit_json(changed=changed, msg="Volume removed") elif volume_type == "snapshot": - if not volume: + snapshot = volume + if not snapshot: module.exit_json(changed=False, msg="Snapshot already absent") else: - snapshot = volume changed = delete_volume(module, snapshot) module.exit_json(changed=changed, msg="Snapshot removed") else: @@ -509,7 +459,7 @@ def handle_absent(module): def execute_state(module): - # Handle different write_protected defaults depending on volume_type. + """ Handle each state. Handle different write_protected defaults depending on volume_type. """ if module.params["volume_type"] == "snapshot": if module.params["write_protected"] in ["True", "true", "Default"]: module.params["write_protected"] = True @@ -521,8 +471,8 @@ def execute_state(module): else: module.params["write_protected"] = True else: - msg = f"An error has occurred handling volume_type '{module.params['volume_type']}' or write_protected '{module.params['write_protected']}' values" - module.fail_json(msg=msg) + msg = f"An error has occurred handling volume_type {module.params['volume_type']} or write_protected {module.params['write_protected']} values" + module.fail_json(msg) state = module.params["state"] try: @@ -533,9 +483,7 @@ def execute_state(module): elif state == "absent": handle_absent(module) else: - module.fail_json( - msg="Internal handler error. Invalid state: {0}".format(state) - ) + module.fail_json(msg=f"Internal handler error. Invalid state: {state}") finally: system = get_system(module) system.logout() @@ -543,22 +491,32 @@ def execute_state(module): def check_options(module): """Verify module options are sane""" + name = module.params["name"] + serial = module.params["serial"] state = module.params["state"] size = module.params["size"] pool = module.params["pool"] volume_type = module.params["volume_type"] parent_volume_name = module.params["parent_volume_name"] + if state == "stat": + if not name and not serial: + msg = "Name or serial parameter must be provided" + module.fail_json(msg) + if state in ["present", "absent"]: + if not name: + msg = "Name parameter must be provided" + module.fail_json(msg=msg) + if state == "present": if volume_type == "master": - if state == "present": - if parent_volume_name: - msg = "parent_volume_name should not be specified " - msg += "if volume_type is 'volume'. Snapshots only." - module.fail_json(msg=msg) - if not size: - msg = "Size is required to create a volume" - module.fail_json(msg=msg) + if parent_volume_name: + msg = "parent_volume_name should not be specified " + msg += "if volume_type is 'master'. Used for snapshots only." + module.fail_json(msg=msg) + if not size: + msg = "Size is required to create a volume" + module.fail_json(msg=msg) elif volume_type == "snapshot": if size or pool: msg = "Neither pool nor size should not be specified " @@ -572,26 +530,28 @@ def check_options(module): else: msg = "A programming error has occurred" module.fail_json(msg=msg) + if not pool and volume_type == "master": + msg = "For state 'present', pool is required" + module.fail_json(msg=msg) def main(): + """ Main """ argument_spec = infinibox_argument_spec() argument_spec.update( dict( - name=dict(required=False), - parent_volume_name=dict(default=None, required=False, type=str), + name=dict(required=False, default=None), + parent_volume_name=dict(default=None, required=False, type="str"), pool=dict(required=False), - size=dict(), - serial=dict(), + restore_volume_from_snapshot=dict(default=False, type="bool"), + serial=dict(required=False, default=None), + size=dict(required=False, default=None), snapshot_lock_expires_at=dict(), - snapshot_lock_only=dict(type="bool", default=False), + snapshot_lock_only=dict(default=False, type="bool"), state=dict(default="present", choices=["stat", "present", "absent"]), thin_provision=dict(type="bool", default=True), - write_protected=dict( - default="Default", choices=["Default", "True", "False"] - ), volume_type=dict(default="master", choices=["master", "snapshot"]), - restore_volume_from_snapshot=dict(default=False, type=bool), + write_protected=dict(default="Default", choices=["Default", "True", "False"]), ) ) @@ -600,16 +560,14 @@ def main(): if not HAS_INFINISDK: module.fail_json(msg=missing_required_lib("infinisdk")) - if not HAS_ARROW: - module.fail_json(msg=missing_required_lib("arrow")) + if not HAS_CAPACITY: + module.fail_json(msg=missing_required_lib("capacity")) if module.params["size"]: try: Capacity(module.params["size"]) - except Exception: - module.fail_json( - msg="size (Physical Capacity) should be defined in MB, GB, TB or PB units" - ) + except Exception: # pylint: disable=broad-exception-caught + module.fail_json(msg="size (Physical Capacity) should be defined in MB, GB, TB or PB units") check_options(module) execute_state(module) |