diff options
author | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-13 12:04:41 +0000 |
---|---|---|
committer | Daniel Baumann <daniel.baumann@progress-linux.org> | 2024-04-13 12:04:41 +0000 |
commit | 975f66f2eebe9dadba04f275774d4ab83f74cf25 (patch) | |
tree | 89bd26a93aaae6a25749145b7e4bca4a1e75b2be /ansible_collections/community/digitalocean/scripts | |
parent | Initial commit. (diff) | |
download | ansible-975f66f2eebe9dadba04f275774d4ab83f74cf25.tar.xz ansible-975f66f2eebe9dadba04f275774d4ab83f74cf25.zip |
Adding upstream version 7.7.0+dfsg.upstream/7.7.0+dfsg
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'ansible_collections/community/digitalocean/scripts')
-rw-r--r-- | ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.ini | 34 | ||||
-rwxr-xr-x | ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.py | 622 |
2 files changed, 656 insertions, 0 deletions
diff --git a/ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.ini b/ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.ini new file mode 100644 index 000000000..b809554b2 --- /dev/null +++ b/ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.ini @@ -0,0 +1,34 @@ +# Ansible DigitalOcean external inventory script settings +# + +[digital_ocean] + +# The module needs your DigitalOcean API Token. +# It may also be specified on the command line via --api-token +# or via the environment variables DO_API_TOKEN or DO_API_KEY +# +#api_token = 123456abcdefg + + +# API calls to DigitalOcean may be slow. For this reason, we cache the results +# of an API call. Set this to the path you want cache files to be written to. +# One file will be written to this directory: +# - ansible-digital_ocean.cache +# +cache_path = /tmp + + +# The number of seconds a cache file is considered valid. After this many +# seconds, a new API call will be made, and the cache file will be updated. +# +cache_max_age = 300 + +# Use the private network IP address instead of the public when available. +# +use_private_network = False + +# Pass variables to every group, e.g.: +# +# group_variables = { 'ansible_user': 'root' } +# +group_variables = {} diff --git a/ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.py b/ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.py new file mode 100755 index 000000000..1d5685c24 --- /dev/null +++ b/ansible_collections/community/digitalocean/scripts/inventory/digital_ocean.py @@ -0,0 +1,622 @@ +#!/usr/bin/env python + +""" +DigitalOcean external inventory script +====================================== + +Generates Ansible inventory of DigitalOcean Droplets. + +In addition to the --list and --host options used by Ansible, there are options +for generating JSON of other DigitalOcean data. This is useful when creating +droplets. For example, --regions will return all the DigitalOcean Regions. +This information can also be easily found in the cache file, whose default +location is /tmp/ansible-digital_ocean.cache). + +The --pretty (-p) option pretty-prints the output for better human readability. + +---- +Although the cache stores all the information received from DigitalOcean, +the cache is not used for current droplet information (in --list, --host, +--all, and --droplets). This is so that accurate droplet information is always +found. You can force this script to use the cache with --force-cache. + +---- +Configuration is read from `digital_ocean.ini`, then from environment variables, +and then from command-line arguments. + +Most notably, the DigitalOcean API Token must be specified. It can be specified +in the INI file or with the following environment variables: + export DO_API_TOKEN='abc123' or + export DO_API_KEY='abc123' + +Alternatively, it can be passed on the command-line with --api-token. + +If you specify DigitalOcean credentials in the INI file, a handy way to +get them into your environment (e.g., to use the digital_ocean module) +is to use the output of the --env option with export: + export $(digital_ocean.py --env) + +---- +The following groups are generated from --list: + - ID (droplet ID) + - NAME (droplet NAME) + - digital_ocean + - image_ID + - image_NAME + - distro_NAME (distribution NAME from image) + - region_NAME + - size_NAME + - status_STATUS + +For each host, the following variables are registered: + - do_backup_ids + - do_created_at + - do_disk + - do_features - list + - do_id + - do_image - object + - do_ip_address + - do_private_ip_address + - do_kernel - object + - do_locked + - do_memory + - do_name + - do_networks - object + - do_next_backup_window + - do_region - object + - do_size - object + - do_size_slug + - do_snapshot_ids - list + - do_status + - do_tags + - do_vcpus + - do_volume_ids + +----- +``` +usage: digital_ocean.py [-h] [--list] [--host HOST] [--all] [--droplets] + [--regions] [--images] [--sizes] [--ssh-keys] + [--domains] [--tags] [--pretty] + [--cache-path CACHE_PATH] + [--cache-max_age CACHE_MAX_AGE] [--force-cache] + [--refresh-cache] [--env] [--api-token API_TOKEN] + +Produce an Ansible Inventory file based on DigitalOcean credentials + +optional arguments: + -h, --help show this help message and exit + --list List all active Droplets as Ansible inventory + (default: True) + --host HOST Get all Ansible inventory variables about the Droplet + with the given ID + --all List all DigitalOcean information as JSON + --droplets, -d List Droplets as JSON + --regions List Regions as JSON + --images List Images as JSON + --sizes List Sizes as JSON + --ssh-keys List SSH keys as JSON + --domains List Domains as JSON + --tags List Tags as JSON + --pretty, -p Pretty-print results + --cache-path CACHE_PATH + Path to the cache files (default: .) + --cache-max_age CACHE_MAX_AGE + Maximum age of the cached items (default: 0) + --force-cache Only use data from the cache + --refresh-cache, -r Force refresh of cache by making API requests to + DigitalOcean (default: False - use cache files) + --env, -e Display DO_API_TOKEN + --api-token API_TOKEN, -a API_TOKEN + DigitalOcean API Token +``` + +""" + +# (c) 2013, Evan Wies <evan@neomantra.net> +# (c) 2017, Ansible Project +# (c) 2017, Abhijeet Kasurde <akasurde@redhat.com> +# +# Inspired by the EC2 inventory plugin: +# https://github.com/ansible/ansible/blob/devel/contrib/inventory/ec2.py +# +# GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) + +from __future__ import absolute_import, division, print_function + +__metaclass__ = type + +###################################################################### + +import argparse +import ast +import os +import re +import requests +import sys +from time import time + +try: + import ConfigParser +except ImportError: + import configparser as ConfigParser + +import json + + +class DoManager: + def __init__(self, api_token): + self.api_token = api_token + self.api_endpoint = "https://api.digitalocean.com/v2" + self.headers = { + "Authorization": "Bearer {0}".format(self.api_token), + "Content-type": "application/json", + } + self.timeout = 60 + + def _url_builder(self, path): + if path[0] == "/": + path = path[1:] + return "%s/%s" % (self.api_endpoint, path) + + def send(self, url, method="GET", data=None, params=None): + url = self._url_builder(url) + data = json.dumps(data) + try: + if method == "GET": + resp_data = {} + incomplete = True + while incomplete: + resp = requests.get( + url, + data=data, + params=params, + headers=self.headers, + timeout=self.timeout, + ) + json_resp = resp.json() + + for key, value in json_resp.items(): + if isinstance(value, list) and key in resp_data: + resp_data[key] += value + else: + resp_data[key] = value + + try: + url = json_resp["links"]["pages"]["next"] + except KeyError: + incomplete = False + + except ValueError as e: + sys.exit("Unable to parse result from %s: %s" % (url, e)) + return resp_data + + def all_active_droplets(self, tag_name=None): + if tag_name is not None: + params = {"tag_name": tag_name} + resp = self.send("droplets/", params=params) + else: + resp = self.send("droplets/") + return resp["droplets"] + + def all_regions(self): + resp = self.send("regions/") + return resp["regions"] + + def all_images(self, filter_name="global"): + params = {"filter": filter_name} + resp = self.send("images/", data=params) + return resp["images"] + + def sizes(self): + resp = self.send("sizes/") + return resp["sizes"] + + def all_ssh_keys(self): + resp = self.send("account/keys") + return resp["ssh_keys"] + + def all_domains(self): + resp = self.send("domains/") + return resp["domains"] + + def show_droplet(self, droplet_id): + resp = self.send("droplets/%s" % droplet_id) + return resp.get("droplet", {}) + + def all_tags(self): + resp = self.send("tags") + return resp["tags"] + + +class DigitalOceanInventory(object): + + ########################################################################### + # Main execution path + ########################################################################### + + def __init__(self): + """Main execution path""" + + # DigitalOceanInventory data + self.data = {} # All DigitalOcean data + self.inventory = {} # Ansible Inventory + + # Define defaults + self.cache_path = "." + self.cache_max_age = 0 + self.use_private_network = False + self.group_variables = {} + self.droplets_tag_name = None + + # Read settings, environment variables, and CLI arguments + self.read_settings() + self.read_environment() + self.read_cli_args() + + # Verify credentials were set + if not hasattr(self, "api_token"): + msg = ( + "Could not find values for DigitalOcean api_token. They must be specified via either ini file, " + "command line argument (--api-token), or environment variables (DO_API_TOKEN)\n" + ) + sys.stderr.write(msg) + sys.exit(-1) + + # env command, show DigitalOcean credentials + if self.args.env: + print("DO_API_TOKEN=%s" % self.api_token) + sys.exit(0) + + # Manage cache + self.cache_filename = self.cache_path + "/ansible-digital_ocean.cache" + self.cache_refreshed = False + + if self.is_cache_valid(): + self.load_from_cache() + if len(self.data) == 0: + if self.args.force_cache: + sys.stderr.write("Cache is empty and --force-cache was specified\n") + sys.exit(-1) + + self.manager = DoManager(self.api_token) + + # Pick the json_data to print based on the CLI command + if self.args.droplets: + self.load_from_digital_ocean("droplets") + json_data = {"droplets": self.data["droplets"]} + elif self.args.regions: + self.load_from_digital_ocean("regions") + json_data = {"regions": self.data["regions"]} + elif self.args.images: + self.load_from_digital_ocean("images") + json_data = {"images": self.data["images"]} + elif self.args.sizes: + self.load_from_digital_ocean("sizes") + json_data = {"sizes": self.data["sizes"]} + elif self.args.ssh_keys: + self.load_from_digital_ocean("ssh_keys") + json_data = {"ssh_keys": self.data["ssh_keys"]} + elif self.args.domains: + self.load_from_digital_ocean("domains") + json_data = {"domains": self.data["domains"]} + elif self.args.tags: + self.load_from_digital_ocean("tags") + json_data = {"tags": self.data["tags"]} + elif self.args.all: + self.load_from_digital_ocean() + json_data = self.data + elif self.args.host: + json_data = self.load_droplet_variables_for_host() + else: # '--list' this is last to make it default + self.load_from_digital_ocean("droplets") + self.build_inventory() + json_data = self.inventory + + if self.cache_refreshed: + self.write_to_cache() + + if self.args.pretty: + print(json.dumps(json_data, indent=2)) + else: + print(json.dumps(json_data)) + + ########################################################################### + # Script configuration + ########################################################################### + + def read_settings(self): + """Reads the settings from the digital_ocean.ini file""" + config = ConfigParser.ConfigParser() + config_path = os.path.join( + os.path.dirname(os.path.realpath(__file__)), "digital_ocean.ini" + ) + config.read(config_path) + + # Credentials + if config.has_option("digital_ocean", "api_token"): + self.api_token = config.get("digital_ocean", "api_token") + + # Cache related + if config.has_option("digital_ocean", "cache_path"): + self.cache_path = config.get("digital_ocean", "cache_path") + if config.has_option("digital_ocean", "cache_max_age"): + self.cache_max_age = config.getint("digital_ocean", "cache_max_age") + + # Private IP Address + if config.has_option("digital_ocean", "use_private_network"): + self.use_private_network = config.getboolean( + "digital_ocean", "use_private_network" + ) + + # Group variables + if config.has_option("digital_ocean", "group_variables"): + self.group_variables = ast.literal_eval( + config.get("digital_ocean", "group_variables") + ) + + # Droplet tag_name + if config.has_option("droplets", "tag_name"): + self.droplets_tag_name = config.get("droplets", "tag_name") + + def read_environment(self): + """Reads the settings from environment variables""" + # Setup credentials + if os.getenv("DO_API_TOKEN"): + self.api_token = os.getenv("DO_API_TOKEN") + if os.getenv("DO_API_KEY"): + self.api_token = os.getenv("DO_API_KEY") + + def read_cli_args(self): + """Command line argument processing""" + parser = argparse.ArgumentParser( + description="Produce an Ansible Inventory file based on DigitalOcean credentials" + ) + + parser.add_argument( + "--list", + action="store_true", + help="List all active Droplets as Ansible inventory (default: True)", + ) + parser.add_argument( + "--host", + action="store", + type=int, + help="Get all Ansible inventory variables about the Droplet with the given ID", + ) + + parser.add_argument( + "--all", + action="store_true", + help="List all DigitalOcean information as JSON", + ) + parser.add_argument( + "--droplets", "-d", action="store_true", help="List Droplets as JSON" + ) + parser.add_argument( + "--regions", action="store_true", help="List Regions as JSON" + ) + parser.add_argument("--images", action="store_true", help="List Images as JSON") + parser.add_argument("--sizes", action="store_true", help="List Sizes as JSON") + parser.add_argument( + "--ssh-keys", action="store_true", help="List SSH keys as JSON" + ) + parser.add_argument( + "--domains", action="store_true", help="List Domains as JSON" + ) + parser.add_argument("--tags", action="store_true", help="List Tags as JSON") + + parser.add_argument( + "--pretty", "-p", action="store_true", help="Pretty-print results" + ) + + parser.add_argument( + "--cache-path", action="store", help="Path to the cache files (default: .)" + ) + parser.add_argument( + "--cache-max_age", + action="store", + help="Maximum age of the cached items (default: 0)", + ) + parser.add_argument( + "--force-cache", + action="store_true", + default=False, + help="Only use data from the cache", + ) + parser.add_argument( + "--refresh-cache", + "-r", + action="store_true", + default=False, + help="Force refresh of cache by making API requests to DigitalOcean (default: False - use cache files)", + ) + + parser.add_argument( + "--env", "-e", action="store_true", help="Display DO_API_TOKEN" + ) + parser.add_argument( + "--api-token", "-a", action="store", help="DigitalOcean API Token" + ) + + self.args = parser.parse_args() + + if self.args.api_token: + self.api_token = self.args.api_token + + # Make --list default if none of the other commands are specified + if ( + not self.args.droplets + and not self.args.regions + and not self.args.images + and not self.args.sizes + and not self.args.ssh_keys + and not self.args.domains + and not self.args.tags + and not self.args.all + and not self.args.host + ): + self.args.list = True + + ########################################################################### + # Data Management + ########################################################################### + + def load_from_digital_ocean(self, resource=None): + """Get JSON from DigitalOcean API""" + if self.args.force_cache and os.path.isfile(self.cache_filename): + return + # We always get fresh droplets + if self.is_cache_valid() and not (resource == "droplets" or resource is None): + return + if self.args.refresh_cache: + resource = None + + if resource == "droplets" or resource is None: + self.data["droplets"] = self.manager.all_active_droplets( + tag_name=self.droplets_tag_name + ) + self.cache_refreshed = True + if resource == "regions" or resource is None: + self.data["regions"] = self.manager.all_regions() + self.cache_refreshed = True + if resource == "images" or resource is None: + self.data["images"] = self.manager.all_images() + self.cache_refreshed = True + if resource == "sizes" or resource is None: + self.data["sizes"] = self.manager.sizes() + self.cache_refreshed = True + if resource == "ssh_keys" or resource is None: + self.data["ssh_keys"] = self.manager.all_ssh_keys() + self.cache_refreshed = True + if resource == "domains" or resource is None: + self.data["domains"] = self.manager.all_domains() + self.cache_refreshed = True + if resource == "tags" or resource is None: + self.data["tags"] = self.manager.all_tags() + self.cache_refreshed = True + + def add_inventory_group(self, key): + """Method to create group dict""" + host_dict = {"hosts": [], "vars": {}} + self.inventory[key] = host_dict + return + + def add_host(self, group, host): + """Helper method to reduce host duplication""" + if group not in self.inventory: + self.add_inventory_group(group) + + if host not in self.inventory[group]["hosts"]: + self.inventory[group]["hosts"].append(host) + return + + def build_inventory(self): + """Build Ansible inventory of droplets""" + self.inventory = { + "all": {"hosts": [], "vars": self.group_variables}, + "_meta": {"hostvars": {}}, + } + + # add all droplets by id and name + for droplet in self.data["droplets"]: + for net in droplet["networks"]["v4"]: + if net["type"] == "public": + droplet["ip_address"] = net["ip_address"] + elif net["type"] == "private": + droplet["private_ip_address"] = net["ip_address"] + + host_indentifier = droplet["ip_address"] + if self.use_private_network and droplet["private_ip_address"]: + host_indentifier = droplet["private_ip_address"] + + self.inventory["all"]["hosts"].append(host_indentifier) + + self.add_host(droplet["id"], host_indentifier) + + self.add_host(droplet["name"], host_indentifier) + + # groups that are always present + for group in ( + "digital_ocean", + "region_" + droplet["region"]["slug"], + "image_" + str(droplet["image"]["id"]), + "size_" + droplet["size"]["slug"], + "distro_" + + DigitalOceanInventory.to_safe(droplet["image"]["distribution"]), + "status_" + droplet["status"], + ): + self.add_host(group, host_indentifier) + + # groups that are not always present + for group in (droplet["image"]["slug"], droplet["image"]["name"]): + if group: + image = "image_" + DigitalOceanInventory.to_safe(group) + self.add_host(image, host_indentifier) + + if droplet["tags"]: + for tag in droplet["tags"]: + self.add_host(tag, host_indentifier) + + # hostvars + info = self.do_namespace(droplet) + self.inventory["_meta"]["hostvars"][host_indentifier] = info + + def load_droplet_variables_for_host(self): + """Generate a JSON response to a --host call""" + droplet = self.manager.show_droplet(self.args.host) + info = self.do_namespace(droplet) + return info + + ########################################################################### + # Cache Management + ########################################################################### + + def is_cache_valid(self): + """Determines if the cache files have expired, or if it is still valid""" + if os.path.isfile(self.cache_filename): + mod_time = os.path.getmtime(self.cache_filename) + current_time = time() + if (mod_time + self.cache_max_age) > current_time: + return True + return False + + def load_from_cache(self): + """Reads the data from the cache file and assigns it to member variables as Python Objects""" + try: + with open(self.cache_filename, "r") as cache: + json_data = cache.read() + data = json.loads(json_data) + except IOError: + data = {"data": {}, "inventory": {}} + + self.data = data["data"] + self.inventory = data["inventory"] + + def write_to_cache(self): + """Writes data in JSON format to a file""" + data = {"data": self.data, "inventory": self.inventory} + json_data = json.dumps(data, indent=2) + + with open(self.cache_filename, "w") as cache: + cache.write(json_data) + + ########################################################################### + # Utilities + ########################################################################### + @staticmethod + def to_safe(word): + """Converts 'bad' characters in a string to underscores so they can be used as Ansible groups""" + return re.sub(r"[^A-Za-z0-9\-.]", "_", word) + + @staticmethod + def do_namespace(data): + """Returns a copy of the dictionary with all the keys put in a 'do_' namespace""" + info = {} + for k, v in data.items(): + info["do_" + k] = v + return info + + +########################################################################### +# Run the script +DigitalOceanInventory() |