2013-06-28 21:11:32 +02:00
|
|
|
#!/usr/bin/env python
|
|
|
|
|
|
|
|
'''
|
|
|
|
Linode external inventory script
|
|
|
|
=================================
|
|
|
|
|
|
|
|
Generates inventory that Ansible can understand by making API request to
|
2014-03-09 02:14:10 +01:00
|
|
|
Linode using the Chube library.
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
NOTE: This script assumes Ansible is being executed where Chube is already
|
|
|
|
installed and has a valid config at ~/.chube. If not, run:
|
|
|
|
|
|
|
|
pip install chube
|
|
|
|
echo -e "---\napi_key: <YOUR API KEY GOES HERE>" > ~/.chube
|
|
|
|
|
|
|
|
For more details, see: https://github.com/exosite/chube
|
|
|
|
|
2016-09-09 15:14:27 +02:00
|
|
|
NOTE: By default, this script also assumes that the Linodes in your account all have
|
2013-06-28 21:51:58 +02:00
|
|
|
labels that correspond to hostnames that are in your resolver search path.
|
|
|
|
Your resolver search path resides in /etc/hosts.
|
2016-09-09 15:14:27 +02:00
|
|
|
Optionally, if you would like to use the hosts public IP instead of it's label use
|
|
|
|
the following setting in linode.ini:
|
|
|
|
|
|
|
|
use_public_ip = true
|
2013-06-28 21:51:58 +02:00
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
When run against a specific host, this script returns the following variables:
|
|
|
|
|
|
|
|
- api_id
|
|
|
|
- datacenter_id
|
|
|
|
- datacenter_city (lowercase city name of data center, e.g. 'tokyo')
|
|
|
|
- label
|
|
|
|
- display_group
|
|
|
|
- create_dt
|
|
|
|
- total_hd
|
|
|
|
- total_xfer
|
|
|
|
- total_ram
|
|
|
|
- status
|
|
|
|
- public_ip (The first public IP found)
|
|
|
|
- private_ip (The first private IP found, or empty string if none)
|
|
|
|
- alert_cpu_enabled
|
|
|
|
- alert_cpu_threshold
|
|
|
|
- alert_diskio_enabled
|
|
|
|
- alert_diskio_threshold
|
|
|
|
- alert_bwin_enabled
|
|
|
|
- alert_bwin_threshold
|
|
|
|
- alert_bwout_enabled
|
|
|
|
- alert_bwout_threshold
|
|
|
|
- alert_bwquota_enabled
|
|
|
|
- alert_bwquota_threshold
|
|
|
|
- backup_weekly_daily
|
|
|
|
- backup_window
|
|
|
|
- watchdog
|
|
|
|
|
|
|
|
Peter Sankauskas did most of the legwork here with his linode plugin; I
|
|
|
|
just adapted that for Linode.
|
|
|
|
'''
|
|
|
|
|
|
|
|
# (c) 2013, Dan Slimmon
|
|
|
|
#
|
|
|
|
# This file is part of Ansible,
|
|
|
|
#
|
|
|
|
# Ansible is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU General Public License as published by
|
|
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
|
|
# (at your option) any later version.
|
|
|
|
#
|
|
|
|
# Ansible is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU General Public License
|
|
|
|
# along with Ansible. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
######################################################################
|
|
|
|
|
|
|
|
# Standard imports
|
2014-03-07 21:35:18 +01:00
|
|
|
import os
|
2013-06-28 21:11:32 +02:00
|
|
|
import re
|
|
|
|
import sys
|
|
|
|
import argparse
|
|
|
|
from time import time
|
2014-03-07 21:35:18 +01:00
|
|
|
|
2018-08-10 18:13:29 +02:00
|
|
|
import json
|
2013-06-28 21:11:32 +02:00
|
|
|
|
2014-03-07 21:35:18 +01:00
|
|
|
try:
|
|
|
|
from chube import load_chube_config
|
|
|
|
from chube import api as chube_api
|
|
|
|
from chube.datacenter import Datacenter
|
|
|
|
from chube.linode_obj import Linode
|
2018-09-08 02:59:46 +02:00
|
|
|
except Exception:
|
2014-03-07 21:35:18 +01:00
|
|
|
try:
|
|
|
|
# remove local paths and other stuff that may
|
|
|
|
# cause an import conflict, as chube is sensitive
|
|
|
|
# to name collisions on importing
|
|
|
|
old_path = sys.path
|
|
|
|
sys.path = [d for d in sys.path if d not in ('', os.getcwd(), os.path.dirname(os.path.realpath(__file__)))]
|
|
|
|
|
|
|
|
from chube import load_chube_config
|
|
|
|
from chube import api as chube_api
|
|
|
|
from chube.datacenter import Datacenter
|
|
|
|
from chube.linode_obj import Linode
|
|
|
|
|
|
|
|
sys.path = old_path
|
2015-08-27 20:58:51 +02:00
|
|
|
except Exception as e:
|
2014-03-07 21:35:18 +01:00
|
|
|
raise Exception("could not import chube")
|
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
load_chube_config()
|
|
|
|
|
|
|
|
# Imports for ansible
|
|
|
|
import ConfigParser
|
|
|
|
|
2017-05-09 23:38:08 +02:00
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
class LinodeInventory(object):
|
2017-08-02 19:26:30 +02:00
|
|
|
def _empty_inventory(self):
|
|
|
|
return {"_meta": {"hostvars": {}}}
|
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
def __init__(self):
|
|
|
|
"""Main execution path."""
|
|
|
|
# Inventory grouped by display group
|
2017-08-02 19:26:30 +02:00
|
|
|
self.inventory = self._empty_inventory()
|
2013-06-28 21:11:32 +02:00
|
|
|
# Index of label to Linode ID
|
|
|
|
self.index = {}
|
2013-07-25 18:58:15 +02:00
|
|
|
# Local cache of Datacenter objects populated by populate_datacenter_cache()
|
|
|
|
self._datacenter_cache = None
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
# Read settings and parse CLI arguments
|
|
|
|
self.read_settings()
|
|
|
|
self.parse_cli_args()
|
|
|
|
|
|
|
|
# Cache
|
|
|
|
if self.args.refresh_cache:
|
|
|
|
self.do_api_calls_update_cache()
|
|
|
|
elif not self.is_cache_valid():
|
|
|
|
self.do_api_calls_update_cache()
|
|
|
|
|
|
|
|
# Data to print
|
|
|
|
if self.args.host:
|
|
|
|
data_to_print = self.get_host_info()
|
|
|
|
elif self.args.list:
|
|
|
|
# Display list of nodes for inventory
|
2017-08-02 19:26:30 +02:00
|
|
|
if len(self.inventory) == 1:
|
2013-06-28 21:11:32 +02:00
|
|
|
data_to_print = self.get_inventory_from_cache()
|
|
|
|
else:
|
|
|
|
data_to_print = self.json_format_dict(self.inventory, True)
|
|
|
|
|
2015-08-28 08:18:13 +02:00
|
|
|
print(data_to_print)
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
def is_cache_valid(self):
|
|
|
|
"""Determines if the cache file has expired, or if it is still valid."""
|
|
|
|
if os.path.isfile(self.cache_path_cache):
|
|
|
|
mod_time = os.path.getmtime(self.cache_path_cache)
|
|
|
|
current_time = time()
|
|
|
|
if (mod_time + self.cache_max_age) > current_time:
|
|
|
|
if os.path.isfile(self.cache_path_index):
|
|
|
|
return True
|
|
|
|
return False
|
2014-02-09 17:33:34 +01:00
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
def read_settings(self):
|
|
|
|
"""Reads the settings from the .ini file."""
|
|
|
|
config = ConfigParser.SafeConfigParser()
|
|
|
|
config.read(os.path.dirname(os.path.realpath(__file__)) + '/linode.ini')
|
|
|
|
|
|
|
|
# Cache related
|
|
|
|
cache_path = config.get('linode', 'cache_path')
|
|
|
|
self.cache_path_cache = cache_path + "/ansible-linode.cache"
|
|
|
|
self.cache_path_index = cache_path + "/ansible-linode.index"
|
|
|
|
self.cache_max_age = config.getint('linode', 'cache_max_age')
|
2016-09-09 15:14:27 +02:00
|
|
|
self.use_public_ip = config.getboolean('linode', 'use_public_ip')
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
def parse_cli_args(self):
|
|
|
|
"""Command line argument processing"""
|
|
|
|
parser = argparse.ArgumentParser(description='Produce an Ansible Inventory file based on Linode')
|
|
|
|
parser.add_argument('--list', action='store_true', default=True,
|
2017-05-09 23:38:08 +02:00
|
|
|
help='List nodes (default: True)')
|
2013-06-28 21:11:32 +02:00
|
|
|
parser.add_argument('--host', action='store',
|
2017-05-09 23:38:08 +02:00
|
|
|
help='Get all the variables about a specific node')
|
2013-06-28 21:11:32 +02:00
|
|
|
parser.add_argument('--refresh-cache', action='store_true', default=False,
|
2017-05-09 23:38:08 +02:00
|
|
|
help='Force refresh of cache by making API requests to Linode (default: False - use cache files)')
|
2013-06-28 21:11:32 +02:00
|
|
|
self.args = parser.parse_args()
|
|
|
|
|
|
|
|
def do_api_calls_update_cache(self):
|
|
|
|
"""Do API calls, and save data in cache files."""
|
|
|
|
self.get_nodes()
|
|
|
|
self.write_to_cache(self.inventory, self.cache_path_cache)
|
|
|
|
self.write_to_cache(self.index, self.cache_path_index)
|
|
|
|
|
|
|
|
def get_nodes(self):
|
|
|
|
"""Makes an Linode API call to get the list of nodes."""
|
|
|
|
try:
|
2013-07-25 18:58:15 +02:00
|
|
|
for node in Linode.search(status=Linode.STATUS_RUNNING):
|
2013-06-28 21:11:32 +02:00
|
|
|
self.add_node(node)
|
2015-08-27 20:58:51 +02:00
|
|
|
except chube_api.linode_api.ApiError as e:
|
2016-08-23 17:09:23 +02:00
|
|
|
sys.exit("Looks like Linode's API is down:\n %s" % e)
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
def get_node(self, linode_id):
|
|
|
|
"""Gets details about a specific node."""
|
|
|
|
try:
|
|
|
|
return Linode.find(api_id=linode_id)
|
2015-08-27 20:58:51 +02:00
|
|
|
except chube_api.linode_api.ApiError as e:
|
2017-04-07 01:58:16 +02:00
|
|
|
sys.exit("Looks like Linode's API is down:\n%s" % e)
|
2013-06-28 21:11:32 +02:00
|
|
|
|
2013-07-25 18:58:15 +02:00
|
|
|
def populate_datacenter_cache(self):
|
|
|
|
"""Creates self._datacenter_cache, containing all Datacenters indexed by ID."""
|
|
|
|
self._datacenter_cache = {}
|
|
|
|
dcs = Datacenter.search()
|
|
|
|
for dc in dcs:
|
|
|
|
self._datacenter_cache[dc.api_id] = dc
|
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
def get_datacenter_city(self, node):
|
|
|
|
"""Returns a the lowercase city name of the node's data center."""
|
2013-07-25 18:58:15 +02:00
|
|
|
if self._datacenter_cache is None:
|
|
|
|
self.populate_datacenter_cache()
|
|
|
|
location = self._datacenter_cache[node.datacenter_id].location
|
2013-06-28 21:11:32 +02:00
|
|
|
location = location.lower()
|
|
|
|
location = location.split(",")[0]
|
|
|
|
return location
|
|
|
|
|
|
|
|
def add_node(self, node):
|
|
|
|
"""Adds an node to the inventory and index."""
|
2016-09-09 15:14:27 +02:00
|
|
|
if self.use_public_ip:
|
|
|
|
dest = self.get_node_public_ip(node)
|
|
|
|
else:
|
|
|
|
dest = node.label
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
# Add to index
|
2014-02-09 17:33:34 +01:00
|
|
|
self.index[dest] = node.api_id
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
# Inventory: Group by node ID (always a group of 1)
|
|
|
|
self.inventory[node.api_id] = [dest]
|
|
|
|
|
|
|
|
# Inventory: Group by datacenter city
|
|
|
|
self.push(self.inventory, self.get_datacenter_city(node), dest)
|
|
|
|
|
2017-08-02 19:26:30 +02:00
|
|
|
# Inventory: Group by display group
|
2013-06-28 21:11:32 +02:00
|
|
|
self.push(self.inventory, node.display_group, dest)
|
|
|
|
|
2017-08-02 19:26:30 +02:00
|
|
|
# Inventory: Add a "linode" global tag group
|
|
|
|
self.push(self.inventory, "linode", dest)
|
|
|
|
|
|
|
|
# Add host info to hostvars
|
2018-01-02 16:24:53 +01:00
|
|
|
self.inventory["_meta"]["hostvars"][dest] = self._get_host_info(node)
|
2017-08-02 19:26:30 +02:00
|
|
|
|
2016-09-09 15:14:27 +02:00
|
|
|
def get_node_public_ip(self, node):
|
|
|
|
"""Returns a the public IP address of the node"""
|
|
|
|
return [addr.address for addr in node.ipaddresses if addr.is_public][0]
|
|
|
|
|
2013-06-28 21:11:32 +02:00
|
|
|
def get_host_info(self):
|
|
|
|
"""Get variables about a specific host."""
|
|
|
|
|
|
|
|
if len(self.index) == 0:
|
|
|
|
# Need to load index from cache
|
|
|
|
self.load_index_from_cache()
|
|
|
|
|
2017-05-09 23:38:08 +02:00
|
|
|
if self.args.host not in self.index:
|
2013-06-28 21:11:32 +02:00
|
|
|
# try updating the cache
|
|
|
|
self.do_api_calls_update_cache()
|
2017-05-09 23:38:08 +02:00
|
|
|
if self.args.host not in self.index:
|
2013-06-28 21:11:32 +02:00
|
|
|
# host might not exist anymore
|
|
|
|
return self.json_format_dict({}, True)
|
|
|
|
|
|
|
|
node_id = self.index[self.args.host]
|
|
|
|
node = self.get_node(node_id)
|
2017-08-02 19:26:30 +02:00
|
|
|
|
2018-01-02 16:24:53 +01:00
|
|
|
return self.json_format_dict(self._get_host_info(node), True)
|
2017-08-02 19:26:30 +02:00
|
|
|
|
2018-01-02 16:24:53 +01:00
|
|
|
def _get_host_info(self, node):
|
2013-06-28 21:11:32 +02:00
|
|
|
node_vars = {}
|
|
|
|
for direct_attr in [
|
|
|
|
"api_id",
|
|
|
|
"datacenter_id",
|
|
|
|
"label",
|
|
|
|
"display_group",
|
|
|
|
"create_dt",
|
|
|
|
"total_hd",
|
|
|
|
"total_xfer",
|
|
|
|
"total_ram",
|
|
|
|
"status",
|
|
|
|
"alert_cpu_enabled",
|
|
|
|
"alert_cpu_threshold",
|
|
|
|
"alert_diskio_enabled",
|
|
|
|
"alert_diskio_threshold",
|
|
|
|
"alert_bwin_enabled",
|
|
|
|
"alert_bwin_threshold",
|
|
|
|
"alert_bwout_enabled",
|
|
|
|
"alert_bwout_threshold",
|
|
|
|
"alert_bwquota_enabled",
|
|
|
|
"alert_bwquota_threshold",
|
|
|
|
"backup_weekly_daily",
|
|
|
|
"backup_window",
|
|
|
|
"watchdog"
|
|
|
|
]:
|
|
|
|
node_vars[direct_attr] = getattr(node, direct_attr)
|
|
|
|
|
|
|
|
node_vars["datacenter_city"] = self.get_datacenter_city(node)
|
2016-09-09 15:14:27 +02:00
|
|
|
node_vars["public_ip"] = self.get_node_public_ip(node)
|
2014-04-15 01:05:44 +02:00
|
|
|
|
2016-02-11 00:54:52 +01:00
|
|
|
# Set the SSH host information, so these inventory items can be used if
|
|
|
|
# their labels aren't FQDNs
|
|
|
|
node_vars['ansible_ssh_host'] = node_vars["public_ip"]
|
|
|
|
node_vars['ansible_host'] = node_vars["public_ip"]
|
|
|
|
|
2014-04-15 01:05:44 +02:00
|
|
|
private_ips = [addr.address for addr in node.ipaddresses if not addr.is_public]
|
|
|
|
|
|
|
|
if private_ips:
|
|
|
|
node_vars["private_ip"] = private_ips[0]
|
2013-06-28 21:11:32 +02:00
|
|
|
|
2017-08-02 19:26:30 +02:00
|
|
|
return node_vars
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
def push(self, my_dict, key, element):
|
|
|
|
"""Pushed an element onto an array that may not have been defined in the dict."""
|
|
|
|
if key in my_dict:
|
2017-01-28 09:12:11 +01:00
|
|
|
my_dict[key].append(element)
|
2013-06-28 21:11:32 +02:00
|
|
|
else:
|
|
|
|
my_dict[key] = [element]
|
|
|
|
|
|
|
|
def get_inventory_from_cache(self):
|
|
|
|
"""Reads the inventory from the cache file and returns it as a JSON object."""
|
|
|
|
cache = open(self.cache_path_cache, 'r')
|
|
|
|
json_inventory = cache.read()
|
|
|
|
return json_inventory
|
|
|
|
|
|
|
|
def load_index_from_cache(self):
|
|
|
|
"""Reads the index from the cache file and sets self.index."""
|
|
|
|
cache = open(self.cache_path_index, 'r')
|
|
|
|
json_index = cache.read()
|
|
|
|
self.index = json.loads(json_index)
|
|
|
|
|
|
|
|
def write_to_cache(self, data, filename):
|
|
|
|
"""Writes data in JSON format to a file."""
|
|
|
|
json_data = self.json_format_dict(data, True)
|
|
|
|
cache = open(filename, 'w')
|
|
|
|
cache.write(json_data)
|
|
|
|
cache.close()
|
|
|
|
|
|
|
|
def to_safe(self, word):
|
|
|
|
"""Escapes any characters that would be invalid in an ansible group name."""
|
2017-11-21 04:08:30 +01:00
|
|
|
return re.sub(r"[^A-Za-z0-9\-]", "_", word)
|
2013-06-28 21:11:32 +02:00
|
|
|
|
|
|
|
def json_format_dict(self, data, pretty=False):
|
|
|
|
"""Converts a dict to a JSON object and dumps it as a formatted string."""
|
|
|
|
if pretty:
|
|
|
|
return json.dumps(data, sort_keys=True, indent=2)
|
|
|
|
else:
|
|
|
|
return json.dumps(data)
|
|
|
|
|
|
|
|
|
|
|
|
LinodeInventory()
|