diff options
author | Kenny Woodson <kwoodson@redhat.com> | 2014-12-15 11:14:56 -0500 |
---|---|---|
committer | Kenny Woodson <kwoodson@redhat.com> | 2014-12-15 11:14:56 -0500 |
commit | c5c69ab47dc0e65c3da590bbc91e0938a2623725 (patch) | |
tree | 67271421c0aeca6bb72b80f0b3e2d49129938926 /inventory/multi_ec2.py | |
parent | ffe702b2e85278d7a1625b9fae33695360aff3ed (diff) | |
parent | d8ccf2a6b937f53e4f564102d4e7f009ade1e26a (diff) | |
download | openshift-c5c69ab47dc0e65c3da590bbc91e0938a2623725.tar.gz openshift-c5c69ab47dc0e65c3da590bbc91e0938a2623725.tar.bz2 openshift-c5c69ab47dc0e65c3da590bbc91e0938a2623725.tar.xz openshift-c5c69ab47dc0e65c3da590bbc91e0938a2623725.zip |
Merge pull request #34 from kwoodson/meta_inventory
Adding a meta inventory to query multiple cloud accounts.
Diffstat (limited to 'inventory/multi_ec2.py')
-rwxr-xr-x | inventory/multi_ec2.py | 196 |
1 files changed, 196 insertions, 0 deletions
diff --git a/inventory/multi_ec2.py b/inventory/multi_ec2.py new file mode 100755 index 000000000..7fbfb0c90 --- /dev/null +++ b/inventory/multi_ec2.py @@ -0,0 +1,196 @@ +#!/usr/bin/env python + +from time import time +import argparse +import yaml +import os +import sys +import pdb +import subprocess +import json +import pprint + + +class MultiEc2(object): + + def __init__(self): + self.config = None + self.results = {} + self.result = {} + self.cache_path_cache = os.path.expanduser('~/.ansible/tmp/multi_ec2_inventory.cache') + self.file_path = os.path.join(os.path.dirname(os.path.realpath(__file__))) + + self.parse_cli_args() + + # load yaml + self.load_yaml_config() + + # if its a host query, fetch and do not cache + if self.args.host: + self.get_inventory() + elif not self.is_cache_valid(): + # go fetch the inventories and cache them if cache is expired + self.get_inventory() + self.write_to_cache() + else: + # get data from disk + self.get_inventory_from_cache() + + def load_yaml_config(self,conf_file=None): + """Load a yaml config file with credentials to query the + respective cloud for inventory. + """ + config = None + if not conf_file: + conf_file = os.path.join(self.file_path,'multi_ec2.yaml') + with open(conf_file) as conf: + self.config = yaml.safe_load(conf) + + def get_provider_tags(self,provider, env={}): + """Call <provider> and query all of the tags that are usuable + by ansible. If environment is empty use the default env. + """ + if not env: + env = os.environ + + # check to see if provider exists + if not os.path.isfile(provider) or not os.access(provider, os.X_OK): + raise RuntimeError("Problem with the provider. Please check path " \ + "and that it is executable. (%s)" % provider) + + cmds = [provider] + if self.args.host: + cmds.append("--host") + cmds.append(self.args.host) + else: + cmds.append('--list') + + cmds.append('--refresh-cache') + + return subprocess.Popen(cmds, stderr=subprocess.PIPE, \ + stdout=subprocess.PIPE, env=env) + def get_inventory(self): + """Create the subprocess to fetch tags from a provider. + Host query: + Query to return a specific host. If > 1 queries have + results then fail. + + List query: + Query all of the different accounts for their tags. Once completed + store all of their results into one merged updated hash. + """ + processes = {} + for account in self.config['accounts']: + env = account['env_vars'] + name = account['name'] + provider = account['provider'] + processes[name] = self.get_provider_tags(provider, env) + + # for each process collect stdout when its available + all_results = [] + for name, process in processes.items(): + out, err = process.communicate() + all_results.append({ + "name": name, + "out": out.strip(), + "err": err.strip(), + "code": process.returncode + }) + + if not self.args.host: + # For any non-zero, raise an error on it + for result in all_results: + if result['code'] != 0: + raise RuntimeError(result['err']) + else: + self.results[result['name']] = json.loads(result['out']) + values = self.results.values() + values.insert(0, self.result) + map(lambda x: self.merge_destructively(self.result, x), values) + else: + # For any 0 result, return it + count = 0 + for results in all_results: + if results['code'] == 0 and results['err'] == '' and results['out'] != '{}': + self.result = json.loads(out) + count += 1 + if count > 1: + raise RuntimeError("Found > 1 results for --host %s. \ + This is an invalid state." % self.args.host) + def merge_destructively(self, a, b): + "merges b into a" + for key in b: + if key in a: + if isinstance(a[key], dict) and isinstance(b[key], dict): + self.merge_destructively(a[key], b[key]) + elif a[key] == b[key]: + pass # same leaf value + # both lists so add each element in b to a if it does ! exist + elif isinstance(a[key], list) and isinstance(b[key],list): + for x in b[key]: + if x not in a[key]: + a[key].append(x) + # a is a list and not b + elif isinstance(a[key], list): + if b[key] not in a[key]: + a[key].append(b[key]) + elif isinstance(b[key], list): + a[key] = [a[key]] + [k for k in b[key] if k != a[key]] + else: + a[key] = [a[key],b[key]] + else: + a[key] = b[key] + return a + + def is_cache_valid(self): + ''' Determines if the cache files have expired, or if it is still valid ''' + + if os.path.isfile(self.cache_path_cache): + mod_time = os.path.getmtime(self.cache_path_cache) + current_time = time() + if (mod_time + self.config['cache_max_age']) > current_time: + #if os.path.isfile(self.cache_path_index): + return True + + return False + + def parse_cli_args(self): + ''' Command line argument processing ''' + + parser = argparse.ArgumentParser(description='Produce an Ansible Inventory file based on a provider') + parser.add_argument('--list', action='store_true', default=True, + help='List instances (default: True)') + parser.add_argument('--host', action='store', + help='Get all the variables about a specific instance') + self.args = parser.parse_args() + + def write_to_cache(self): + ''' Writes data in JSON format to a file ''' + + json_data = self.json_format_dict(self.result, True) + with open(self.cache_path_cache, 'w') as cache: + cache.write(json_data) + + def get_inventory_from_cache(self): + ''' Reads the inventory from the cache file and returns it as a JSON + object ''' + + with open(self.cache_path_cache, 'r') as cache: + self.result = json.loads(cache.read()) + + def json_format_dict(self, data, pretty=False): + ''' Converts a dict to a JSON object and dumps it as a formatted + string ''' + + if pretty: + return json.dumps(data, sort_keys=True, indent=2) + else: + return json.dumps(data) + + +if __name__ == "__main__": + mi = MultiEc2() + #print mi.result + pp = pprint.PrettyPrinter(indent=2) + pp.pprint(mi.result) + |