# tests are fairly 'live' (but safe to run) # setup authorized_keys for logged in user such # that the user can log in as themselves before running tests import unittest import getpass import ansible.playbook import ansible.utils as utils import ansible.callbacks as ans_callbacks import os import shutil import time try: import json except: import simplejson as json EVENTS = [] class TestCallbacks(object): # using same callbacks class for both runner and playbook def __init__(self): pass def set_playbook(self, playbook): self.playbook = playbook def on_start(self): EVENTS.append('start') def on_setup_primary(self): EVENTS.append([ 'primary_setup' ]) def on_setup_secondary(self): EVENTS.append([ 'secondary_setup' ]) def on_skipped(self, host): EVENTS.append([ 'skipped', [ host ]]) def on_import_for_host(self, host, filename): EVENTS.append([ 'import', [ host, filename ]]) def on_error(self, host, msg): EVENTS.append([ 'stderr', [ host, msg ]]) def on_not_import_for_host(self, host, missing_filename): pass def on_notify(self, host, handler): EVENTS.append([ 'notify', [ host, handler ]]) def on_task_start(self, name, is_conditional): EVENTS.append([ 'task start', [ name, is_conditional ]]) def on_unreachable(self, host, msg): EVENTS.append([ 'unreachable', [ host, msg ]]) def on_failed(self, host, results): EVENTS.append([ 'failed', [ host, results ]]) def on_ok(self, host, result): # delete certain info from host_result to make test comparisons easier host_result = result.copy() for k in [ 'ansible_job_id', 'results_file', 'invocation', 'md5sum', 'delta', 'start', 'end' ]: if k in host_result: del host_result[k] for k in host_result.keys(): if k.startswith('facter_') or k.startswith('ohai_'): del host_result[k] EVENTS.append([ 'ok', [ host, host_result ]]) def on_play_start(self, pattern): EVENTS.append([ 'play start', [ pattern ]]) def on_async_ok(self, host, res, jid): EVENTS.append([ 'async ok', [ host ]]) def on_async_poll(self, host, res, jid, clock): EVENTS.append([ 'async poll', [ host ]]) def on_async_failed(self, host, res, jid): EVENTS.append([ 'async failed', [ host ]]) def on_unreachable(self, host, msg): EVENTS.append([ 'failed/dark', [ host, msg ]]) def on_setup_primary(self): pass def on_setup_secondary(self): pass def on_no_hosts(self): pass class TestPlaybook(unittest.TestCase): def setUp(self): self.user = getpass.getuser() self.cwd = os.getcwd() self.test_dir = os.path.join(self.cwd, 'test') self.stage_dir = self._prepare_stage_dir() if os.path.exists('/tmp/ansible_test_data_copy.out'): os.unlink('/tmp/ansible_test_data_copy.out') if os.path.exists('/tmp/ansible_test_data_template.out'): os.unlink('/tmp/ansible_test_data_template.out') def _prepare_stage_dir(self): stage_path = os.path.join(self.test_dir, 'test_data') if os.path.exists(stage_path): shutil.rmtree(stage_path, ignore_errors=False) assert not os.path.exists(stage_path) os.makedirs(stage_path) assert os.path.exists(stage_path) return stage_path def _get_test_file(self, filename): # get a file inside the test input directory filename = os.path.join(self.test_dir, filename) assert os.path.exists(filename) return filename def _get_stage_file(self, filename): # get a file inside the test output directory filename = os.path.join(self.stage_dir, filename) return filename def _run(self, test_playbook): ''' run a module and get the localhost results ''' self.test_callbacks = TestCallbacks() self.playbook = ansible.playbook.PlayBook( playbook = test_playbook, host_list = 'test/ansible_hosts', module_path = 'library/', forks = 1, timeout = 5, remote_user = self.user, remote_pass = None, stats = ans_callbacks.AggregateStats(), callbacks = self.test_callbacks, runner_callbacks = self.test_callbacks ) result = self.playbook.run() # print utils.bigjson(dict(events=EVENTS)) return result def test_one(self): pb = os.path.join(self.test_dir, 'playbook1.yml') actual = self._run(pb) # if different, this will output to screen print "**ACTUAL**" print utils.bigjson(actual) expected = { "127.0.0.2": { "changed": 9, "failures": 0, "ok": 12, "skipped": 1, "unreachable": 0 } } print "**EXPECTED**" print utils.bigjson(expected) assert utils.bigjson(expected) == utils.bigjson(actual) # make sure the template module took options from the vars section data = file('/tmp/ansible_test_data_template.out').read() print data assert data.find("ears") != -1, "template success" def test_includes(self): pb = os.path.join(self.test_dir, 'playbook_includes.yml') actual = self._run(pb) for i, event in enumerate(EVENTS): print "EVENT %s"%(i) print event if i == 6: assert 'blue' in event[1][1]['cmd'] if i == 8: assert 'quack' in event[1][1]['cmd'] if i == 10: assert 'RAN' in event[1][1]['cmd'] if i == 12: assert 'RAN' in event[1][1]['cmd'] if i == 14: assert 'red' in event[1][1]['cmd'] if i == 18 or i == 20: assert 'skipped' in event[0]