
This enables basic clustering functionality. We add: tools/cluster/cluster/daemon.py: A server that handles validation of cluster passwords. tools/cluster/cluster/client.py: A client for this server. Important Note: This prototype does not support TLS, and the functionality in the client and server is basic. Before we roll clustering out to production, we need to have those two chat over TLS, and be much more careful about verifying credentials. Also included ... Various fixes and changes to the init script and config templates to support cluster configuration, and allow for the fact that we may have endpoint references for two network ips. Updates to snapcraft.yaml, adding the new tooling. A more formalized config infrastructure. It's still a TODO to move the specification out of the implicit definition in the install hook, and into a nice, explicit, well documented yaml file. Added nesting to the Question classes in the init script, as well as strings pointing at config keys, rather than having the config be implicitly indicated by the Question subclass' name. (This allows us to put together a config spec that doesn't require the person reading the spec to understand what Questions are, and how they are implemented.) Renamed and unified the "unit" and "lint" tox environments, to allow for the multiple Python tools that we want to lint and test. Added hooks in the init script to make it possible to do automated testing, and added an automated test for a cluster. Run with "tox -e cluster". Added cirros image to snap, to work around sporadic issues downloading it from download.cirros.net. Removed ping logic from snap, to workaround failures in gate. Need to add it back in once we fix them. Change-Id: I44ccd16168a7ed41486464df8c9e22a14d71ccfd
132 lines
4.3 KiB
Python
Executable File
132 lines
4.3 KiB
Python
Executable File
#!/usr/bin/env python
|
|
"""
|
|
cluster_test.py
|
|
|
|
This is a test to verify that we can setup a small, two node cluster.
|
|
|
|
The host running this test must have at least 16GB of RAM, four cpu
|
|
cores, a large amount of disk space, and the ability to run multipass
|
|
vms.
|
|
|
|
"""
|
|
|
|
import json
|
|
import os
|
|
import petname
|
|
import sys
|
|
import unittest
|
|
|
|
sys.path.append(os.getcwd())
|
|
|
|
from tests.framework import Framework, check, check_output, call # noqa E402
|
|
|
|
|
|
os.environ['MULTIPASS'] = 'true' # TODO better way to do this.
|
|
|
|
|
|
class TestCluster(Framework):
|
|
|
|
INIT_FLAG = 'control'
|
|
|
|
def _compute_node(self, channel='dangerous'):
|
|
"""Make a compute node.
|
|
|
|
TODO: refactor framework so that we can fold a lot of this
|
|
into the parent framework. There's a lot of dupe code here.
|
|
|
|
"""
|
|
machine = petname.generate()
|
|
prefix = ['multipass', 'exec', machine, '--']
|
|
|
|
check('multipass', 'launch', '--cpus', '2', '--mem', '8G',
|
|
self.DISTRO, '--name', machine)
|
|
|
|
check('multipass', 'copy-files', self.SNAP, '{}:'.format(machine))
|
|
check(*prefix, 'sudo', 'snap', 'install', '--classic',
|
|
'--{}'.format(channel), self.SNAP)
|
|
|
|
return machine, prefix
|
|
|
|
def test_cluster(self):
|
|
|
|
# After the setUp step, we should have a control node running
|
|
# in a multipass vm. Let's look up its cluster password and ip
|
|
# address.
|
|
|
|
openstack = '/snap/bin/microstack.openstack'
|
|
|
|
cluster_password = check_output(*self.PREFIX, 'sudo', 'snap',
|
|
'get', 'microstack',
|
|
'config.cluster.password')
|
|
control_ip = check_output(*self.PREFIX, 'sudo', 'snap',
|
|
'get', 'microstack',
|
|
'config.network.control-ip')
|
|
|
|
self.assertTrue(cluster_password)
|
|
self.assertTrue(control_ip)
|
|
|
|
compute_machine, compute_prefix = self._compute_node()
|
|
|
|
# TODO add the following to args for init
|
|
check(*compute_prefix, 'sudo', 'snap', 'set', 'microstack',
|
|
'config.network.control-ip={}'.format(control_ip))
|
|
|
|
check(*compute_prefix, 'sudo', 'microstack.init', '--compute',
|
|
'--cluster-password', cluster_password, '--debug')
|
|
|
|
# Verify that our services look setup properly on compute node.
|
|
services = check_output(
|
|
*compute_prefix, 'systemctl', 'status', 'snap.microstack.*',
|
|
'--no-page')
|
|
|
|
self.assertTrue('nova-compute' in services)
|
|
self.assertFalse('keystone-' in services)
|
|
|
|
check(*compute_prefix, '/snap/bin/microstack.launch', 'cirros',
|
|
'--name', 'breakfast', '--retry',
|
|
'--availability-zone', 'nova:{}'.format(compute_machine))
|
|
|
|
# TODO: verify horizon dashboard on control node.
|
|
|
|
# Verify endpoints
|
|
compute_ip = check_output(*compute_prefix, 'sudo', 'snap',
|
|
'get', 'microstack',
|
|
'config.network.compute-ip')
|
|
self.assertFalse(compute_ip == control_ip)
|
|
|
|
# Ping the instance
|
|
ip = None
|
|
servers = check_output(*compute_prefix, openstack,
|
|
'server', 'list', '--format', 'json')
|
|
servers = json.loads(servers)
|
|
for server in servers:
|
|
if server['Name'] == 'breakfast':
|
|
ip = server['Networks'].split(",")[1].strip()
|
|
break
|
|
|
|
self.assertTrue(ip)
|
|
|
|
pings = 1
|
|
max_pings = 60 # ~1 minutes
|
|
# Ping the machine from the control node (we don't have
|
|
# networking wired up for the other nodes).
|
|
while not call(*self.PREFIX, 'ping', '-c1', '-w1', ip):
|
|
pings += 1
|
|
if pings > max_pings:
|
|
self.assertFalse(
|
|
True,
|
|
msg='Max pings reached for instance on {}!'.format(
|
|
compute_machine))
|
|
|
|
self.passed = True
|
|
|
|
# Compute machine cleanup
|
|
check('sudo', 'multipass', 'delete', compute_machine)
|
|
|
|
|
|
if __name__ == '__main__':
|
|
# Run our tests, ignoring deprecation warnings and warnings about
|
|
# unclosed sockets. (TODO: setup a selenium server so that we can
|
|
# move from PhantomJS, which is deprecated, to to Selenium headless.)
|
|
unittest.main(warnings='ignore')
|