Merge "docker-cmd hook"
This commit is contained in:
commit
a6a6861ff1
@ -0,0 +1,9 @@
|
||||
A hook which uses the `docker` command to deploy containers.
|
||||
|
||||
The hook currently supports specifying containers in the `docker-compose v1
|
||||
format <https://docs.docker.com/compose/compose-file/#/version-1>`_. The
|
||||
intention is for this hook to also support the kubernetes pod format.
|
||||
|
||||
A dedicated os-refresh-config script will remove running containers if a
|
||||
deployment is removed or changed, then the docker-cmd hook will run any
|
||||
containers in new or updated deployments.
|
@ -0,0 +1,2 @@
|
||||
os-apply-config
|
||||
os-refresh-config
|
@ -0,0 +1,6 @@
|
||||
#!/bin/bash
|
||||
set -x
|
||||
|
||||
SCRIPTDIR=$(dirname $0)
|
||||
|
||||
install -D -g root -o root -m 0755 ${SCRIPTDIR}/hook-docker-cmd.py /var/lib/heat-config/hooks/docker-cmd
|
138
hot/software-config/elements/heat-config-docker-cmd/install.d/hook-docker-cmd.py
Executable file
138
hot/software-config/elements/heat-config-docker-cmd/install.d/hook-docker-cmd.py
Executable file
@ -0,0 +1,138 @@
|
||||
#!/usr/bin/env python
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
import logging
|
||||
import os
|
||||
import subprocess
|
||||
import sys
|
||||
import yaml
|
||||
|
||||
|
||||
DOCKER_CMD = os.environ.get('HEAT_DOCKER_CMD', 'docker')
|
||||
|
||||
|
||||
log = None
|
||||
|
||||
|
||||
def build_response(deploy_stdout, deploy_stderr, deploy_status_code):
|
||||
return {
|
||||
'deploy_stdout': deploy_stdout,
|
||||
'deploy_stderr': deploy_stderr,
|
||||
'deploy_status_code': deploy_status_code,
|
||||
}
|
||||
|
||||
|
||||
def docker_arg_map(key, value):
|
||||
value = str(value).encode('ascii', 'ignore')
|
||||
return {
|
||||
'container_step_config': None,
|
||||
'environment': "--env=%s" % value,
|
||||
'image': value,
|
||||
'net': "--net=%s" % value,
|
||||
'pid': "--pid=%s" % value,
|
||||
'privileged': "--privileged=%s" % 'true' if value else 'false',
|
||||
'restart': "--restart=%s" % value,
|
||||
'user': "--user=%s" % value,
|
||||
'volumes': "--volume=%s" % value,
|
||||
'volumes_from': "--volumes-from=%s" % value,
|
||||
}.get(key, None)
|
||||
|
||||
|
||||
def main(argv=sys.argv):
|
||||
global log
|
||||
log = logging.getLogger('heat-config')
|
||||
handler = logging.StreamHandler(sys.stderr)
|
||||
handler.setFormatter(
|
||||
logging.Formatter(
|
||||
'[%(asctime)s] (%(name)s) [%(levelname)s] %(message)s'))
|
||||
log.addHandler(handler)
|
||||
log.setLevel('DEBUG')
|
||||
|
||||
c = json.load(sys.stdin)
|
||||
|
||||
input_values = dict((i['name'], i['value']) for i in c.get('inputs', {}))
|
||||
|
||||
if input_values.get('deploy_action') == 'DELETE':
|
||||
json.dump(build_response(
|
||||
'', '', 0), sys.stdout)
|
||||
return
|
||||
|
||||
config = c.get('config', '')
|
||||
if not config:
|
||||
log.debug("No 'config' input found, nothing to do.")
|
||||
json.dump(build_response(
|
||||
'', '', 0), sys.stdout)
|
||||
return
|
||||
|
||||
stdout = []
|
||||
stderr = []
|
||||
deploy_status_code = 0
|
||||
|
||||
# convert config to dict
|
||||
if not isinstance(config, dict):
|
||||
config = yaml.safe_load(config)
|
||||
|
||||
for container in config:
|
||||
container_name = '%s__%s' % (c['name'], container)
|
||||
cmd = [
|
||||
DOCKER_CMD,
|
||||
'run',
|
||||
'--detach=true',
|
||||
'--name',
|
||||
container_name.encode('ascii', 'ignore'),
|
||||
]
|
||||
image_name = ''
|
||||
for key in sorted(config[container]):
|
||||
# These ones contain a list of values
|
||||
if key in ['environment', 'volumes', 'volumes_from']:
|
||||
for value in config[container][key]:
|
||||
# Somehow the lists get empty values sometimes
|
||||
if type(value) is unicode and not value.strip():
|
||||
continue
|
||||
cmd.append(docker_arg_map(key, value))
|
||||
elif key == 'image':
|
||||
image_name = config[container][key].encode('ascii', 'ignore')
|
||||
else:
|
||||
arg = docker_arg_map(key, config[container][key])
|
||||
if arg:
|
||||
cmd.append(arg)
|
||||
|
||||
# Image name must come last.
|
||||
cmd.append(image_name)
|
||||
|
||||
log.debug(' '.join(cmd))
|
||||
subproc = subprocess.Popen(cmd, stdout=subprocess.PIPE,
|
||||
stderr=subprocess.PIPE)
|
||||
cmd_stdout, cmd_stderr = subproc.communicate()
|
||||
log.debug(cmd_stdout)
|
||||
log.debug(cmd_stderr)
|
||||
if cmd_stdout:
|
||||
stdout.append(cmd_stdout)
|
||||
if cmd_stderr:
|
||||
stderr.append(cmd_stderr)
|
||||
|
||||
if subproc.returncode:
|
||||
log.error("Error running %s. [%s]\n" % (cmd, subproc.returncode))
|
||||
else:
|
||||
log.debug('Completed %s' % cmd)
|
||||
|
||||
if subproc.returncode != 0:
|
||||
deploy_status_code = subproc.returncode
|
||||
|
||||
json.dump(build_response(
|
||||
'\n'.join(stdout), '\n'.join(stderr), deploy_status_code), sys.stdout)
|
||||
|
||||
if __name__ == '__main__':
|
||||
sys.exit(main(sys.argv))
|
@ -0,0 +1,144 @@
|
||||
#!/usr/bin/env python
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import json
|
||||
import logging
|
||||
import os
|
||||
import subprocess
|
||||
import sys
|
||||
|
||||
import yaml
|
||||
|
||||
|
||||
CONF_FILE = os.environ.get('HEAT_SHELL_CONFIG',
|
||||
'/var/run/heat-config/heat-config')
|
||||
|
||||
WORKING_DIR = os.environ.get(
|
||||
'HEAT_DOCKER_CMD_WORKING',
|
||||
'/var/lib/heat-config/heat-config-docker-cmd')
|
||||
|
||||
DOCKER_CMD = os.environ.get('HEAT_DOCKER_CMD', 'docker')
|
||||
|
||||
|
||||
log = None
|
||||
|
||||
|
||||
def main(argv=sys.argv):
|
||||
global log
|
||||
log = logging.getLogger('heat-config')
|
||||
handler = logging.StreamHandler(sys.stderr)
|
||||
handler.setFormatter(
|
||||
logging.Formatter(
|
||||
'[%(asctime)s] (%(name)s) [%(levelname)s] %(message)s'))
|
||||
log.addHandler(handler)
|
||||
log.setLevel('DEBUG')
|
||||
|
||||
if not os.path.exists(CONF_FILE):
|
||||
log.error('No config file %s' % CONF_FILE)
|
||||
return 1
|
||||
|
||||
if not os.path.isdir(WORKING_DIR):
|
||||
os.makedirs(WORKING_DIR, 0o700)
|
||||
|
||||
try:
|
||||
configs = json.load(open(CONF_FILE))
|
||||
except ValueError:
|
||||
pass
|
||||
|
||||
cmd_configs = list(build_configs(configs))
|
||||
try:
|
||||
delete_missing_projects(cmd_configs)
|
||||
for c in cmd_configs:
|
||||
delete_changed_project(c)
|
||||
write_project(c)
|
||||
except Exception as e:
|
||||
log.exception(e)
|
||||
|
||||
|
||||
def build_configs(configs):
|
||||
for c in configs:
|
||||
if c['group'] != 'docker-cmd':
|
||||
continue
|
||||
if not isinstance(c['config'], dict):
|
||||
# convert config to dict
|
||||
c['config'] = yaml.safe_load(c['config'])
|
||||
yield c
|
||||
|
||||
|
||||
def current_projects():
|
||||
for proj_file in os.listdir(WORKING_DIR):
|
||||
if proj_file.endswith('.json'):
|
||||
proj = proj_file[:-5]
|
||||
yield proj
|
||||
|
||||
|
||||
def remove_project(proj):
|
||||
proj_file = os.path.join(WORKING_DIR, '%s.json' % proj)
|
||||
with open(proj_file, 'r') as f:
|
||||
proj_data = json.load(f)
|
||||
for name in extract_container_names(proj, proj_data):
|
||||
remove_container(name)
|
||||
os.remove(proj_file)
|
||||
|
||||
|
||||
def remove_container(name):
|
||||
cmd = [DOCKER_CMD, 'rm', '-f', name]
|
||||
log.debug(' '.join(cmd))
|
||||
subproc = subprocess.Popen(cmd, stdout=subprocess.PIPE,
|
||||
stderr=subprocess.PIPE)
|
||||
stdout, stderr = subproc.communicate()
|
||||
log.info(stdout)
|
||||
log.debug(stderr)
|
||||
|
||||
|
||||
def delete_missing_projects(configs):
|
||||
config_names = [c['name'] for c in configs]
|
||||
for proj in current_projects():
|
||||
if proj not in config_names:
|
||||
log.debug('%s no longer exists, deleting containers' % proj)
|
||||
remove_project(proj)
|
||||
|
||||
|
||||
def extract_container_names(proj, proj_data):
|
||||
# For now, assume a docker-compose v1 format where the
|
||||
# root keys are service names
|
||||
for name in proj_data:
|
||||
yield '%s__%s' % (proj, name)
|
||||
|
||||
|
||||
def delete_changed_project(c):
|
||||
proj = c['name']
|
||||
proj_file = os.path.join(WORKING_DIR, '%s.json' % proj)
|
||||
proj_data = c.get('config', {})
|
||||
if os.path.isfile(proj_file):
|
||||
with open(proj_file, 'r') as f:
|
||||
prev_proj_data = json.load(f)
|
||||
if proj_data != prev_proj_data:
|
||||
log.debug('%s has changed, deleting containers' % proj)
|
||||
remove_project(proj)
|
||||
|
||||
|
||||
def write_project(c):
|
||||
proj = c['name']
|
||||
proj_file = os.path.join(WORKING_DIR, '%s.json' % proj)
|
||||
proj_data = c.get('config', {})
|
||||
|
||||
with os.fdopen(os.open(
|
||||
proj_file, os.O_CREAT | os.O_WRONLY | os.O_TRUNC, 0o600),
|
||||
'w') as f:
|
||||
json.dump(proj_data, f, indent=2)
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
sys.exit(main(sys.argv))
|
@ -28,7 +28,16 @@ import sys
|
||||
|
||||
|
||||
def main(argv=sys.argv):
|
||||
with open(os.environ.get('TEST_STATE_PATH'), 'w') as f:
|
||||
|
||||
state_path = os.environ.get('TEST_STATE_PATH')
|
||||
|
||||
# handle multiple invocations by writing to numbered state path files
|
||||
suffix = 0
|
||||
while os.path.isfile(state_path):
|
||||
suffix += 1
|
||||
state_path = '%s_%s' % (os.environ.get('TEST_STATE_PATH'), suffix)
|
||||
|
||||
with open(state_path, 'w') as f:
|
||||
json.dump({'env': dict(os.environ), 'args': argv}, f)
|
||||
|
||||
if 'TEST_RESPONSE' not in os.environ:
|
||||
|
267
tests/software_config/test_hook_docker_cmd.py
Normal file
267
tests/software_config/test_hook_docker_cmd.py
Normal file
@ -0,0 +1,267 @@
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
||||
# not use this file except in compliance with the License. You may obtain
|
||||
# a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
||||
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
||||
# License for the specific language governing permissions and limitations
|
||||
# under the License.
|
||||
|
||||
import copy
|
||||
import json
|
||||
import os
|
||||
import tempfile
|
||||
|
||||
import fixtures
|
||||
from testtools import matchers
|
||||
|
||||
from tests.software_config import common
|
||||
|
||||
|
||||
class HookDockerComposeTest(common.RunScriptTest):
|
||||
data = {
|
||||
"name": "abcdef001",
|
||||
"group": "docker-cmd",
|
||||
"config": {
|
||||
"web": {
|
||||
"name": "x",
|
||||
"image": "xxx"
|
||||
},
|
||||
"db": {
|
||||
"name": "y",
|
||||
"image": "xxx",
|
||||
"net": "host",
|
||||
"restart": "always",
|
||||
"privileged": True,
|
||||
"user": "root",
|
||||
"volumes": [
|
||||
"/run:/run",
|
||||
"db:/var/lib/db"
|
||||
],
|
||||
"environment": [
|
||||
"KOLLA_CONFIG_STRATEGY=COPY_ALWAYS",
|
||||
"FOO=BAR"
|
||||
]
|
||||
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
def setUp(self):
|
||||
super(HookDockerComposeTest, self).setUp()
|
||||
self.hook_path = self.relative_path(
|
||||
__file__,
|
||||
'../..',
|
||||
'hot/software-config/elements',
|
||||
'heat-config-docker-cmd/install.d/hook-docker-cmd.py')
|
||||
|
||||
self.cleanup_path = self.relative_path(
|
||||
__file__,
|
||||
'../..',
|
||||
'hot/software-config/elements/heat-config-docker-cmd/',
|
||||
'os-refresh-config/configure.d/50-heat-config-docker-cmd')
|
||||
|
||||
self.fake_tool_path = self.relative_path(
|
||||
__file__,
|
||||
'config-tool-fake.py')
|
||||
|
||||
self.working_dir = self.useFixture(fixtures.TempDir())
|
||||
self.outputs_dir = self.useFixture(fixtures.TempDir())
|
||||
self.test_state_path = self.outputs_dir.join('test_state.json')
|
||||
|
||||
self.env = os.environ.copy()
|
||||
self.env.update({
|
||||
'HEAT_DOCKER_CMD_WORKING': self.working_dir.join(),
|
||||
'HEAT_DOCKER_CMD': self.fake_tool_path,
|
||||
'TEST_STATE_PATH': self.test_state_path,
|
||||
})
|
||||
|
||||
def test_hook(self):
|
||||
|
||||
self.env.update({
|
||||
'TEST_RESPONSE': json.dumps({
|
||||
'stdout': '',
|
||||
'stderr': 'Creating abcdef001_db_1...'
|
||||
})
|
||||
})
|
||||
returncode, stdout, stderr = self.run_cmd(
|
||||
[self.hook_path], self.env, json.dumps(self.data))
|
||||
|
||||
self.assertEqual(0, returncode, stderr)
|
||||
|
||||
self.assertEqual({
|
||||
'deploy_stdout': '',
|
||||
'deploy_stderr': 'Creating abcdef001_db_1...\n'
|
||||
'Creating abcdef001_db_1...',
|
||||
'deploy_status_code': 0
|
||||
}, json.loads(stdout))
|
||||
|
||||
state_0 = self.json_from_file(self.test_state_path)
|
||||
state_1 = self.json_from_file('%s_1' % self.test_state_path)
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'run',
|
||||
'--detach=true',
|
||||
'--name',
|
||||
'abcdef001__web',
|
||||
'xxx'
|
||||
], state_0['args'])
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'run',
|
||||
'--detach=true',
|
||||
'--name',
|
||||
'abcdef001__db',
|
||||
'--env=KOLLA_CONFIG_STRATEGY=COPY_ALWAYS',
|
||||
'--env=FOO=BAR',
|
||||
'--net=host',
|
||||
'--privileged=true',
|
||||
'--restart=always',
|
||||
'--user=root',
|
||||
'--volume=/run:/run',
|
||||
'--volume=db:/var/lib/db',
|
||||
'xxx'
|
||||
], state_1['args'])
|
||||
|
||||
def test_hook_failed(self):
|
||||
|
||||
self.env.update({
|
||||
'TEST_RESPONSE': json.dumps({
|
||||
'stdout': '',
|
||||
'stderr': 'Error: image library/xxx:latest not found',
|
||||
'returncode': 1
|
||||
})
|
||||
})
|
||||
returncode, stdout, stderr = self.run_cmd(
|
||||
[self.hook_path], self.env, json.dumps(self.data))
|
||||
|
||||
self.assertEqual({
|
||||
'deploy_stdout': '',
|
||||
'deploy_stderr': 'Error: image library/xxx:latest not found\n'
|
||||
'Error: image library/xxx:latest not found',
|
||||
'deploy_status_code': 1
|
||||
}, json.loads(stdout))
|
||||
|
||||
state_0 = self.json_from_file(self.test_state_path)
|
||||
state_1 = self.json_from_file('%s_1' % self.test_state_path)
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'run',
|
||||
'--detach=true',
|
||||
'--name',
|
||||
'abcdef001__web',
|
||||
'xxx'
|
||||
], state_0['args'])
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'run',
|
||||
'--detach=true',
|
||||
'--name',
|
||||
'abcdef001__db',
|
||||
'--env=KOLLA_CONFIG_STRATEGY=COPY_ALWAYS',
|
||||
'--env=FOO=BAR',
|
||||
'--net=host',
|
||||
'--privileged=true',
|
||||
'--restart=always',
|
||||
'--user=root',
|
||||
'--volume=/run:/run',
|
||||
'--volume=db:/var/lib/db',
|
||||
'xxx'
|
||||
], state_1['args'])
|
||||
|
||||
def test_cleanup_deleted(self):
|
||||
with tempfile.NamedTemporaryFile(delete=False) as f:
|
||||
f.write(json.dumps([self.data]))
|
||||
f.flush()
|
||||
self.env['HEAT_SHELL_CONFIG'] = f.name
|
||||
|
||||
returncode, stdout, stderr = self.run_cmd(
|
||||
[self.cleanup_path], self.env)
|
||||
|
||||
# on the first run, abcdef001.json is written out, no docker calls made
|
||||
configs_path = os.path.join(self.env['HEAT_DOCKER_CMD_WORKING'],
|
||||
'abcdef001.json')
|
||||
self.assertThat(configs_path, matchers.FileExists())
|
||||
self.assertThat(self.test_state_path,
|
||||
matchers.Not(matchers.FileExists()))
|
||||
|
||||
# run again with empty config data
|
||||
with tempfile.NamedTemporaryFile(delete=False) as f:
|
||||
f.write(json.dumps([]))
|
||||
f.flush()
|
||||
self.env['HEAT_SHELL_CONFIG'] = f.name
|
||||
|
||||
returncode, stdout, stderr = self.run_cmd(
|
||||
[self.cleanup_path], self.env)
|
||||
|
||||
# on the second run, abcdef001.json is deleted, docker rm is run on
|
||||
# both containers
|
||||
configs_path = os.path.join(self.env['HEAT_DOCKER_CMD_WORKING'],
|
||||
'abcdef001.json')
|
||||
self.assertThat(configs_path,
|
||||
matchers.Not(matchers.FileExists()))
|
||||
state_0 = self.json_from_file(self.test_state_path)
|
||||
state_1 = self.json_from_file('%s_1' % self.test_state_path)
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'rm',
|
||||
'-f',
|
||||
'abcdef001__web',
|
||||
], state_0['args'])
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'rm',
|
||||
'-f',
|
||||
'abcdef001__db',
|
||||
], state_1['args'])
|
||||
|
||||
def test_cleanup_changed(self):
|
||||
with tempfile.NamedTemporaryFile(delete=False) as f:
|
||||
f.write(json.dumps([self.data]))
|
||||
f.flush()
|
||||
self.env['HEAT_SHELL_CONFIG'] = f.name
|
||||
|
||||
returncode, stdout, stderr = self.run_cmd(
|
||||
[self.cleanup_path], self.env)
|
||||
|
||||
# on the first run, abcdef001.json is written out, no docker calls made
|
||||
configs_path = os.path.join(self.env['HEAT_DOCKER_CMD_WORKING'],
|
||||
'abcdef001.json')
|
||||
self.assertThat(configs_path, matchers.FileExists())
|
||||
self.assertThat(self.test_state_path,
|
||||
matchers.Not(matchers.FileExists()))
|
||||
|
||||
# run again with changed config data
|
||||
new_data = copy.deepcopy(self.data)
|
||||
new_data['config']['web']['image'] = 'yyy'
|
||||
with tempfile.NamedTemporaryFile(delete=False) as f:
|
||||
f.write(json.dumps([new_data]))
|
||||
f.flush()
|
||||
self.env['HEAT_SHELL_CONFIG'] = f.name
|
||||
|
||||
returncode, stdout, stderr = self.run_cmd(
|
||||
[self.cleanup_path], self.env)
|
||||
|
||||
# on the second run, abcdef001.json is written with the new data,
|
||||
# docker rm is run on both containers
|
||||
configs_path = os.path.join(self.env['HEAT_DOCKER_CMD_WORKING'],
|
||||
'abcdef001.json')
|
||||
self.assertThat(configs_path, matchers.FileExists())
|
||||
state_0 = self.json_from_file(self.test_state_path)
|
||||
state_1 = self.json_from_file('%s_1' % self.test_state_path)
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'rm',
|
||||
'-f',
|
||||
'abcdef001__web',
|
||||
], state_0['args'])
|
||||
self.assertEqual([
|
||||
self.fake_tool_path,
|
||||
'rm',
|
||||
'-f',
|
||||
'abcdef001__db',
|
||||
], state_1['args'])
|
Loading…
Reference in New Issue
Block a user