Add ability to save logs received from the ramdisk
Change-Id: I44abe5f41c83eafd5042434bf2556b8309f4a42a Partial-Bug: #1439647
This commit is contained in:
parent
d80c7d5ee5
commit
69cf5306aa
|
@ -76,6 +76,8 @@ the ramdisk. Request body: JSON dictionary with at least these keys:
|
||||||
* ``error`` optional error happened during ramdisk run, interpreted by
|
* ``error`` optional error happened during ramdisk run, interpreted by
|
||||||
``ramdisk_error`` plugin
|
``ramdisk_error`` plugin
|
||||||
|
|
||||||
|
* ``logs`` optional base64-encoded logs from the ramdisk
|
||||||
|
|
||||||
* ``block_devices`` optional block devices information for
|
* ``block_devices`` optional block devices information for
|
||||||
``root_device_hint`` plugin, dictionary with keys:
|
``root_device_hint`` plugin, dictionary with keys:
|
||||||
|
|
||||||
|
|
|
@ -294,7 +294,8 @@ These are plugins that are enabled by default and should not be disabled,
|
||||||
unless you understand what you're doing:
|
unless you understand what you're doing:
|
||||||
|
|
||||||
``ramdisk_error``
|
``ramdisk_error``
|
||||||
reports error, if ``error`` field is set by the ramdisk.
|
reports error, if ``error`` field is set by the ramdisk, also optionally
|
||||||
|
stores logs from ``logs`` field, see `HTTP API`_ for details.
|
||||||
``scheduler``
|
``scheduler``
|
||||||
validates and updates basic hardware scheduling properties: CPU number and
|
validates and updates basic hardware scheduling properties: CPU number and
|
||||||
architecture, memory and disk size.
|
architecture, memory and disk size.
|
||||||
|
|
|
@ -120,6 +120,9 @@ SERVICE_OPTS = [
|
||||||
cfg.BoolOpt('debug',
|
cfg.BoolOpt('debug',
|
||||||
default=False,
|
default=False,
|
||||||
help='Debug mode enabled/disabled.'),
|
help='Debug mode enabled/disabled.'),
|
||||||
|
cfg.StrOpt('ramdisk_logs_dir',
|
||||||
|
help='If set, logs from ramdisk will be stored in this '
|
||||||
|
'directory'),
|
||||||
cfg.BoolOpt('ports_for_inactive_interfaces',
|
cfg.BoolOpt('ports_for_inactive_interfaces',
|
||||||
default=False,
|
default=False,
|
||||||
help='DEPRECATED: use add_ports.'),
|
help='DEPRECATED: use add_ports.'),
|
||||||
|
|
|
@ -13,7 +13,10 @@
|
||||||
|
|
||||||
"""Standard set of plugins."""
|
"""Standard set of plugins."""
|
||||||
|
|
||||||
|
import base64
|
||||||
|
import datetime
|
||||||
import logging
|
import logging
|
||||||
|
import os
|
||||||
import sys
|
import sys
|
||||||
|
|
||||||
from oslo_config import cfg
|
from oslo_config import cfg
|
||||||
|
@ -160,8 +163,31 @@ class ValidateInterfacesHook(base.ProcessingHook):
|
||||||
class RamdiskErrorHook(base.ProcessingHook):
|
class RamdiskErrorHook(base.ProcessingHook):
|
||||||
"""Hook to process error send from the ramdisk."""
|
"""Hook to process error send from the ramdisk."""
|
||||||
|
|
||||||
|
DATETIME_FORMAT = '%Y.%m.%d_%H.%M.%S_%f'
|
||||||
|
|
||||||
def before_processing(self, node_info):
|
def before_processing(self, node_info):
|
||||||
if not node_info.get('error'):
|
if not node_info.get('error'):
|
||||||
return
|
return
|
||||||
|
|
||||||
|
logs = node_info.get('logs')
|
||||||
|
if logs:
|
||||||
|
self._store_logs(logs, node_info)
|
||||||
|
|
||||||
raise utils.Error(_('Ramdisk reported error: %s') % node_info['error'])
|
raise utils.Error(_('Ramdisk reported error: %s') % node_info['error'])
|
||||||
|
|
||||||
|
def _store_logs(self, logs, node_info):
|
||||||
|
if not CONF.discoverd.ramdisk_logs_dir:
|
||||||
|
LOG.warn(_LW('Failed to store logs received from the discovery '
|
||||||
|
'ramdisk because ramdisk_logs_dir configuration '
|
||||||
|
'option is not set'))
|
||||||
|
return
|
||||||
|
|
||||||
|
if not os.path.exists(CONF.discoverd.ramdisk_logs_dir):
|
||||||
|
os.makedirs(CONF.discoverd.ramdisk_logs_dir)
|
||||||
|
|
||||||
|
time_fmt = datetime.datetime.utcnow().strftime(self.DATETIME_FORMAT)
|
||||||
|
bmc_address = node_info.get('ipmi_address', 'unknown')
|
||||||
|
file_name = 'bmc_%s_%s' % (bmc_address, time_fmt)
|
||||||
|
with open(os.path.join(CONF.discoverd.ramdisk_logs_dir, file_name),
|
||||||
|
'wb') as fp:
|
||||||
|
fp.write(base64.b64decode(logs))
|
||||||
|
|
|
@ -0,0 +1,83 @@
|
||||||
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
# you may not use this file except in compliance with the License.
|
||||||
|
# You may obtain a copy of the License at
|
||||||
|
#
|
||||||
|
# http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
#
|
||||||
|
# Unless required by applicable law or agreed to in writing, software
|
||||||
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
||||||
|
# implied.
|
||||||
|
# See the License for the specific language governing permissions and
|
||||||
|
# limitations under the License.
|
||||||
|
|
||||||
|
import base64
|
||||||
|
import os
|
||||||
|
import shutil
|
||||||
|
import tempfile
|
||||||
|
|
||||||
|
from oslo_config import cfg
|
||||||
|
|
||||||
|
from ironic_discoverd import process
|
||||||
|
from ironic_discoverd.test import base as test_base
|
||||||
|
from ironic_discoverd import utils
|
||||||
|
|
||||||
|
CONF = cfg.CONF
|
||||||
|
|
||||||
|
|
||||||
|
class TestRamdiskError(test_base.BaseTest):
|
||||||
|
def setUp(self):
|
||||||
|
super(TestRamdiskError, self).setUp()
|
||||||
|
self.msg = 'BOOM'
|
||||||
|
self.bmc_address = '1.2.3.4'
|
||||||
|
self.data = {
|
||||||
|
'error': self.msg,
|
||||||
|
'ipmi_address': self.bmc_address,
|
||||||
|
}
|
||||||
|
|
||||||
|
self.tempdir = tempfile.mkdtemp()
|
||||||
|
self.addCleanup(lambda: shutil.rmtree(self.tempdir))
|
||||||
|
CONF.set_override('ramdisk_logs_dir', self.tempdir, 'discoverd')
|
||||||
|
|
||||||
|
def test_no_logs(self):
|
||||||
|
self.assertRaisesRegexp(utils.Error,
|
||||||
|
self.msg,
|
||||||
|
process.process, self.data)
|
||||||
|
self.assertEqual([], os.listdir(self.tempdir))
|
||||||
|
|
||||||
|
def test_logs_disabled(self):
|
||||||
|
self.data['logs'] = 'some log'
|
||||||
|
CONF.set_override('ramdisk_logs_dir', None, 'discoverd')
|
||||||
|
|
||||||
|
self.assertRaisesRegexp(utils.Error,
|
||||||
|
self.msg,
|
||||||
|
process.process, self.data)
|
||||||
|
self.assertEqual([], os.listdir(self.tempdir))
|
||||||
|
|
||||||
|
def test_logs(self):
|
||||||
|
log = b'log contents'
|
||||||
|
self.data['logs'] = base64.b64encode(log)
|
||||||
|
|
||||||
|
self.assertRaisesRegexp(utils.Error,
|
||||||
|
self.msg,
|
||||||
|
process.process, self.data)
|
||||||
|
|
||||||
|
files = os.listdir(self.tempdir)
|
||||||
|
self.assertEqual(1, len(files))
|
||||||
|
filename = files[0]
|
||||||
|
self.assertTrue(filename.startswith('bmc_%s_' % self.bmc_address),
|
||||||
|
'%s does not start with bmc_%s'
|
||||||
|
% (filename, self.bmc_address))
|
||||||
|
with open(os.path.join(self.tempdir, filename), 'rb') as fp:
|
||||||
|
self.assertEqual(log, fp.read())
|
||||||
|
|
||||||
|
def test_logs_create_dir(self):
|
||||||
|
shutil.rmtree(self.tempdir)
|
||||||
|
self.data['logs'] = base64.b64encode(b'log')
|
||||||
|
|
||||||
|
self.assertRaisesRegexp(utils.Error,
|
||||||
|
self.msg,
|
||||||
|
process.process, self.data)
|
||||||
|
|
||||||
|
files = os.listdir(self.tempdir)
|
||||||
|
self.assertEqual(1, len(files))
|
|
@ -241,15 +241,6 @@ class TestProcess(BaseTest):
|
||||||
process_mock.assert_called_once_with(cli, cli.node.get.return_value,
|
process_mock.assert_called_once_with(cli, cli.node.get.return_value,
|
||||||
self.data, pop_mock.return_value)
|
self.data, pop_mock.return_value)
|
||||||
|
|
||||||
@prepare_mocks
|
|
||||||
def test_error(self, cli, pop_mock, process_mock):
|
|
||||||
self.data['error'] = 'BOOM'
|
|
||||||
|
|
||||||
self.assertRaisesRegexp(utils.Error,
|
|
||||||
'BOOM',
|
|
||||||
process.process, self.data)
|
|
||||||
self.assertFalse(process_mock.called)
|
|
||||||
|
|
||||||
@prepare_mocks
|
@prepare_mocks
|
||||||
def test_missing_required(self, cli, pop_mock, process_mock):
|
def test_missing_required(self, cli, pop_mock, process_mock):
|
||||||
del self.data['cpus']
|
del self.data['cpus']
|
||||||
|
|
Loading…
Reference in New Issue