deb-gnocchi/gnocchi/cli.py

205 lines
6.7 KiB
Python

# Copyright (c) 2013 Mirantis Inc.
# Copyright (c) 2015 Red Hat
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
import logging
import multiprocessing
import signal
import sys
import time
from oslo_config import cfg
from oslo_utils import timeutils
import retrying
import six
from gnocchi import archive_policy
from gnocchi import indexer
from gnocchi.rest import app
from gnocchi import service
from gnocchi import statsd as statsd_service
from gnocchi import storage
LOG = logging.getLogger(__name__)
def upgrade():
conf = cfg.ConfigOpts()
conf.register_cli_opts([
cfg.BoolOpt("skip-index", default=False,
help="Skip index upgrade."),
cfg.BoolOpt("skip-storage", default=False,
help="Skip storage upgrade."),
cfg.BoolOpt("skip-archive-policies-creation", default=False,
help="Skip default archive policies creation."),
cfg.BoolOpt("create-legacy-resource-types", default=False,
help="Creation of Ceilometer legacy resource types.")
])
conf = service.prepare_service(conf=conf)
index = indexer.get_driver(conf)
index.connect()
if not conf.skip_index:
LOG.info("Upgrading indexer %s" % index)
index.upgrade(
create_legacy_resource_types=conf.create_legacy_resource_types)
if not conf.skip_storage:
s = storage.get_driver(conf)
LOG.info("Upgrading storage %s" % s)
s.upgrade(index)
if (not conf.skip_archive_policies_creation
and not index.list_archive_policies()
and not index.list_archive_policy_rules()):
for name, ap in six.iteritems(archive_policy.DEFAULT_ARCHIVE_POLICIES):
index.create_archive_policy(ap)
index.create_archive_policy_rule("default", "*", "low")
def api():
app.build_server()
def statsd():
statsd_service.start()
class MetricProcessBase(multiprocessing.Process):
def __init__(self, conf, worker_id=0, interval_delay=0):
super(MetricProcessBase, self).__init__()
self.conf = conf
self.worker_id = worker_id
self.startup_delay = worker_id
self.interval_delay = interval_delay
# Retry with exponential backoff for up to 5 minutes
@retrying.retry(wait_exponential_multiplier=500,
wait_exponential_max=60000,
stop_max_delay=300000)
def _configure(self):
self.store = storage.get_driver(self.conf)
self.store.partition = self.worker_id
self.index = indexer.get_driver(self.conf)
self.index.connect()
def run(self):
self._configure()
# Delay startup so workers are jittered.
time.sleep(self.startup_delay)
while True:
try:
with timeutils.StopWatch() as timer:
self._run_job()
time.sleep(max(0, self.interval_delay - timer.elapsed()))
except KeyboardInterrupt:
# Ignore KeyboardInterrupt so parent handler can kill
# all children.
pass
@staticmethod
def _run_job():
raise NotImplementedError
class MetricReporting(MetricProcessBase):
def __init__(self, conf, worker_id=0, interval_delay=0, queues=None):
super(MetricReporting, self).__init__(conf, worker_id, interval_delay)
self.queues = queues
def _run_job(self):
try:
report = self.store.measures_report(details=False)
if self.queues:
block_size = max(16, min(
256, report['summary']['metrics'] // len(self.queues)))
for queue in self.queues:
queue.put(block_size)
LOG.info("Metricd reporting: %d measurements bundles across %d "
"metrics wait to be processed.",
report['summary']['measures'],
report['summary']['metrics'])
except Exception:
LOG.error("Unexpected error during pending measures reporting",
exc_info=True)
class MetricProcessor(MetricProcessBase):
def __init__(self, conf, worker_id=0, interval_delay=0, queue=None):
super(MetricProcessor, self).__init__(conf, worker_id, interval_delay)
self.queue = queue
self.block_size = 128
def _run_job(self):
try:
if self.queue:
while not self.queue.empty():
self.block_size = self.queue.get()
LOG.debug("Re-configuring worker to handle up to %s "
"metrics", self.block_size)
self.store.process_background_tasks(self.index, self.block_size)
except Exception:
LOG.error("Unexpected error during measures processing",
exc_info=True)
def metricd():
conf = service.prepare_service()
if (conf.storage.metric_reporting_delay <
conf.storage.metric_processing_delay):
LOG.error("Metric reporting must run less frequently then processing")
sys.exit(0)
signal.signal(signal.SIGTERM, _metricd_terminate)
try:
queues = []
workers = []
for worker in range(conf.metricd.workers):
queue = multiprocessing.Queue()
metric_worker = MetricProcessor(
conf, worker, conf.storage.metric_processing_delay, queue)
metric_worker.start()
queues.append(queue)
workers.append(metric_worker)
metric_report = MetricReporting(
conf, 0, conf.storage.metric_reporting_delay, queues)
metric_report.start()
workers.append(metric_report)
for worker in workers:
worker.join()
except KeyboardInterrupt:
_metricd_cleanup(workers)
sys.exit(0)
except Exception:
LOG.warning("exiting", exc_info=True)
_metricd_cleanup(workers)
sys.exit(1)
def _metricd_cleanup(workers):
for worker in workers:
if hasattr(worker, 'queue'):
worker.queue.close()
worker.terminate()
for worker in workers:
worker.join()
def _metricd_terminate(signum, frame):
_metricd_cleanup(multiprocessing.active_children())
sys.exit(0)