baece23027
Following setup adds version of monasca-persister into oslo setup Change-Id: I22aaaac2fcfdd2849b44a95979c23473488dd2c3
200 lines
7.3 KiB
Python
200 lines
7.3 KiB
Python
# (C) Copyright 2014-2017 Hewlett Packard Enterprise Development LP
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
|
# implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
|
|
"""Persister Module
|
|
|
|
The Persister reads metrics and alarms from Kafka and then stores them
|
|
in into either Influxdb or Cassandra
|
|
|
|
Start the perister as stand-alone process by running 'persister.py
|
|
--config-file <config file>'
|
|
"""
|
|
import multiprocessing
|
|
import os
|
|
import signal
|
|
import sys
|
|
import time
|
|
|
|
from monasca_common.simport import simport
|
|
from oslo_config import cfg
|
|
from oslo_log import log
|
|
|
|
from monasca_persister.repositories import persister
|
|
from monasca_persister import version
|
|
|
|
LOG = log.getLogger(__name__)
|
|
|
|
zookeeper_opts = [cfg.StrOpt('uri'),
|
|
cfg.IntOpt('partition_interval_recheck_seconds')]
|
|
|
|
zookeeper_group = cfg.OptGroup(name='zookeeper', title='zookeeper')
|
|
cfg.CONF.register_group(zookeeper_group)
|
|
cfg.CONF.register_opts(zookeeper_opts, zookeeper_group)
|
|
|
|
kafka_common_opts = [cfg.StrOpt('uri'),
|
|
cfg.StrOpt('group_id'),
|
|
cfg.StrOpt('topic'),
|
|
cfg.StrOpt('consumer_id'),
|
|
cfg.StrOpt('client_id'),
|
|
cfg.IntOpt('database_batch_size'),
|
|
cfg.IntOpt('max_wait_time_seconds'),
|
|
cfg.IntOpt('fetch_size_bytes'),
|
|
cfg.IntOpt('buffer_size'),
|
|
cfg.IntOpt('max_buffer_size'),
|
|
cfg.StrOpt('zookeeper_path'),
|
|
cfg.IntOpt('num_processors')]
|
|
|
|
kafka_metrics_opts = kafka_common_opts
|
|
kafka_alarm_history_opts = kafka_common_opts
|
|
|
|
kafka_metrics_group = cfg.OptGroup(name='kafka_metrics', title='kafka_metrics')
|
|
kafka_alarm_history_group = cfg.OptGroup(name='kafka_alarm_history',
|
|
title='kafka_alarm_history')
|
|
|
|
cfg.CONF.register_group(kafka_metrics_group)
|
|
cfg.CONF.register_group(kafka_alarm_history_group)
|
|
cfg.CONF.register_opts(kafka_metrics_opts, kafka_metrics_group)
|
|
cfg.CONF.register_opts(kafka_alarm_history_opts, kafka_alarm_history_group)
|
|
|
|
repositories_opts = [
|
|
cfg.StrOpt('metrics_driver', help='The repository driver to use for metrics'),
|
|
cfg.StrOpt('alarm_state_history_driver', help='The repository driver to use for alarm state history')]
|
|
|
|
repositories_group = cfg.OptGroup(name='repositories', title='repositories')
|
|
cfg.CONF.register_group(repositories_group)
|
|
cfg.CONF.register_opts(repositories_opts, repositories_group)
|
|
|
|
processors = [] # global list to facilitate clean signal handling
|
|
exiting = False
|
|
|
|
|
|
def clean_exit(signum, frame=None):
|
|
"""Exit all processes attempting to finish uncommitted active work before exit.
|
|
Can be called on an os signal or no zookeeper losing connection.
|
|
"""
|
|
global exiting
|
|
if exiting:
|
|
# Since this is set up as a handler for SIGCHLD when this kills one
|
|
# child it gets another signal, the global exiting avoids this running
|
|
# multiple times.
|
|
LOG.debug('Exit in progress clean_exit received additional signal %s' % signum)
|
|
return
|
|
|
|
LOG.info('Received signal %s, beginning graceful shutdown.' % signum)
|
|
exiting = True
|
|
wait_for_exit = False
|
|
|
|
for process in processors:
|
|
try:
|
|
if process.is_alive():
|
|
process.terminate() # Sends sigterm which any processes after a notification is sent attempt to handle
|
|
wait_for_exit = True
|
|
except Exception: # nosec
|
|
# There is really nothing to do if the kill fails, so just go on.
|
|
# The # nosec keeps bandit from reporting this as a security issue
|
|
pass
|
|
|
|
# wait for a couple seconds to give the subprocesses a chance to shut down correctly.
|
|
if wait_for_exit:
|
|
time.sleep(2)
|
|
|
|
# Kill everything, that didn't already die
|
|
for child in multiprocessing.active_children():
|
|
LOG.debug('Killing pid %s' % child.pid)
|
|
try:
|
|
os.kill(child.pid, signal.SIGKILL)
|
|
except Exception: # nosec
|
|
# There is really nothing to do if the kill fails, so just go on.
|
|
# The # nosec keeps bandit from reporting this as a security issue
|
|
pass
|
|
|
|
if signum == signal.SIGTERM:
|
|
sys.exit(0)
|
|
|
|
sys.exit(signum)
|
|
|
|
|
|
def start_process(respository, kafka_config):
|
|
LOG.info("start process: {}".format(respository))
|
|
m_persister = persister.Persister(kafka_config, cfg.CONF.zookeeper,
|
|
respository)
|
|
m_persister.run()
|
|
|
|
|
|
def _init_config():
|
|
log.register_options(cfg.CONF)
|
|
log.set_defaults()
|
|
cfg.CONF(prog='persister',
|
|
project='monasca',
|
|
version=version.version_str,
|
|
description='Persists metrics & alarm history in TSDB')
|
|
log.setup(cfg.CONF, 'monasca-persister', version=version.version_str)
|
|
|
|
|
|
def main():
|
|
"""Start persister."""
|
|
|
|
_init_config()
|
|
|
|
metric_repository = simport.load(cfg.CONF.repositories.metrics_driver)
|
|
alarm_state_history_repository = simport.load(cfg.CONF.repositories.alarm_state_history_driver)
|
|
|
|
# Add processors for metrics topic
|
|
for proc in range(0, cfg.CONF.kafka_metrics.num_processors):
|
|
processors.append(multiprocessing.Process(
|
|
target=start_process, args=(metric_repository, cfg.CONF.kafka_metrics)))
|
|
|
|
# Add processors for alarm history topic
|
|
for proc in range(0, cfg.CONF.kafka_alarm_history.num_processors):
|
|
processors.append(multiprocessing.Process(
|
|
target=start_process, args=(alarm_state_history_repository, cfg.CONF.kafka_alarm_history)))
|
|
|
|
# Start
|
|
try:
|
|
LOG.info('''
|
|
|
|
_____
|
|
/ \ ____ ____ _____ ______ ____ _____
|
|
/ \ / \ / _ \ / \\\__ \ / ___// ___\\\__ \\
|
|
/ Y ( <_> ) | \/ __ \_\___ \\ \___ / __ \\_
|
|
\____|__ /\____/|___| (____ /____ >\___ >____ /
|
|
\/ \/ \/ \/ \/ \/
|
|
__________ .__ __
|
|
\______ \ ___________ _____|__| _______/ |_ ___________
|
|
| ___// __ \_ __ \/ ___/ |/ ___/\ __\/ __ \_ __ \\
|
|
| | \ ___/| | \/\___ \| |\___ \ | | \ ___/| | \/
|
|
|____| \___ >__| /____ >__/____ > |__| \___ >__|
|
|
\/ \/ \/ \/
|
|
|
|
''')
|
|
for process in processors:
|
|
process.start()
|
|
|
|
# The signal handlers must be added after the processes start otherwise
|
|
# they run on all processes
|
|
signal.signal(signal.SIGCHLD, clean_exit)
|
|
signal.signal(signal.SIGINT, clean_exit)
|
|
signal.signal(signal.SIGTERM, clean_exit)
|
|
|
|
while True:
|
|
time.sleep(10)
|
|
|
|
except Exception:
|
|
LOG.exception('Error! Exiting.')
|
|
clean_exit(signal.SIGKILL)
|
|
|
|
if __name__ == "__main__":
|
|
sys.exit(main())
|