379 lines
16 KiB
Python
379 lines
16 KiB
Python
# Copyright 2015 Hewlett-Packard Development Company, L.P.
|
|
#
|
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
|
|
# not use this file except in compliance with the License. You may obtain
|
|
# a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
|
|
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
|
|
# License for the specific language governing permissions and limitations
|
|
# under the License.
|
|
|
|
import datetime
|
|
import time
|
|
|
|
from oslo_config import cfg
|
|
from oslo_log import log as logging
|
|
from oslo_utils import excutils
|
|
import sqlalchemy
|
|
from stevedore import driver as stevedore_driver
|
|
|
|
from octavia.common import constants
|
|
from octavia.common import stats
|
|
from octavia.controller.healthmanager.health_drivers import update_base
|
|
from octavia.controller.healthmanager import update_serializer
|
|
from octavia.db import api as db_api
|
|
from octavia.db import repositories as repo
|
|
|
|
CONF = cfg.CONF
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
|
|
class UpdateHealthDb(update_base.HealthUpdateBase):
|
|
def __init__(self):
|
|
super(UpdateHealthDb, self).__init__()
|
|
# first setup repo for amphora, listener,member(nodes),pool repo
|
|
self.event_streamer = stevedore_driver.DriverManager(
|
|
namespace='octavia.controller.queues',
|
|
name=CONF.health_manager.event_streamer_driver,
|
|
invoke_on_load=True).driver
|
|
self.amphora_repo = repo.AmphoraRepository()
|
|
self.amphora_health_repo = repo.AmphoraHealthRepository()
|
|
self.listener_repo = repo.ListenerRepository()
|
|
self.loadbalancer_repo = repo.LoadBalancerRepository()
|
|
self.member_repo = repo.MemberRepository()
|
|
self.pool_repo = repo.PoolRepository()
|
|
self.sync_prv_status = CONF.health_manager.sync_provisioning_status
|
|
|
|
def emit(self, info_type, info_id, info_obj):
|
|
cnt = update_serializer.InfoContainer(info_type, info_id, info_obj)
|
|
self.event_streamer.emit(cnt)
|
|
|
|
def _update_status_and_emit_event(self, session, repo, entity_type,
|
|
entity_id, new_op_status, old_op_status,
|
|
current_prov_status):
|
|
message = {}
|
|
if old_op_status.lower() != new_op_status.lower():
|
|
LOG.debug("%s %s status has changed from %s to "
|
|
"%s. Updating db and sending event.",
|
|
entity_type, entity_id, old_op_status,
|
|
new_op_status)
|
|
repo.update(session, entity_id, operating_status=new_op_status)
|
|
if new_op_status == constants.DRAINING:
|
|
new_op_status = constants.ONLINE
|
|
message.update({constants.OPERATING_STATUS: new_op_status})
|
|
if self.sync_prv_status:
|
|
LOG.debug("%s %s provisioning_status %s. "
|
|
"Updating db and sending event.",
|
|
entity_type, entity_id, current_prov_status)
|
|
message.update(
|
|
{constants.PROVISIONING_STATUS: current_prov_status})
|
|
if message:
|
|
self.emit(entity_type, entity_id, message)
|
|
|
|
def update_health(self, health):
|
|
"""This function is to update db info based on amphora status
|
|
|
|
:param health: map object that contains amphora, listener, member info
|
|
:type map: string
|
|
:returns: null
|
|
|
|
The input health data structure is shown as below::
|
|
|
|
health = {
|
|
"id": self.FAKE_UUID_1,
|
|
"listeners": {
|
|
"listener-id-1": {"status": constants.OPEN, "pools": {
|
|
"pool-id-1": {"status": constants.UP,
|
|
"members": {
|
|
"member-id-1": constants.ONLINE}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
"""
|
|
session = db_api.get_session()
|
|
|
|
# We need to see if all of the listeners are reporting in
|
|
expected_listener_count = 0
|
|
db_lbs_on_amp = self.amphora_repo.get_all_lbs_on_amphora(session,
|
|
health['id'])
|
|
listeners = health['listeners']
|
|
|
|
# We need to loop over the lbs here to make sure we update the
|
|
# amphora_health record as soon as possible to prevent unnecessary
|
|
# failovers. Unfortunately that means looping over this list twice.
|
|
for db_lb in db_lbs_on_amp:
|
|
expected_listener_count += len(db_lb.listeners)
|
|
|
|
# Do not update amphora health if the reporting listener count
|
|
# does not match the expected listener count
|
|
if len(listeners) == expected_listener_count:
|
|
|
|
lock_session = db_api.get_session(autocommit=False)
|
|
|
|
# if we're running too far behind, warn and bail
|
|
proc_delay = time.time() - health['recv_time']
|
|
hb_interval = CONF.health_manager.heartbeat_interval
|
|
if proc_delay >= hb_interval:
|
|
LOG.warning('Amphora %(id)s health message was processed too '
|
|
'slowly: %(delay)ss! The system may be overloaded '
|
|
'or otherwise malfunctioning. This heartbeat has '
|
|
'been ignored and no update was made to the '
|
|
'amphora health entry. THIS IS NOT GOOD.',
|
|
{'id': health['id'], 'delay': proc_delay})
|
|
return
|
|
|
|
# if the input amphora is healthy, we update its db info
|
|
try:
|
|
self.amphora_health_repo.replace(
|
|
lock_session, health['id'],
|
|
last_update=(datetime.datetime.utcnow()))
|
|
lock_session.commit()
|
|
except Exception:
|
|
with excutils.save_and_reraise_exception():
|
|
lock_session.rollback()
|
|
else:
|
|
LOG.warning('Amphora %(id)s health message reports %(found)i '
|
|
'listeners when %(expected)i expected',
|
|
{'id': health['id'], 'found': len(listeners),
|
|
'expected': expected_listener_count})
|
|
|
|
for db_lb in db_lbs_on_amp:
|
|
|
|
processed_pools = []
|
|
|
|
# We got a heartbeat so lb is healthy until proven otherwise
|
|
if db_lb.enabled is False:
|
|
lb_status = constants.OFFLINE
|
|
else:
|
|
lb_status = constants.ONLINE
|
|
|
|
for db_listener in db_lb.listeners:
|
|
listener_status = None
|
|
listener_id = db_listener.id
|
|
listener = None
|
|
|
|
if listener_id not in listeners:
|
|
listener_status = constants.OFFLINE
|
|
else:
|
|
listener = listeners[listener_id]
|
|
|
|
# OPEN = HAProxy listener status nbconn < maxconn
|
|
if listener.get('status') == constants.OPEN:
|
|
listener_status = constants.ONLINE
|
|
# FULL = HAProxy listener status not nbconn < maxconn
|
|
elif listener.get('status') == constants.FULL:
|
|
listener_status = constants.DEGRADED
|
|
if lb_status == constants.ONLINE:
|
|
lb_status = constants.DEGRADED
|
|
else:
|
|
LOG.warning(('Listener %(list)s reported status of '
|
|
'%(status)s'),
|
|
{'list': listener_id,
|
|
'status': listener.get('status')})
|
|
|
|
try:
|
|
if listener_status is not None:
|
|
self._update_status_and_emit_event(
|
|
session, self.listener_repo, constants.LISTENER,
|
|
listener_id, listener_status,
|
|
db_listener.operating_status,
|
|
db_listener.provisioning_status
|
|
)
|
|
except sqlalchemy.orm.exc.NoResultFound:
|
|
LOG.error("Listener %s is not in DB", listener_id)
|
|
|
|
if not listener:
|
|
continue
|
|
|
|
pools = listener['pools']
|
|
|
|
# Process pools bound to listeners
|
|
for db_pool in db_listener.pools:
|
|
lb_status = self._process_pool_status(
|
|
session, db_pool, pools, lb_status, processed_pools)
|
|
|
|
# Process pools bound to the load balancer
|
|
for db_pool in db_lb.pools:
|
|
# Don't re-process pools shared with listeners
|
|
if db_pool.id in processed_pools:
|
|
continue
|
|
lb_status = self._process_pool_status(
|
|
session, db_pool, [], lb_status, processed_pools)
|
|
|
|
# Update the load balancer status last
|
|
try:
|
|
self._update_status_and_emit_event(
|
|
session, self.loadbalancer_repo,
|
|
constants.LOADBALANCER, db_lb.id, lb_status,
|
|
db_lb.operating_status, db_lb.provisioning_status
|
|
)
|
|
except sqlalchemy.orm.exc.NoResultFound:
|
|
LOG.error("Load balancer %s is not in DB", db_lb.id)
|
|
|
|
def _process_pool_status(self, session, db_pool, pools, lb_status,
|
|
processed_pools):
|
|
pool_status = None
|
|
pool_id = db_pool.id
|
|
|
|
if pool_id not in pools:
|
|
pool_status = constants.OFFLINE
|
|
else:
|
|
pool = pools[pool_id]
|
|
|
|
processed_pools.append(pool_id)
|
|
|
|
# UP = HAProxy backend has working or no servers
|
|
if pool.get('status') == constants.UP:
|
|
pool_status = constants.ONLINE
|
|
# DOWN = HAProxy backend has no working servers
|
|
elif pool.get('status') == constants.DOWN:
|
|
pool_status = constants.ERROR
|
|
lb_status = constants.ERROR
|
|
else:
|
|
LOG.warning(('Pool %(pool)s reported status of '
|
|
'%(status)s'),
|
|
{'pool': pool_id,
|
|
'status': pool.get('status')})
|
|
|
|
# Deal with the members that are reporting from
|
|
# the Amphora
|
|
members = pool['members']
|
|
for db_member in db_pool.members:
|
|
member_status = None
|
|
member_id = db_member.id
|
|
|
|
if member_id not in members:
|
|
member_status = constants.OFFLINE
|
|
else:
|
|
status = members[member_id]
|
|
|
|
# Member status can be "UP" or "UP #/#"
|
|
# (transitional)
|
|
if status.startswith(constants.UP):
|
|
member_status = constants.ONLINE
|
|
# Member status can be "DOWN" or "DOWN #/#"
|
|
# (transitional)
|
|
elif status.startswith(constants.DOWN):
|
|
member_status = constants.ERROR
|
|
if pool_status == constants.ONLINE:
|
|
pool_status = constants.DEGRADED
|
|
if lb_status == constants.ONLINE:
|
|
lb_status = constants.DEGRADED
|
|
elif status == constants.DRAIN:
|
|
member_status = constants.DRAINING
|
|
elif status == constants.MAINT:
|
|
member_status = constants.OFFLINE
|
|
elif status == constants.NO_CHECK:
|
|
member_status = constants.NO_MONITOR
|
|
else:
|
|
LOG.warning('Member %(mem)s reported '
|
|
'status of %(status)s',
|
|
{'mem': member_id,
|
|
'status': status})
|
|
|
|
try:
|
|
if member_status is not None:
|
|
self._update_status_and_emit_event(
|
|
session, self.member_repo,
|
|
constants.MEMBER,
|
|
member_id, member_status,
|
|
db_member.operating_status,
|
|
db_member.provisioning_status
|
|
)
|
|
except sqlalchemy.orm.exc.NoResultFound:
|
|
LOG.error("Member %s is not able to update "
|
|
"in DB", member_id)
|
|
|
|
try:
|
|
if pool_status is not None:
|
|
self._update_status_and_emit_event(
|
|
session, self.pool_repo, constants.POOL,
|
|
pool_id, pool_status, db_pool.operating_status,
|
|
db_pool.provisioning_status
|
|
)
|
|
except sqlalchemy.orm.exc.NoResultFound:
|
|
LOG.error("Pool %s is not in DB", pool_id)
|
|
|
|
return lb_status
|
|
|
|
|
|
class UpdateStatsDb(update_base.StatsUpdateBase, stats.StatsMixin):
|
|
|
|
def __init__(self):
|
|
super(UpdateStatsDb, self).__init__()
|
|
self.event_streamer = stevedore_driver.DriverManager(
|
|
namespace='octavia.controller.queues',
|
|
name=CONF.health_manager.event_streamer_driver,
|
|
invoke_on_load=True).driver
|
|
self.repo_listener = repo.ListenerRepository()
|
|
|
|
def emit(self, info_type, info_id, info_obj):
|
|
cnt = update_serializer.InfoContainer(info_type, info_id, info_obj)
|
|
self.event_streamer.emit(cnt)
|
|
|
|
def update_stats(self, health_message):
|
|
"""This function is to update the db with listener stats
|
|
|
|
:param health_message: The health message containing the listener stats
|
|
:type map: string
|
|
:returns: null
|
|
|
|
Example::
|
|
|
|
health = {
|
|
"id": self.FAKE_UUID_1,
|
|
"listeners": {
|
|
"listener-id-1": {
|
|
"status": constants.OPEN,
|
|
"stats": {
|
|
"ereq":0,
|
|
"conns": 0,
|
|
"totconns": 0,
|
|
"rx": 0,
|
|
"tx": 0,
|
|
},
|
|
"pools": {
|
|
"pool-id-1": {
|
|
"status": constants.UP,
|
|
"members": {"member-id-1": constants.ONLINE}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
"""
|
|
session = db_api.get_session()
|
|
|
|
amphora_id = health_message['id']
|
|
listeners = health_message['listeners']
|
|
for listener_id, listener in listeners.items():
|
|
|
|
stats = listener.get('stats')
|
|
stats = {'bytes_in': stats['rx'], 'bytes_out': stats['tx'],
|
|
'active_connections': stats['conns'],
|
|
'total_connections': stats['totconns'],
|
|
'request_errors': stats['ereq']}
|
|
LOG.debug("Updating listener stats in db and sending event.")
|
|
LOG.debug("Listener %s / Amphora %s stats: %s",
|
|
listener_id, amphora_id, stats)
|
|
self.listener_stats_repo.replace(
|
|
session, listener_id, amphora_id, **stats)
|
|
|
|
listener_stats = self.get_listener_stats(session, listener_id)
|
|
self.emit(
|
|
'listener_stats', listener_id, listener_stats.get_stats())
|
|
|
|
listener_db = self.repo_listener.get(session, id=listener_id)
|
|
lb_stats = self.get_loadbalancer_stats(
|
|
session, listener_db.load_balancer_id)
|
|
self.emit('loadbalancer_stats',
|
|
listener_db.load_balancer_id, lb_stats.get_stats())
|