d2b8556aa6
The current open file limit of 1024 is not enough when executing operations that involve a large amount of subclouds in parallel (e.g. patching orchestration). dcmanager-audit already has its limit increased through the existing global option 'worker_rlimit_nofile', but the other services are not using this option. This commit moves the function that sets the limit to the utils.py module and makes the other dcmanager services use the defined limit. Test Plan: 1. PASS - Verify that the open file limit gets updated to the defined value by checking the service log file and by verifying the /proc/<pid>/limits file; 2. PASS - Verify that the new limit it enough to support DC patch orchestration with MAX_SUBCLOUD_GROUP_MAX_PARALLEL_SUBCLOUDS subclouds; 3. PASS - Verify that the limit can be changed by modifying the /etc/dcmanager/dcmanager.conf file with worker_rlimit_nofile option; 4. PASS - Verify that it's not possible to increase the soft limit past the hard limit. Story: 2010584 Task: 47751 Signed-off-by: Gustavo Herzmann <gustavo.herzmann@windriver.com> Change-Id: Ia52e532fd14cd2e61714477a561674d16ca45195
142 lines
5.1 KiB
Python
142 lines
5.1 KiB
Python
# Copyright (c) 2020-2021 Wind River Systems, Inc.
|
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
# you may not use this file except in compliance with the License.
|
|
# You may obtain a copy of the License at
|
|
#
|
|
# http://www.apache.org/licenses/LICENSE-2.0
|
|
#
|
|
# Unless required by applicable law or agreed to in writing, software
|
|
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
|
|
# implied.
|
|
# See the License for the specific language governing permissions and
|
|
# limitations under the License.
|
|
#
|
|
|
|
import six
|
|
|
|
import functools
|
|
from oslo_config import cfg
|
|
from oslo_log import log as logging
|
|
import oslo_messaging
|
|
from oslo_service import service
|
|
|
|
from dccommon.subprocess_cleanup import SubprocessCleanup
|
|
from dcmanager.common import consts
|
|
from dcmanager.common import context
|
|
from dcmanager.common import exceptions
|
|
from dcmanager.common import messaging as rpc_messaging
|
|
from dcmanager.common import scheduler
|
|
from dcmanager.common import utils
|
|
from dcmanager.orchestrator.sw_update_manager import SwUpdateManager
|
|
|
|
CONF = cfg.CONF
|
|
LOG = logging.getLogger(__name__)
|
|
|
|
|
|
def request_context(func):
|
|
@functools.wraps(func)
|
|
def wrapped(self, ctx, *args, **kwargs):
|
|
if ctx is not None and not isinstance(ctx, context.RequestContext):
|
|
ctx = context.RequestContext.from_dict(ctx.to_dict())
|
|
try:
|
|
return func(self, ctx, *args, **kwargs)
|
|
except exceptions.DCManagerException:
|
|
raise oslo_messaging.rpc.dispatcher.ExpectedException()
|
|
|
|
return wrapped
|
|
|
|
|
|
class DCManagerOrchestratorService(service.Service):
|
|
"""Lifecycle manager for a running orchestrator service."""
|
|
|
|
def __init__(self):
|
|
|
|
super(DCManagerOrchestratorService, self).__init__()
|
|
self.host = cfg.CONF.host
|
|
self.rpc_api_version = consts.RPC_API_VERSION
|
|
self.topic = consts.TOPIC_DC_MANAGER_ORCHESTRATOR
|
|
# The following are initialized here, but assigned in start() which
|
|
# happens after the fork when spawning multiple worker processes
|
|
self.TG = None
|
|
self.target = None
|
|
self._rpc_server = None
|
|
self.sw_update_manager = None
|
|
|
|
def start(self):
|
|
utils.set_open_file_limit(cfg.CONF.worker_rlimit_nofile)
|
|
self.init_tgm()
|
|
self.init_manager()
|
|
target = oslo_messaging.Target(version=self.rpc_api_version,
|
|
server=self.host,
|
|
topic=self.topic)
|
|
self.target = target
|
|
self._rpc_server = rpc_messaging.get_rpc_server(self.target, self)
|
|
self._rpc_server.start()
|
|
super(DCManagerOrchestratorService, self).start()
|
|
|
|
def init_tgm(self):
|
|
self.TG = scheduler.ThreadGroupManager()
|
|
|
|
def init_manager(self):
|
|
self.sw_update_manager = SwUpdateManager()
|
|
|
|
def _stop_rpc_server(self):
|
|
# Stop RPC connection to prevent new requests
|
|
LOG.debug("Attempting to stop RPC service...")
|
|
if self._rpc_server is not None:
|
|
try:
|
|
self._rpc_server.stop()
|
|
self._rpc_server.wait()
|
|
self._rpc_server = None
|
|
LOG.info('RPC service stopped successfully')
|
|
except Exception as ex:
|
|
LOG.error('Failed to stop engine service: %s',
|
|
six.text_type(ex))
|
|
|
|
def stop(self):
|
|
"""Stop anything initiated by start"""
|
|
SubprocessCleanup.shutdown_cleanup(origin="service")
|
|
self._stop_rpc_server()
|
|
if self.TG is not None:
|
|
self.TG.stop()
|
|
self.TG = None
|
|
if self.sw_update_manager is not None:
|
|
self.sw_update_manager.stop()
|
|
self.sw_update_manager = None
|
|
# Terminate the engine process
|
|
LOG.info("All threads were gone, terminating engine")
|
|
super(DCManagerOrchestratorService, self).stop()
|
|
|
|
@request_context
|
|
def create_sw_update_strategy(self, context, payload):
|
|
# Creates a software update strategy
|
|
LOG.info("Handling create_sw_update_strategy request of type %s" %
|
|
payload.get('type'))
|
|
return self.sw_update_manager.create_sw_update_strategy(
|
|
context, payload)
|
|
|
|
@request_context
|
|
def delete_sw_update_strategy(self, context, update_type=None):
|
|
# Deletes the software update strategy
|
|
LOG.info("Handling delete_sw_update_strategy request")
|
|
return self.sw_update_manager.delete_sw_update_strategy(
|
|
context,
|
|
update_type=update_type)
|
|
|
|
@request_context
|
|
def apply_sw_update_strategy(self, context, update_type=None):
|
|
# Applies the software update strategy
|
|
LOG.info("Handling apply_sw_update_strategy request")
|
|
return self.sw_update_manager.apply_sw_update_strategy(
|
|
context,
|
|
update_type=update_type)
|
|
|
|
@request_context
|
|
def abort_sw_update_strategy(self, context, update_type=None):
|
|
# Aborts the software update strategy
|
|
LOG.info("Handling abort_sw_update_strategy request")
|
|
return self.sw_update_manager.abort_sw_update_strategy(
|
|
context,
|
|
update_type=update_type)
|