diff --git a/stackalytics/dashboard/decorators.py b/stackalytics/dashboard/decorators.py index ab76c724c..7c30bf153 100644 --- a/stackalytics/dashboard/decorators.py +++ b/stackalytics/dashboard/decorators.py @@ -32,6 +32,7 @@ from stackalytics.processor import utils from stackalytics import version as stackalytics_version +CONF = cfg.CONF LOG = logging.getLogger(__name__) @@ -498,7 +499,7 @@ def profiler_decorator(func): @functools.wraps(func) def profiler_decorated_function(*args, **kwargs): profiler = None - profile_filename = cfg.CONF.collect_profiler_stats + profile_filename = CONF.collect_profiler_stats if profile_filename: LOG.debug('Profiler is enabled') diff --git a/stackalytics/dashboard/parameters.py b/stackalytics/dashboard/parameters.py index 3e9528ca9..433dd86e5 100644 --- a/stackalytics/dashboard/parameters.py +++ b/stackalytics/dashboard/parameters.py @@ -21,6 +21,8 @@ import time from stackalytics.dashboard import vault +CONF = cfg.CONF + DEFAULTS = { 'review_nth': 5, } @@ -66,7 +68,7 @@ DEFAULT_STATIC_ACTIVITY_SIZE = 100 def get_default(param_name): if 'release' not in DEFAULTS: - release = cfg.CONF.default_release + release = CONF.default_release if not release: runtime_storage_inst = vault.get_runtime_storage() releases = runtime_storage_inst.get_by_key('releases') @@ -80,8 +82,8 @@ def get_default(param_name): else: release = 'all' DEFAULTS['release'] = release.lower() - DEFAULTS['metric'] = cfg.CONF.default_metric.lower() - DEFAULTS['project_type'] = cfg.CONF.default_project_type.lower() + DEFAULTS['metric'] = CONF.default_metric.lower() + DEFAULTS['project_type'] = CONF.default_project_type.lower() if param_name in DEFAULTS: return DEFAULTS[param_name] diff --git a/stackalytics/dashboard/vault.py b/stackalytics/dashboard/vault.py index 5c6707e54..8384f73e2 100644 --- a/stackalytics/dashboard/vault.py +++ b/stackalytics/dashboard/vault.py @@ -28,6 +28,7 @@ from stackalytics.processor import user_processor from stackalytics.processor import utils +CONF = cfg.CONF LOG = logging.getLogger(__name__) @@ -79,7 +80,7 @@ def get_vault(): try: vault = {} runtime_storage_inst = runtime_storage.get_runtime_storage( - cfg.CONF.runtime_storage_uri) + CONF.runtime_storage_uri) vault['runtime_storage'] = runtime_storage_inst vault['memory_storage'] = memory_storage.get_memory_storage( memory_storage.MEMORY_STORAGE_CACHED) @@ -97,7 +98,7 @@ def get_vault(): flask.request.stackalytics_updated = True vault['vault_update_time'] = time_now vault['vault_next_update_time'] = ( - time_now + cfg.CONF.dashboard_update_interval) + time_now + CONF.dashboard_update_interval) memory_storage_inst = vault['memory_storage'] have_updates = memory_storage_inst.update(compact_records( vault['runtime_storage'].get_update(os.getpid()))) diff --git a/stackalytics/dashboard/web.py b/stackalytics/dashboard/web.py index f4d61c8fb..8db4c7a94 100644 --- a/stackalytics/dashboard/web.py +++ b/stackalytics/dashboard/web.py @@ -42,8 +42,8 @@ app.register_blueprint(kpi.blueprint) LOG = logging.getLogger(__name__) -conf = cfg.CONF -conf.register_opts(config.CONNECTION_OPTS + config.DASHBOARD_OPTS) +CONF = cfg.CONF +CONF.register_opts(config.CONNECTION_OPTS + config.DASHBOARD_OPTS) # Handlers --------- @@ -680,20 +680,20 @@ def too_old(timestamp): def main(): - logging.register_options(conf) + logging.register_options(CONF) logging.set_defaults() conf_file = os.getenv('STACKALYTICS_CONF') if conf_file and os.path.isfile(conf_file): - conf(default_config_files=[conf_file]) - app.config['DEBUG'] = cfg.CONF.debug + CONF(default_config_files=[conf_file]) + app.config['DEBUG'] = CONF.debug LOG.info('Stackalytics.dashboard is configured via "%s"', conf_file) else: - conf(project='stackalytics') + CONF(project='stackalytics') - logging.setup(conf, 'stackalytics.dashboard') + logging.setup(CONF, 'stackalytics.dashboard') - app.run(cfg.CONF.listen_host, cfg.CONF.listen_port) + app.run(CONF.listen_host, CONF.listen_port) if __name__ == '__main__': main() diff --git a/stackalytics/processor/default_data_processor.py b/stackalytics/processor/default_data_processor.py index 130a3c669..1012fb6f9 100644 --- a/stackalytics/processor/default_data_processor.py +++ b/stackalytics/processor/default_data_processor.py @@ -28,6 +28,7 @@ from stackalytics.processor import rcs from stackalytics.processor import user_processor from stackalytics.processor import utils +CONF = cfg.CONF LOG = logging.getLogger(__name__) GITHUB_URI_PREFIX = r'^github:\/\/' @@ -50,7 +51,7 @@ def _check_default_data_change(runtime_storage_inst, default_data): def _retrieve_project_list_from_sources(project_sources): for project_source in project_sources: - uri = project_source.get('uri') or cfg.CONF.review_uri + uri = project_source.get('uri') or CONF.review_uri repo_iterator = [] if re.search(rcs.GERRIT_URI_PREFIX, uri): repo_iterator = _retrieve_project_list_from_gerrit(project_source) @@ -66,11 +67,11 @@ def _retrieve_project_list_from_sources(project_sources): def _retrieve_project_list_from_gerrit(project_source): LOG.info('Retrieving project list from Gerrit') try: - uri = project_source.get('uri') or cfg.CONF.review_uri + uri = project_source.get('uri') or CONF.review_uri gerrit_inst = rcs.Gerrit(uri) key_filename = (project_source.get('ssh_key_filename') or - cfg.CONF.ssh_key_filename) - username = project_source.get('ssh_username') or cfg.CONF.ssh_username + CONF.ssh_key_filename) + username = project_source.get('ssh_username') or CONF.ssh_username gerrit_inst.setup(key_filename=key_filename, username=username) project_list = gerrit_inst.get_project_list() @@ -83,7 +84,7 @@ def _retrieve_project_list_from_gerrit(project_source): LOG.debug('Get list of projects for organization %s', organization) git_repos = [f for f in project_list if f.startswith(organization + "/")] - git_base_uri = project_source.get('git_base_uri') or cfg.CONF.git_base_uri + git_base_uri = project_source.get('git_base_uri') or CONF.git_base_uri for repo in git_repos: (org, name) = repo.split('/') diff --git a/stackalytics/processor/dump.py b/stackalytics/processor/dump.py index fe0c8cf6e..13e1bed99 100644 --- a/stackalytics/processor/dump.py +++ b/stackalytics/processor/dump.py @@ -26,6 +26,7 @@ from stackalytics.processor import config from stackalytics.processor import utils +CONF = cfg.CONF LOG = logging.getLogger(__name__) OPTS = [ @@ -61,7 +62,7 @@ def read_records_from_fd(fd): def store_bucket(memcached_inst, bucket): LOG.debug('Store bucket of records into memcached') res = memcached_inst.set_multi(bucket, - min_compress_len=cfg.CONF.min_compress_len) + min_compress_len=CONF.min_compress_len) if res: LOG.critical('Failed to set values in memcached: %s', res) raise Exception('memcached set_multi operation is failed') @@ -158,11 +159,11 @@ def _connect_to_memcached(uri): def main(): utils.init_config_and_logging(config.CONNECTION_OPTS + OPTS) - memcached_inst = _connect_to_memcached(cfg.CONF.runtime_storage_uri) + memcached_inst = _connect_to_memcached(CONF.runtime_storage_uri) - filename = cfg.CONF.file + filename = CONF.file - if cfg.CONF.restore: + if CONF.restore: if filename: fd = open(filename, 'r') else: diff --git a/stackalytics/processor/main.py b/stackalytics/processor/main.py index 4dcde72c4..6186df2f2 100644 --- a/stackalytics/processor/main.py +++ b/stackalytics/processor/main.py @@ -36,6 +36,7 @@ from stackalytics.processor import utils from stackalytics.processor import vcs from stackalytics.processor import zanata +CONF = cfg.CONF LOG = logging.getLogger(__name__) @@ -107,7 +108,7 @@ def _process_repo(repo, runtime_storage_inst, record_processor_inst, runtime_storage_inst.set_by_key( 'bug_modified_since-%s' % repo['module'], current_date) - vcs_inst = vcs.get_vcs(repo, cfg.CONF.sources_root) + vcs_inst = vcs.get_vcs(repo, CONF.sources_root) vcs_inst.fetch() branches = {repo.get('default_branch', 'master')} @@ -186,7 +187,7 @@ def _process_mail_list(uri, runtime_storage_inst, record_processor_inst): def _process_translation_stats(runtime_storage_inst, record_processor_inst): translation_iterator = zanata.log(runtime_storage_inst, - cfg.CONF.translation_team_uri) + CONF.translation_team_uri) translation_iterator_typed = _record_typer(translation_iterator, 'i18n') processed_translation_iterator = record_processor_inst.process( translation_iterator_typed) @@ -195,8 +196,8 @@ def _process_translation_stats(runtime_storage_inst, record_processor_inst): def _process_member_list(uri, runtime_storage_inst, record_processor_inst): member_iterator = mps.log(uri, runtime_storage_inst, - cfg.CONF.days_to_update_members, - cfg.CONF.members_look_ahead) + CONF.days_to_update_members, + CONF.members_look_ahead) member_iterator_typed = _record_typer(member_iterator, 'member') processed_member_iterator = record_processor_inst.process( member_iterator_typed) @@ -214,7 +215,7 @@ def _post_process_records(record_processor_inst, repos): LOG.debug('Build release index') release_index = {} for repo in repos: - vcs_inst = vcs.get_vcs(repo, cfg.CONF.sources_root) + vcs_inst = vcs.get_vcs(repo, CONF.sources_root) release_index.update(vcs_inst.fetch()) LOG.debug('Post-process all records') @@ -224,10 +225,10 @@ def _post_process_records(record_processor_inst, repos): def process(runtime_storage_inst, record_processor_inst): repos = utils.load_repos(runtime_storage_inst) - rcs_inst = rcs.get_rcs(cfg.CONF.review_uri) - rcs_inst.setup(key_filename=cfg.CONF.ssh_key_filename, - username=cfg.CONF.ssh_username, - gerrit_retry=cfg.CONF.gerrit_retry) + rcs_inst = rcs.get_rcs(CONF.review_uri) + rcs_inst.setup(key_filename=CONF.ssh_key_filename, + username=CONF.ssh_username, + gerrit_retry=CONF.gerrit_retry) for repo in repos: _process_repo(repo, runtime_storage_inst, record_processor_inst, @@ -310,19 +311,19 @@ def main(): config.PROCESSOR_OPTS) runtime_storage_inst = runtime_storage.get_runtime_storage( - cfg.CONF.runtime_storage_uri) + CONF.runtime_storage_uri) - if cfg.CONF.default_data_file: - default_data = utils.read_json_from_file(cfg.CONF.default_data_file) + if CONF.default_data_file: + default_data = utils.read_json_from_file(CONF.default_data_file) else: - default_data = utils.read_json_from_uri(cfg.CONF.default_data_uri) + default_data = utils.read_json_from_uri(CONF.default_data_uri) if not default_data: LOG.critical('Unable to load default data') return not 0 default_data_processor.process(runtime_storage_inst, default_data, - cfg.CONF.driverlog_data_uri) + CONF.driverlog_data_uri) process_project_list(runtime_storage_inst) @@ -333,7 +334,7 @@ def main(): process(runtime_storage_inst, record_processor_inst) - apply_corrections(cfg.CONF.corrections_uri, runtime_storage_inst) + apply_corrections(CONF.corrections_uri, runtime_storage_inst) # long operation should be the last update_members(runtime_storage_inst, record_processor_inst) diff --git a/stackalytics/processor/utils.py b/stackalytics/processor/utils.py index 6920eb179..d725afde5 100644 --- a/stackalytics/processor/utils.py +++ b/stackalytics/processor/utils.py @@ -31,21 +31,21 @@ import six import yaml +CONF = cfg.CONF LOG = logging.getLogger(__name__) def init_config_and_logging(opts): - conf = cfg.CONF - conf.register_cli_opts(opts) - conf.register_opts(opts) - logging.register_options(conf) + CONF.register_cli_opts(opts) + CONF.register_opts(opts) + logging.register_options(CONF) logging.set_defaults() - conf(project='stackalytics') + CONF(project='stackalytics') - logging.setup(conf, 'stackalytics') + logging.setup(CONF, 'stackalytics') LOG.info('Logging enabled') - conf.log_opt_values(LOG, logging.DEBUG) + CONF.log_opt_values(LOG, logging.DEBUG) def date_to_timestamp(d):