tempest/tools/check_logs.py
Masayuki Igawa 134d9f748b
Introduce flake8-import-order
This commit introduces flake8-import-order to check import orders
automatically. We shouldn't do it manually.

Depends-On: Ia2140f5566ce169b076ffa5c9ebfbdbfc41e5bed
Change-Id: Iff75ebec42b295870edd1c04001adfb17460a682
2017-02-13 17:16:18 +09:00

198 lines
6.1 KiB
Python
Executable File

#!/usr/bin/env python
# Copyright 2013 Red Hat, Inc.
# All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import argparse
import gzip
import os
import re
import sys
import six
import six.moves.urllib.request as urlreq
import yaml
# DEVSTACK_GATE_GRENADE is either unset if grenade is not running
# or a string describing what type of grenade run to perform.
is_grenade = os.environ.get('DEVSTACK_GATE_GRENADE') is not None
dump_all_errors = True
# As logs are made clean, remove from this set
allowed_dirty = set([
'c-api',
'ceilometer-acentral',
'ceilometer-acompute',
'ceilometer-alarm-evaluator',
'ceilometer-anotification',
'ceilometer-api',
'ceilometer-collector',
'c-vol',
'g-api',
'h-api',
'h-eng',
'ir-cond',
'n-api',
'n-cpu',
'n-net',
'q-agt',
'q-dhcp',
'q-lbaas',
'q-meta',
'q-metering',
'q-svc',
's-proxy'])
def process_files(file_specs, url_specs, whitelists):
regexp = re.compile(r"^.* (ERROR|CRITICAL|TRACE) .*\[.*\-.*\]")
logs_with_errors = []
for (name, filename) in file_specs:
whitelist = whitelists.get(name, [])
with open(filename) as content:
if scan_content(name, content, regexp, whitelist):
logs_with_errors.append(name)
for (name, url) in url_specs:
whitelist = whitelists.get(name, [])
req = urlreq.Request(url)
req.add_header('Accept-Encoding', 'gzip')
page = urlreq.urlopen(req)
buf = six.StringIO(page.read())
f = gzip.GzipFile(fileobj=buf)
if scan_content(name, f.read().splitlines(), regexp, whitelist):
logs_with_errors.append(name)
return logs_with_errors
def scan_content(name, content, regexp, whitelist):
had_errors = False
for line in content:
if not line.startswith("Stderr:") and regexp.match(line):
whitelisted = False
for w in whitelist:
pat = ".*%s.*%s.*" % (w['module'].replace('.', '\\.'),
w['message'])
if re.match(pat, line):
whitelisted = True
break
if not whitelisted or dump_all_errors:
if not whitelisted:
had_errors = True
return had_errors
def collect_url_logs(url):
page = urlreq.urlopen(url)
content = page.read()
logs = re.findall('(screen-[\w-]+\.txt\.gz)</a>', content)
return logs
def main(opts):
if opts.directory and opts.url or not (opts.directory or opts.url):
print("Must provide exactly one of -d or -u")
return 1
print("Checking logs...")
WHITELIST_FILE = os.path.join(
os.path.abspath(os.path.dirname(os.path.dirname(__file__))),
"etc", "whitelist.yaml")
file_matcher = re.compile(r".*screen-([\w-]+)\.log")
files = []
if opts.directory:
d = opts.directory
for f in os.listdir(d):
files.append(os.path.join(d, f))
files_to_process = []
for f in files:
m = file_matcher.match(f)
if m:
files_to_process.append((m.group(1), f))
url_matcher = re.compile(r".*screen-([\w-]+)\.txt\.gz")
urls = []
if opts.url:
for logfile in collect_url_logs(opts.url):
urls.append("%s/%s" % (opts.url, logfile))
urls_to_process = []
for u in urls:
m = url_matcher.match(u)
if m:
urls_to_process.append((m.group(1), u))
whitelists = {}
with open(WHITELIST_FILE) as stream:
loaded = yaml.safe_load(stream)
if loaded:
for (name, l) in loaded.iteritems():
for w in l:
assert 'module' in w, 'no module in %s' % name
assert 'message' in w, 'no message in %s' % name
whitelists = loaded
logs_with_errors = process_files(files_to_process, urls_to_process,
whitelists)
failed = False
if logs_with_errors:
log_files = set(logs_with_errors)
for log in log_files:
msg = '%s log file has errors' % log
if log not in allowed_dirty:
msg += ' and is not allowed to have them'
failed = True
print(msg)
print("\nPlease check the respective log files to see the errors")
if failed:
if is_grenade:
print("Currently not failing grenade runs with errors")
return 0
return 1
print("ok")
return 0
usage = """
Find non-white-listed log errors in log files from a devstack-gate run.
Log files will be searched for ERROR or CRITICAL messages. If any
error messages do not match any of the whitelist entries contained in
etc/whitelist.yaml, those messages will be printed to the console and
failure will be returned. A file directory containing logs or a url to the
log files of an OpenStack gate job can be provided.
The whitelist yaml looks like:
log-name:
- module: "a.b.c"
message: "regexp"
- module: "a.b.c"
message: "regexp"
repeated for each log file with a whitelist.
"""
parser = argparse.ArgumentParser(description=usage)
parser.add_argument('-d', '--directory',
help="Directory containing log files")
parser.add_argument('-u', '--url',
help="url containing logs from an OpenStack gate job")
if __name__ == "__main__":
try:
sys.exit(main(parser.parse_args()))
except Exception as e:
print("Failure in script: %s" % e)
# Don't fail if there is a problem with the script.
sys.exit(0)