Ihar Hrachyshka 226de0b8f8 Fixed a comment on dirty logs list
We've recently changed default behaviour to forbid any dirty logs for
services outside the list instead of explicitly enabling log checks for
specific services. The comment was not updated to reflect the change

Change-Id: I5d5ba64d156a9b5318cfc2740b2a69d75de3c127
2014-10-31 13:56:36 +01:00

199 lines
6.1 KiB
Executable File

#!/usr/bin/env python
# Copyright 2013 Red Hat, Inc.
# All Rights Reserved.
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
# http://www.apache.org/licenses/LICENSE-2.0
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
import argparse
import gzip
import os
import re
import StringIO
import sys
import urllib2
import yaml
# DEVSTACK_GATE_GRENADE is either unset if grenade is not running
# or a string describing what type of grenade run to perform.
is_grenade = os.environ.get('DEVSTACK_GATE_GRENADE') is not None
dump_all_errors = True
# As logs are made clean, remove from this set
allowed_dirty = set([
def process_files(file_specs, url_specs, whitelists):
regexp = re.compile(r"^.* (ERROR|CRITICAL|TRACE) .*\[.*\-.*\]")
logs_with_errors = []
for (name, filename) in file_specs:
whitelist = whitelists.get(name, [])
with open(filename) as content:
if scan_content(name, content, regexp, whitelist):
for (name, url) in url_specs:
whitelist = whitelists.get(name, [])
req = urllib2.Request(url)
req.add_header('Accept-Encoding', 'gzip')
page = urllib2.urlopen(req)
buf = StringIO.StringIO(page.read())
f = gzip.GzipFile(fileobj=buf)
if scan_content(name, f.read().splitlines(), regexp, whitelist):
return logs_with_errors
def scan_content(name, content, regexp, whitelist):
had_errors = False
for line in content:
if not line.startswith("Stderr:") and regexp.match(line):
whitelisted = False
for w in whitelist:
pat = ".*%s.*%s.*" % (w['module'].replace('.', '\\.'),
if re.match(pat, line):
whitelisted = True
if not whitelisted or dump_all_errors:
if not whitelisted:
had_errors = True
return had_errors
def collect_url_logs(url):
page = urllib2.urlopen(url)
content = page.read()
logs = re.findall('(screen-[\w-]+\.txt\.gz)</a>', content)
return logs
def main(opts):
if opts.directory and opts.url or not (opts.directory or opts.url):
print("Must provide exactly one of -d or -u")
print("Checking logs...")
WHITELIST_FILE = os.path.join(
"etc", "whitelist.yaml")
file_matcher = re.compile(r".*screen-([\w-]+)\.log")
files = []
if opts.directory:
d = opts.directory
for f in os.listdir(d):
files.append(os.path.join(d, f))
files_to_process = []
for f in files:
m = file_matcher.match(f)
if m:
files_to_process.append((m.group(1), f))
url_matcher = re.compile(r".*screen-([\w-]+)\.txt\.gz")
urls = []
if opts.url:
for logfile in collect_url_logs(opts.url):
urls.append("%s/%s" % (opts.url, logfile))
urls_to_process = []
for u in urls:
m = url_matcher.match(u)
if m:
urls_to_process.append((m.group(1), u))
whitelists = {}
with open(WHITELIST_FILE) as stream:
loaded = yaml.safe_load(stream)
if loaded:
for (name, l) in loaded.iteritems():
for w in l:
assert 'module' in w, 'no module in %s' % name
assert 'message' in w, 'no message in %s' % name
whitelists = loaded
logs_with_errors = process_files(files_to_process, urls_to_process,
failed = False
if logs_with_errors:
log_files = set(logs_with_errors)
for log in log_files:
msg = '%s log file has errors' % log
if log not in allowed_dirty:
msg += ' and is not allowed to have them'
failed = True
print("\nPlease check the respective log files to see the errors")
if failed:
if is_grenade:
print("Currently not failing grenade runs with errors")
return 0
return 1
return 0
usage = """
Find non-white-listed log errors in log files from a devstack-gate run.
Log files will be searched for ERROR or CRITICAL messages. If any
error messages do not match any of the whitelist entries contained in
etc/whitelist.yaml, those messages will be printed to the console and
failure will be returned. A file directory containing logs or a url to the
log files of an OpenStack gate job can be provided.
The whitelist yaml looks like:
- module: "a.b.c"
message: "regexp"
- module: "a.b.c"
message: "regexp"
repeated for each log file with a whitelist.
parser = argparse.ArgumentParser(description=usage)
parser.add_argument('-d', '--directory',
help="Directory containing log files")
parser.add_argument('-u', '--url',
help="url containing logs from an OpenStack gate job")
if __name__ == "__main__":
except Exception as e:
print("Failure in script: %s" % e)
# Don't fail if there is a problem with the script.