diff options
Diffstat (limited to 'deprecated/generate-waterfall-reports.py')
-rwxr-xr-x | deprecated/generate-waterfall-reports.py | 853 |
1 files changed, 853 insertions, 0 deletions
diff --git a/deprecated/generate-waterfall-reports.py b/deprecated/generate-waterfall-reports.py new file mode 100755 index 00000000..a67cd6ca --- /dev/null +++ b/deprecated/generate-waterfall-reports.py @@ -0,0 +1,853 @@ +#!/usr/bin/env python2 +"""Generate summary report for ChromeOS toolchain waterfalls.""" + +# Desired future features (to be added): +# - arguments to allow generating only the main waterfall report, +# or only the rotating builder reports, or only the failures +# report; or the waterfall reports without the failures report. +# - Better way of figuring out which dates/builds to generate +# reports for: probably an argument specifying a date or a date +# range, then use something like the new buildbot utils to +# query the build logs to find the right build numbers for the +# builders for the specified dates. +# - Store/get the json/data files in mobiletc-prebuild's x20 area. +# - Update data in json file to reflect, for each testsuite, which +# tests are not expected to run on which boards; update this +# script to use that data appropriately. +# - Make sure user's prodaccess is up-to-date before trying to use +# this script. +# - Add some nice formatting/highlighting to reports. + +from __future__ import print_function + +import argparse +import getpass +import json +import os +import re +import shutil +import sys +import time + +from cros_utils import command_executer + +# All the test suites whose data we might want for the reports. +TESTS = (('bvt-inline', 'HWTest'), ('bvt-cq', 'HWTest'), ('security', 'HWTest'), + ('kernel_daily_regression', 'HWTest'), ('kernel_daily_benchmarks', + 'HWTest'),) + +# The main waterfall builders, IN THE ORDER IN WHICH WE WANT THEM +# LISTED IN THE REPORT. +WATERFALL_BUILDERS = [ + 'amd64-llvm-next-toolchain', + 'arm-llvm-next-toolchain', + 'arm64-llvm-next-toolchain', +] + +DATA_DIR = '/google/data/rw/users/mo/mobiletc-prebuild/waterfall-report-data/' +ARCHIVE_DIR = '/google/data/rw/users/mo/mobiletc-prebuild/waterfall-reports/' +DOWNLOAD_DIR = '/tmp/waterfall-logs' +MAX_SAVE_RECORDS = 7 +BUILD_DATA_FILE = '%s/build-data.txt' % DATA_DIR +GCC_ROTATING_BUILDER = 'gcc_toolchain' +LLVM_ROTATING_BUILDER = 'llvm_next_toolchain' +ROTATING_BUILDERS = [GCC_ROTATING_BUILDER, LLVM_ROTATING_BUILDER] + +# For int-to-string date conversion. Note, the index of the month in this +# list needs to correspond to the month's integer value. i.e. 'Sep' must +# be as MONTHS[9]. +MONTHS = [ + '', 'Jan', 'Feb', 'Mar', 'Apr', 'May', 'Jun', 'Jul', 'Aug', 'Sep', 'Oct', + 'Nov', 'Dec' +] + + +def format_date(int_date): + """Convert an integer date to a string date. YYYYMMDD -> YYYY-MMM-DD""" + + if int_date == 0: + return 'today' + + tmp_date = int_date + day = tmp_date % 100 + tmp_date = tmp_date / 100 + month = tmp_date % 100 + year = tmp_date / 100 + + month_str = MONTHS[month] + date_str = '%d-%s-%d' % (year, month_str, day) + return date_str + + +def EmailReport(report_file, report_type, date, email_to): + subject = '%s Waterfall Summary report, %s' % (report_type, date) + sendgmr_path = '/google/data/ro/projects/gws-sre/sendgmr' + command = ('%s --to=%s --subject="%s" --body_file=%s' % + (sendgmr_path, email_to, subject, report_file)) + command_executer.GetCommandExecuter().RunCommand(command) + + +def PruneOldFailures(failure_dict, int_date): + earliest_date = int_date - MAX_SAVE_RECORDS + for suite in failure_dict: + suite_dict = failure_dict[suite] + test_keys_to_remove = [] + for test in suite_dict: + test_dict = suite_dict[test] + msg_keys_to_remove = [] + for msg in test_dict: + fails = test_dict[msg] + i = 0 + while i < len(fails) and fails[i][0] <= earliest_date: + i += 1 + new_fails = fails[i:] + test_dict[msg] = new_fails + if len(new_fails) == 0: + msg_keys_to_remove.append(msg) + + for k in msg_keys_to_remove: + del test_dict[k] + + suite_dict[test] = test_dict + if len(test_dict) == 0: + test_keys_to_remove.append(test) + + for k in test_keys_to_remove: + del suite_dict[k] + + failure_dict[suite] = suite_dict + + +def GetBuildID(build_bot, date): + """Get the build id for a build_bot at a given date.""" + day = '{day:02d}'.format(day=date % 100) + mon = MONTHS[date / 100 % 100] + date_string = mon + ' ' + day + if build_bot in WATERFALL_BUILDERS: + url = 'https://uberchromegw.corp.google.com/i/chromeos/' + \ + 'builders/%s?numbuilds=200' % build_bot + if build_bot in ROTATING_BUILDERS: + url = 'https://uberchromegw.corp.google.com/i/chromiumos.tryserver/' + \ + 'builders/%s?numbuilds=200' % build_bot + command = 'sso_client %s' % url + retval = 1 + retry_time = 3 + while retval and retry_time: + retval, output, _ = \ + command_executer.GetCommandExecuter().RunCommandWOutput(command, \ + print_to_console=False) + retry_time -= 1 + + if retval: + return [] + + out = output.split('\n') + line_num = 0 + build_id = [] + # Parse the output like this + # <td>Dec 14 10:55</td> + # <td class="revision">??</td> + # <td failure</td><td><a href="../builders/gcc_toolchain/builds/109">#109</a> + while line_num < len(out): + if date_string in out[line_num]: + if line_num + 2 < len(out): + build_num_line = out[line_num + 2] + raw_num = re.findall(r'builds/\d+', build_num_line) + # raw_num is ['builds/109'] in the example. + if raw_num: + build_id.append(int(raw_num[0].split('/')[1])) + line_num += 1 + return build_id + + +def GenerateFailuresReport(fail_dict, date): + filename = 'waterfall_report.failures.%s.txt' % date + date_string = format_date(date) + with open(filename, 'w') as out_file: + # Write failure report section. + out_file.write('\n\nSummary of Test Failures as of %s\n\n' % date_string) + + # We want to sort the errors and output them in order of the ones that occur + # most often. So we have to collect the data about all of them, then sort + # it. + error_groups = [] + for suite in fail_dict: + suite_dict = fail_dict[suite] + if suite_dict: + for test in suite_dict: + test_dict = suite_dict[test] + for err_msg in test_dict: + err_list = test_dict[err_msg] + sorted_list = sorted(err_list, key=lambda x: x[0], reverse=True) + err_group = [len(sorted_list), suite, test, err_msg, sorted_list] + error_groups.append(err_group) + + # Sort the errors by the number of errors of each type. Then output them in + # order. + sorted_errors = sorted(error_groups, key=lambda x: x[0], reverse=True) + for i in range(0, len(sorted_errors)): + err_group = sorted_errors[i] + suite = err_group[1] + test = err_group[2] + err_msg = err_group[3] + err_list = err_group[4] + out_file.write('Suite: %s\n' % suite) + out_file.write(' %s (%d failures)\n' % (test, len(err_list))) + out_file.write(' (%s)\n' % err_msg) + for i in range(0, len(err_list)): + err = err_list[i] + out_file.write(' %s, %s, %s\n' % (format_date(err[0]), err[1], + err[2])) + out_file.write('\n') + + print('Report generated in %s.' % filename) + return filename + + +def GenerateWaterfallReport(report_dict, fail_dict, waterfall_type, date, + omit_failures): + """Write out the actual formatted report.""" + + filename = 'waterfall_report.%s_waterfall.%s.txt' % (waterfall_type, date) + + date_string = '' + date_list = report_dict['date'] + num_dates = len(date_list) + i = 0 + for d in date_list: + date_string += d + if i < num_dates - 1: + date_string += ', ' + i += 1 + + if waterfall_type == 'main': + report_list = WATERFALL_BUILDERS + else: + report_list = report_dict.keys() + + with open(filename, 'w') as out_file: + # Write Report Header + out_file.write('\nStatus of %s Waterfall Builds from %s\n\n' % + (waterfall_type, date_string)) + out_file.write(' ' + ' kernel kernel\n') + out_file.write(' Build bvt- bvt-cq ' + ' security daily daily\n') + out_file.write(' status inline ' + ' regression benchmarks\n') + out_file.write(' [P/ F/ DR]* [P/ F /DR]* ' + '[P/ F/ DR]* [P/ F/ DR]* [P/ F/ DR]*\n\n') + + # Write daily waterfall status section. + for i in range(0, len(report_list)): + builder = report_list[i] + if builder == 'date': + continue + + if builder not in report_dict: + out_file.write('Unable to find information for %s.\n\n' % builder) + continue + + build_dict = report_dict[builder] + status = build_dict.get('build_status', 'bad') + inline = build_dict.get('bvt-inline', '[??/ ?? /??]') + cq = build_dict.get('bvt-cq', '[??/ ?? /??]') + inline_color = build_dict.get('bvt-inline-color', '') + cq_color = build_dict.get('bvt-cq-color', '') + if 'x86' not in builder: + security = build_dict.get('security', '[??/ ?? /??]') + security_color = build_dict.get('security-color', '') + if 'gcc' in builder: + regression = build_dict.get('kernel_daily_regression', '[??/ ?? /??]') + bench = build_dict.get('kernel_daily_benchmarks', '[??/ ?? /??]') + regression_color = build_dict.get('kernel_daily_regression-color', '') + bench_color = build_dict.get('kernel_daily_benchmarks-color', '') + out_file.write(' %6s %6s' + ' %6s %6s %6s\n' % + (inline_color, cq_color, security_color, + regression_color, bench_color)) + out_file.write('%25s %3s %s %s %s %s %s\n' % + (builder, status, inline, cq, security, regression, + bench)) + else: + out_file.write(' %6s %6s' + ' %6s\n' % (inline_color, cq_color, + security_color)) + out_file.write('%25s %3s %s %s %s\n' % (builder, status, inline, cq, + security)) + else: + out_file.write(' %6s %6s\n' % + (inline_color, cq_color)) + out_file.write('%25s %3s %s %s\n' % (builder, status, inline, cq)) + if 'build_link' in build_dict: + out_file.write('%s\n\n' % build_dict['build_link']) + + out_file.write('\n\n*P = Number of tests in suite that Passed; F = ' + 'Number of tests in suite that Failed; DR = Number of tests' + ' in suite that Didn\'t Run.\n') + + if omit_failures: + print('Report generated in %s.' % filename) + return filename + + # Write failure report section. + out_file.write('\n\nSummary of Test Failures as of %s\n\n' % date_string) + + # We want to sort the errors and output them in order of the ones that occur + # most often. So we have to collect the data about all of them, then sort + # it. + error_groups = [] + for suite in fail_dict: + suite_dict = fail_dict[suite] + if suite_dict: + for test in suite_dict: + test_dict = suite_dict[test] + for err_msg in test_dict: + err_list = test_dict[err_msg] + sorted_list = sorted(err_list, key=lambda x: x[0], reverse=True) + err_group = [len(sorted_list), suite, test, err_msg, sorted_list] + error_groups.append(err_group) + + # Sort the errors by the number of errors of each type. Then output them in + # order. + sorted_errors = sorted(error_groups, key=lambda x: x[0], reverse=True) + for i in range(0, len(sorted_errors)): + err_group = sorted_errors[i] + suite = err_group[1] + test = err_group[2] + err_msg = err_group[3] + err_list = err_group[4] + out_file.write('Suite: %s\n' % suite) + out_file.write(' %s (%d failures)\n' % (test, len(err_list))) + out_file.write(' (%s)\n' % err_msg) + for i in range(0, len(err_list)): + err = err_list[i] + out_file.write(' %s, %s, %s\n' % (format_date(err[0]), err[1], + err[2])) + out_file.write('\n') + + print('Report generated in %s.' % filename) + return filename + + +def UpdateReport(report_dict, builder, test, report_date, build_link, + test_summary, board, color): + """Update the data in our report dictionary with current test's data.""" + + if 'date' not in report_dict: + report_dict['date'] = [report_date] + elif report_date not in report_dict['date']: + # It is possible that some of the builders started/finished on different + # days, so we allow for multiple dates in the reports. + report_dict['date'].append(report_date) + + build_key = '' + if builder == GCC_ROTATING_BUILDER: + build_key = '%s-gcc-toolchain' % board + elif builder == LLVM_ROTATING_BUILDER: + build_key = '%s-llvm-next-toolchain' % board + else: + build_key = builder + + if build_key not in report_dict.keys(): + build_dict = dict() + else: + build_dict = report_dict[build_key] + + if 'build_link' not in build_dict: + build_dict['build_link'] = build_link + + if 'date' not in build_dict: + build_dict['date'] = report_date + + if 'board' in build_dict and build_dict['board'] != board: + raise RuntimeError( + 'Error: Two different boards (%s,%s) in one build (%s)!' % + (board, build_dict['board'], build_link)) + build_dict['board'] = board + + color_key = '%s-color' % test + build_dict[color_key] = color + + # Check to see if we already have a build status for this build_key + status = '' + if 'build_status' in build_dict.keys(): + # Use current build_status, unless current test failed (see below). + status = build_dict['build_status'] + + if not test_summary: + # Current test data was not available, so something was bad with build. + build_dict['build_status'] = 'bad' + build_dict[test] = '[ no data ]' + else: + build_dict[test] = test_summary + if not status: + # Current test ok; no other data, so assume build was ok. + build_dict['build_status'] = 'ok' + + report_dict[build_key] = build_dict + + +def UpdateBuilds(builds): + """Update the data in our build-data.txt file.""" + + # The build data file records the last build number for which we + # generated a report. When we generate the next report, we read + # this data and increment it to get the new data; when we finish + # generating the reports, we write the updated values into this file. + # NOTE: One side effect of doing this at the end: If the script + # fails in the middle of generating a report, this data does not get + # updated. + with open(BUILD_DATA_FILE, 'w') as fp: + gcc_max = 0 + llvm_max = 0 + for b in builds: + if b[0] == GCC_ROTATING_BUILDER: + gcc_max = max(gcc_max, b[1]) + elif b[0] == LLVM_ROTATING_BUILDER: + llvm_max = max(llvm_max, b[1]) + else: + fp.write('%s,%d\n' % (b[0], b[1])) + if gcc_max > 0: + fp.write('%s,%d\n' % (GCC_ROTATING_BUILDER, gcc_max)) + if llvm_max > 0: + fp.write('%s,%d\n' % (LLVM_ROTATING_BUILDER, llvm_max)) + + +def GetBuilds(date=0): + """Get build id from builds.""" + + # If date is set, get the build id from waterfall. + builds = [] + + if date: + for builder in WATERFALL_BUILDERS + ROTATING_BUILDERS: + build_ids = GetBuildID(builder, date) + for build_id in build_ids: + builds.append((builder, build_id)) + return builds + + # If date is not set, we try to get the most recent builds. + # Read the values of the last builds used to generate a report, and + # increment them appropriately, to get values for generating the + # current report. (See comments in UpdateBuilds). + with open(BUILD_DATA_FILE, 'r') as fp: + lines = fp.readlines() + + for l in lines: + l = l.rstrip() + words = l.split(',') + builder = words[0] + build = int(words[1]) + builds.append((builder, build + 1)) + # NOTE: We are assuming here that there are always 2 daily builds in + # each of the rotating builders. I am not convinced this is a valid + # assumption. + if builder in ROTATING_BUILDERS: + builds.append((builder, build + 2)) + + return builds + + +def RecordFailures(failure_dict, platform, suite, builder, int_date, log_file, + build_num, failed): + """Read and update the stored data about test failures.""" + + # Get the dictionary for this particular test suite from the failures + # dictionary. + suite_dict = failure_dict[suite] + + # Read in the entire log file for this test/build. + with open(log_file, 'r') as in_file: + lines = in_file.readlines() + + # Update the entries in the failure dictionary for each test within this suite + # that failed. + for test in failed: + # Check to see if there is already an entry in the suite dictionary for this + # test; if so use that, otherwise create a new entry. + if test in suite_dict: + test_dict = suite_dict[test] + else: + test_dict = dict() + # Parse the lines from the log file, looking for lines that indicate this + # test failed. + msg = '' + for l in lines: + words = l.split() + if len(words) < 3: + continue + if ((words[0] == test and words[1] == 'ERROR:') or + (words[0] == 'provision' and words[1] == 'FAIL:')): + words = words[2:] + # Get the error message for the failure. + msg = ' '.join(words) + if not msg: + msg = 'Unknown_Error' + + # Look for an existing entry for this error message in the test dictionary. + # If found use that, otherwise create a new entry for this error message. + if msg in test_dict: + error_list = test_dict[msg] + else: + error_list = list() + # Create an entry for this new failure + new_item = [int_date, platform, builder, build_num] + # Add this failure to the error list if it's not already there. + if new_item not in error_list: + error_list.append([int_date, platform, builder, build_num]) + # Sort the error list by date. + error_list.sort(key=lambda x: x[0]) + # Calculate the earliest date to save; delete records for older failures. + earliest_date = int_date - MAX_SAVE_RECORDS + i = 0 + while i < len(error_list) and error_list[i][0] <= earliest_date: + i += 1 + if i > 0: + error_list = error_list[i:] + # Save the error list in the test's dictionary, keyed on error_msg. + test_dict[msg] = error_list + + # Save the updated test dictionary in the test_suite dictionary. + suite_dict[test] = test_dict + + # Save the updated test_suite dictionary in the failure dictionary. + failure_dict[suite] = suite_dict + + +def ParseLogFile(log_file, test_data_dict, failure_dict, test, builder, + build_num, build_link): + """Parse the log file from the given builder, build_num and test. + + Also adds the results for this test to our test results dictionary, + and calls RecordFailures, to update our test failure data. + """ + + print('Parsing file %s' % log_file) + lines = [] + with open(log_file, 'r') as infile: + lines = infile.readlines() + + passed = {} + failed = {} + not_run = {} + date = '' + status = '' + board = '' + num_provision_errors = 0 + build_ok = True + afe_line = '' + + for line in lines: + if line.rstrip() == '<title>404 Not Found</title>': + print('Warning: File for %s (build number %d), %s was not found.' % + (builder, build_num, test)) + build_ok = False + break + if '[ PASSED ]' in line: + test_name = line.split()[0] + if test_name != 'Suite': + passed[test_name] = True + elif '[ FAILED ]' in line: + test_name = line.split()[0] + if test_name == 'provision': + num_provision_errors += 1 + not_run[test_name] = True + elif test_name != 'Suite': + failed[test_name] = True + elif line.startswith('started: '): + date = line.rstrip() + date = date[9:] + date_obj = time.strptime(date, '%a %b %d %H:%M:%S %Y') + int_date = ( + date_obj.tm_year * 10000 + date_obj.tm_mon * 100 + date_obj.tm_mday) + date = time.strftime('%a %b %d %Y', date_obj) + elif not status and line.startswith('status: '): + status = line.rstrip() + words = status.split(':') + status = words[-1] + elif line.find('Suite passed with a warning') != -1: + status = 'WARNING' + elif line.startswith('@@@STEP_LINK@Link to suite@'): + afe_line = line.rstrip() + words = afe_line.split('@') + for w in words: + if w.startswith('http'): + afe_line = w + afe_line = afe_line.replace('&', '&') + elif 'INFO: RunCommand:' in line: + words = line.split() + for i in range(0, len(words) - 1): + if words[i] == '--board': + board = words[i + 1] + + test_dict = test_data_dict[test] + test_list = test_dict['tests'] + + if build_ok: + for t in test_list: + if not t in passed and not t in failed: + not_run[t] = True + + total_pass = len(passed) + total_fail = len(failed) + total_notrun = len(not_run) + + else: + total_pass = 0 + total_fail = 0 + total_notrun = 0 + status = 'Not found.' + if not build_ok: + return [], date, board, 0, ' ' + + build_dict = dict() + build_dict['id'] = build_num + build_dict['builder'] = builder + build_dict['date'] = date + build_dict['build_link'] = build_link + build_dict['total_pass'] = total_pass + build_dict['total_fail'] = total_fail + build_dict['total_not_run'] = total_notrun + build_dict['afe_job_link'] = afe_line + build_dict['provision_errors'] = num_provision_errors + if status.strip() == 'SUCCESS': + build_dict['color'] = 'green ' + elif status.strip() == 'FAILURE': + build_dict['color'] = ' red ' + elif status.strip() == 'WARNING': + build_dict['color'] = 'orange' + else: + build_dict['color'] = ' ' + + # Use YYYYMMDD (integer) as the build record key + if build_ok: + if board in test_dict: + board_dict = test_dict[board] + else: + board_dict = dict() + board_dict[int_date] = build_dict + + # Only keep the last 5 records (based on date) + keys_list = board_dict.keys() + if len(keys_list) > MAX_SAVE_RECORDS: + min_key = min(keys_list) + del board_dict[min_key] + + # Make sure changes get back into the main dictionary + test_dict[board] = board_dict + test_data_dict[test] = test_dict + + if len(failed) > 0: + RecordFailures(failure_dict, board, test, builder, int_date, log_file, + build_num, failed) + + summary_result = '[%2d/ %2d/ %2d]' % (total_pass, total_fail, total_notrun) + + return summary_result, date, board, int_date, build_dict['color'] + + +def DownloadLogFile(builder, buildnum, test, test_family): + + ce = command_executer.GetCommandExecuter() + os.system('mkdir -p %s/%s/%s' % (DOWNLOAD_DIR, builder, test)) + if builder in ROTATING_BUILDERS: + source = ('https://uberchromegw.corp.google.com/i/chromiumos.tryserver' + '/builders/%s/builds/%d/steps/%s%%20%%5B%s%%5D/logs/stdio' % + (builder, buildnum, test_family, test)) + build_link = ('https://uberchromegw.corp.google.com/i/chromiumos.tryserver' + '/builders/%s/builds/%d' % (builder, buildnum)) + else: + source = ('https://uberchromegw.corp.google.com/i/chromeos/builders/%s/' + 'builds/%d/steps/%s%%20%%5B%s%%5D/logs/stdio' % + (builder, buildnum, test_family, test)) + build_link = ('https://uberchromegw.corp.google.com/i/chromeos/builders/%s' + '/builds/%d' % (builder, buildnum)) + + target = '%s/%s/%s/%d' % (DOWNLOAD_DIR, builder, test, buildnum) + if not os.path.isfile(target) or os.path.getsize(target) == 0: + cmd = 'sso_client %s > %s' % (source, target) + status = ce.RunCommand(cmd) + if status != 0: + return '', '' + + return target, build_link + + +# Check for prodaccess. +def CheckProdAccess(): + status, output, _ = command_executer.GetCommandExecuter().RunCommandWOutput( + 'prodcertstatus') + if status != 0: + return False + # Verify that status is not expired + if 'expires' in output: + return True + return False + + +def ValidOptions(parser, options): + too_many_options = False + if options.main: + if options.rotating or options.failures_report: + too_many_options = True + elif options.rotating and options.failures_report: + too_many_options = True + + if too_many_options: + parser.error('Can only specify one of --main, --rotating or' + ' --failures_report.') + + conflicting_failure_options = False + if options.failures_report and options.omit_failures: + conflicting_failure_options = True + parser.error('Cannot specify both --failures_report and --omit_failures.') + + email_ok = True + if options.email and options.email.find('@') == -1: + email_ok = False + parser.error('"%s" is not a valid email address; it must contain "@..."' % + options.email) + + return not too_many_options and not conflicting_failure_options and email_ok + + +def Main(argv): + """Main function for this script.""" + parser = argparse.ArgumentParser() + parser.add_argument( + '--main', + dest='main', + default=False, + action='store_true', + help='Generate report only for main waterfall ' + 'builders.') + parser.add_argument( + '--rotating', + dest='rotating', + default=False, + action='store_true', + help='Generate report only for rotating builders.') + parser.add_argument( + '--failures_report', + dest='failures_report', + default=False, + action='store_true', + help='Only generate the failures section of the report.') + parser.add_argument( + '--omit_failures', + dest='omit_failures', + default=False, + action='store_true', + help='Do not generate the failures section of the report.') + parser.add_argument( + '--no_update', + dest='no_update', + default=False, + action='store_true', + help='Run reports, but do not update the data files.') + parser.add_argument( + '--date', + dest='date', + default=0, + type=int, + help='The date YYYYMMDD of waterfall report.') + parser.add_argument( + '--email', + dest='email', + default='', + help='Email address to use for sending the report.') + + options = parser.parse_args(argv) + + if not ValidOptions(parser, options): + return 1 + + main_only = options.main + rotating_only = options.rotating + failures_report = options.failures_report + omit_failures = options.omit_failures + date = options.date + + test_data_dict = dict() + failure_dict = dict() + + prod_access = CheckProdAccess() + if not prod_access: + print('ERROR: Please run prodaccess first.') + return + + with open('%s/waterfall-test-data.json' % DATA_DIR, 'r') as input_file: + test_data_dict = json.load(input_file) + + with open('%s/test-failure-data.json' % DATA_DIR, 'r') as fp: + failure_dict = json.load(fp) + + builds = GetBuilds(date) + + waterfall_report_dict = dict() + rotating_report_dict = dict() + int_date = 0 + for test_desc in TESTS: + test, test_family = test_desc + for build in builds: + (builder, buildnum) = build + if test.startswith('kernel') and 'llvm' in builder: + continue + if 'x86' in builder and not test.startswith('bvt'): + continue + target, build_link = DownloadLogFile(builder, buildnum, test, test_family) + + if os.path.exists(target): + test_summary, report_date, board, tmp_date, color = ParseLogFile( + target, test_data_dict, failure_dict, test, builder, buildnum, + build_link) + if not test_summary: + continue + + if tmp_date != 0: + int_date = tmp_date + + if builder in ROTATING_BUILDERS: + UpdateReport(rotating_report_dict, builder, test, report_date, + build_link, test_summary, board, color) + else: + UpdateReport(waterfall_report_dict, builder, test, report_date, + build_link, test_summary, board, color) + + PruneOldFailures(failure_dict, int_date) + + if options.email: + email_to = options.email + else: + email_to = getpass.getuser() + + if waterfall_report_dict and not rotating_only and not failures_report: + main_report = GenerateWaterfallReport(waterfall_report_dict, failure_dict, + 'main', int_date, omit_failures) + EmailReport(main_report, 'Main', format_date(int_date), email_to) + shutil.copy(main_report, ARCHIVE_DIR) + if rotating_report_dict and not main_only and not failures_report: + rotating_report = GenerateWaterfallReport( + rotating_report_dict, failure_dict, 'rotating', int_date, omit_failures) + EmailReport(rotating_report, 'Rotating', format_date(int_date), email_to) + shutil.copy(rotating_report, ARCHIVE_DIR) + + if failures_report: + failures_report = GenerateFailuresReport(failure_dict, int_date) + EmailReport(failures_report, 'Failures', format_date(int_date), email_to) + shutil.copy(failures_report, ARCHIVE_DIR) + + if not options.no_update: + with open('%s/waterfall-test-data.json' % DATA_DIR, 'w') as out_file: + json.dump(test_data_dict, out_file, indent=2) + + with open('%s/test-failure-data.json' % DATA_DIR, 'w') as out_file: + json.dump(failure_dict, out_file, indent=2) + + UpdateBuilds(builds) + + +if __name__ == '__main__': + Main(sys.argv[1:]) + sys.exit(0) |