aboutsummaryrefslogtreecommitdiffhomepage
diff options
context:
space:
mode:
authorGravatar bensong@google.com <bensong@google.com@2bbb7eff-a529-9590-31e7-b0007b416f81>2013-10-15 00:08:14 +0000
committerGravatar bensong@google.com <bensong@google.com@2bbb7eff-a529-9590-31e7-b0007b416f81>2013-10-15 00:08:14 +0000
commit0d9e6defbc403dfb614b91f782d988938aa9d7e1 (patch)
tree30a3a149fa32ee19f005f2a8fe6d63b9fadb5c7b
parentadacc7067ad617cdc7bbef39192ca80f4b4d27f9 (diff)
Changes bench to output with git hash filename; cleans up bench_graph_svg to only handle bench alerts; renames it to check_bench_regressions.
Review URL: https://codereview.chromium.org/26592010 git-svn-id: http://skia.googlecode.com/svn/trunk@11765 2bbb7eff-a529-9590-31e7-b0007b416f81
-rw-r--r--bench/bench_graph_svg.py1062
-rw-r--r--bench/check_bench_regressions.py218
2 files changed, 218 insertions, 1062 deletions
diff --git a/bench/bench_graph_svg.py b/bench/bench_graph_svg.py
deleted file mode 100644
index 02ecbe17c9..0000000000
--- a/bench/bench_graph_svg.py
+++ /dev/null
@@ -1,1062 +0,0 @@
-'''
-Created on May 16, 2011
-
-@author: bungeman
-'''
-import bench_util
-import getopt
-import httplib
-import itertools
-import json
-import os
-import re
-import sys
-import urllib
-import urllib2
-import xml.sax.saxutils
-
-# We throw out any measurement outside this range, and log a warning.
-MIN_REASONABLE_TIME = 0
-MAX_REASONABLE_TIME = 99999
-
-# Constants for prefixes in output title used in buildbot.
-TITLE_PREAMBLE = 'Bench_Performance_for_'
-TITLE_PREAMBLE_LENGTH = len(TITLE_PREAMBLE)
-
-def usage():
- """Prints simple usage information."""
-
- print '-a <url> the url to use for adding bench values to app engine app.'
- print ' Example: "https://skiadash.appspot.com/add_point".'
- print ' If not set, will skip this step.'
- print '-b <bench> the bench to show.'
- print '-c <config> the config to show (GPU, 8888, 565, etc).'
- print '-d <dir> a directory containing bench_r<revision>_<scalar> files.'
- print '-e <file> file containing expected bench builder values/ranges.'
- print ' Will raise exception if actual bench values are out of range.'
- print ' See bench_expectations_<builder>.txt for data format / examples.'
- print '-f <revision>[:<revision>] the revisions to use for fitting.'
- print ' Negative <revision> is taken as offset from most recent revision.'
- print '-i <time> the time to ignore (w, c, g, etc).'
- print ' The flag is ignored when -t is set; otherwise we plot all the'
- print ' times except the one specified here.'
- print '-l <title> title to use for the output graph'
- print '-m <representation> representation of bench value.'
- print ' See _ListAlgorithm class in bench_util.py.'
- print '-o <path> path to which to write output.'
- print '-r <revision>[:<revision>] the revisions to show.'
- print ' Negative <revision> is taken as offset from most recent revision.'
- print '-s <setting>[=<value>] a setting to show (alpha, scalar, etc).'
- print '-t <time> the time to show (w, c, g, etc).'
- print '-x <int> the desired width of the svg.'
- print '-y <int> the desired height of the svg.'
- print '--default-setting <setting>[=<value>] setting for those without.'
-
-
-class Label:
- """The information in a label.
-
- (str, str, str, str, {str:str})"""
- def __init__(self, bench, config, time_type, settings):
- self.bench = bench
- self.config = config
- self.time_type = time_type
- self.settings = settings
-
- def __repr__(self):
- return "Label(%s, %s, %s, %s)" % (
- str(self.bench),
- str(self.config),
- str(self.time_type),
- str(self.settings),
- )
-
- def __str__(self):
- return "%s_%s_%s_%s" % (
- str(self.bench),
- str(self.config),
- str(self.time_type),
- str(self.settings),
- )
-
- def __eq__(self, other):
- return (self.bench == other.bench and
- self.config == other.config and
- self.time_type == other.time_type and
- self.settings == other.settings)
-
- def __hash__(self):
- return (hash(self.bench) ^
- hash(self.config) ^
- hash(self.time_type) ^
- hash(frozenset(self.settings.iteritems())))
-
-def get_latest_revision(directory):
- """Returns the latest revision number found within this directory.
- """
- latest_revision_found = -1
- for bench_file in os.listdir(directory):
- file_name_match = re.match('bench_r(\d+)_(\S+)', bench_file)
- if (file_name_match is None):
- continue
- revision = int(file_name_match.group(1))
- if revision > latest_revision_found:
- latest_revision_found = revision
- if latest_revision_found < 0:
- return None
- else:
- return latest_revision_found
-
-def parse_dir(directory, default_settings, oldest_revision, newest_revision,
- rep):
- """Parses bench data from files like bench_r<revision>_<scalar>.
-
- (str, {str, str}, Number, Number) -> {int:[BenchDataPoints]}"""
- revision_data_points = {} # {revision : [BenchDataPoints]}
- file_list = os.listdir(directory)
- file_list.sort()
- for bench_file in file_list:
- file_name_match = re.match('bench_r(\d+)_(\S+)', bench_file)
- if (file_name_match is None):
- continue
-
- revision = int(file_name_match.group(1))
- scalar_type = file_name_match.group(2)
-
- if (revision < oldest_revision or revision > newest_revision):
- continue
-
- file_handle = open(directory + '/' + bench_file, 'r')
-
- if (revision not in revision_data_points):
- revision_data_points[revision] = []
- default_settings['scalar'] = scalar_type
- revision_data_points[revision].extend(
- bench_util.parse(default_settings, file_handle, rep))
- file_handle.close()
- return revision_data_points
-
-def add_to_revision_data_points(new_point, revision, revision_data_points):
- """Add new_point to set of revision_data_points we are building up.
- """
- if (revision not in revision_data_points):
- revision_data_points[revision] = []
- revision_data_points[revision].append(new_point)
-
-def filter_data_points(unfiltered_revision_data_points):
- """Filter out any data points that are utterly bogus.
-
- Returns (allowed_revision_data_points, ignored_revision_data_points):
- allowed_revision_data_points: points that survived the filter
- ignored_revision_data_points: points that did NOT survive the filter
- """
- allowed_revision_data_points = {} # {revision : [BenchDataPoints]}
- ignored_revision_data_points = {} # {revision : [BenchDataPoints]}
- revisions = unfiltered_revision_data_points.keys()
- revisions.sort()
- for revision in revisions:
- for point in unfiltered_revision_data_points[revision]:
- if point.time < MIN_REASONABLE_TIME or point.time > MAX_REASONABLE_TIME:
- add_to_revision_data_points(point, revision, ignored_revision_data_points)
- else:
- add_to_revision_data_points(point, revision, allowed_revision_data_points)
- return (allowed_revision_data_points, ignored_revision_data_points)
-
-def get_abs_path(relative_path):
- """My own implementation of os.path.abspath() that better handles paths
- which approach Window's 260-character limit.
- See https://code.google.com/p/skia/issues/detail?id=674
-
- This implementation adds path components one at a time, resolving the
- absolute path each time, to take advantage of any chdirs into outer
- directories that will shorten the total path length.
-
- TODO: share a single implementation with upload_to_bucket.py, instead
- of pasting this same code into both files."""
- if os.path.isabs(relative_path):
- return relative_path
- path_parts = relative_path.split(os.sep)
- abs_path = os.path.abspath('.')
- for path_part in path_parts:
- abs_path = os.path.abspath(os.path.join(abs_path, path_part))
- return abs_path
-
-def redirect_stdout(output_path):
- """Redirect all following stdout to a file.
-
- You may be asking yourself, why redirect stdout within Python rather than
- redirecting the script's output in the calling shell?
- The answer lies in https://code.google.com/p/skia/issues/detail?id=674
- ('buildbot: windows GenerateBenchGraphs step fails due to filename length'):
- On Windows, we need to generate the absolute path within Python to avoid
- the operating system's 260-character pathname limit, including chdirs."""
- abs_path = get_abs_path(output_path)
- sys.stdout = open(abs_path, 'w')
-
-def create_lines(revision_data_points, settings
- , bench_of_interest, config_of_interest, time_of_interest
- , time_to_ignore):
- """Convert revision data into a dictionary of line data.
-
- Args:
- revision_data_points: a dictionary with integer keys (revision #) and a
- list of bench data points as values
- settings: a dictionary of setting names to value
- bench_of_interest: optional filter parameters: which bench type is of
- interest. If None, process them all.
- config_of_interest: optional filter parameters: which config type is of
- interest. If None, process them all.
- time_of_interest: optional filter parameters: which timer type is of
- interest. If None, process them all.
- time_to_ignore: optional timer type to ignore
-
- Returns:
- a dictionary of this form:
- keys = Label objects
- values = a list of (x, y) tuples sorted such that x values increase
- monotonically
- """
- revisions = revision_data_points.keys()
- revisions.sort()
- lines = {} # {Label:[(x,y)] | x[n] <= x[n+1]}
- for revision in revisions:
- for point in revision_data_points[revision]:
- if (bench_of_interest is not None and
- not bench_of_interest == point.bench):
- continue
-
- if (config_of_interest is not None and
- not config_of_interest == point.config):
- continue
-
- if (time_of_interest is not None and
- not time_of_interest == point.time_type):
- continue
- elif (time_to_ignore is not None and
- time_to_ignore == point.time_type):
- continue
-
- skip = False
- for key, value in settings.items():
- if key in point.settings and point.settings[key] != value:
- skip = True
- break
- if skip:
- continue
-
- line_name = Label(point.bench
- , point.config
- , point.time_type
- , point.settings)
-
- if line_name not in lines:
- lines[line_name] = []
-
- lines[line_name].append((revision, point.time))
-
- return lines
-
-def bounds(lines):
- """Finds the bounding rectangle for the lines.
-
- {Label:[(x,y)]} -> ((min_x, min_y),(max_x,max_y))"""
- min_x = bench_util.Max
- min_y = bench_util.Max
- max_x = bench_util.Min
- max_y = bench_util.Min
-
- for line in lines.itervalues():
- for x, y in line:
- min_x = min(min_x, x)
- min_y = min(min_y, y)
- max_x = max(max_x, x)
- max_y = max(max_y, y)
-
- return ((min_x, min_y), (max_x, max_y))
-
-def create_regressions(lines, start_x, end_x):
- """Creates regression data from line segments.
-
- ({Label:[(x,y)] | [n].x <= [n+1].x}, Number, Number)
- -> {Label:LinearRegression}"""
- regressions = {} # {Label : LinearRegression}
-
- for label, line in lines.iteritems():
- regression_line = [p for p in line if start_x <= p[0] <= end_x]
-
- if (len(regression_line) < 2):
- continue
- regression = bench_util.LinearRegression(regression_line)
- regressions[label] = regression
-
- return regressions
-
-def bounds_slope(regressions):
- """Finds the extreme up and down slopes of a set of linear regressions.
-
- ({Label:LinearRegression}) -> (max_up_slope, min_down_slope)"""
- max_up_slope = 0
- min_down_slope = 0
- for regression in regressions.itervalues():
- min_slope = regression.find_min_slope()
- max_up_slope = max(max_up_slope, min_slope)
- min_down_slope = min(min_down_slope, min_slope)
-
- return (max_up_slope, min_down_slope)
-
-def main():
- """Parses command line and writes output."""
-
- try:
- opts, _ = getopt.getopt(sys.argv[1:]
- , "a:b:c:d:e:f:i:l:m:o:r:s:t:x:y:"
- , "default-setting=")
- except getopt.GetoptError, err:
- print str(err)
- usage()
- sys.exit(2)
-
- directory = None
- config_of_interest = None
- bench_of_interest = None
- time_of_interest = None
- time_to_ignore = None
- output_path = None
- bench_expectations = {}
- appengine_url = None # used for adding data to appengine datastore
- rep = None # bench representation algorithm
- revision_range = '0:'
- regression_range = '0:'
- latest_revision = None
- requested_height = None
- requested_width = None
- title = 'Bench graph'
- settings = {}
- default_settings = {}
-
- def parse_range(range):
- """Takes '<old>[:<new>]' as a string and returns (old, new).
- Any revision numbers that are dependent on the latest revision number
- will be filled in based on latest_revision.
- """
- old, _, new = range.partition(":")
- old = int(old)
- if old < 0:
- old += latest_revision;
- if not new:
- new = latest_revision;
- new = int(new)
- if new < 0:
- new += latest_revision;
- return (old, new)
-
- def add_setting(settings, setting):
- """Takes <key>[=<value>] adds {key:value} or {key:True} to settings."""
- name, _, value = setting.partition('=')
- if not value:
- settings[name] = True
- else:
- settings[name] = value
-
- def read_expectations(expectations, filename):
- """Reads expectations data from file and put in expectations dict."""
- for expectation in open(filename).readlines():
- elements = expectation.strip().split(',')
- if not elements[0] or elements[0].startswith('#'):
- continue
- if len(elements) != 5:
- raise Exception("Invalid expectation line format: %s" %
- expectation)
- bench_entry = elements[0] + ',' + elements[1]
- if bench_entry in expectations:
- raise Exception("Dup entries for bench expectation %s" %
- bench_entry)
- # [<Bench_BmpConfig_TimeType>,<Platform-Alg>] -> (LB, UB)
- expectations[bench_entry] = (float(elements[-2]),
- float(elements[-1]))
-
- def check_expectations(lines, expectations, newest_revision, key_suffix):
- """Check if there are benches in latest rev outside expected range.
- For exceptions, also outputs URL link for the dashboard plot.
- The link history token format here only works for single-line plots.
- """
- # The platform for this bot, to pass to the dashboard plot.
- platform = key_suffix[ : key_suffix.rfind('-')]
- # Starting revision for the dashboard plot.
- start_rev = str(newest_revision - 100) # Displays about 100 revisions.
- exceptions = []
- for line in lines:
- line_str = str(line)
- line_str = line_str[ : line_str.find('_{')]
- bench_platform_key = line_str + ',' + key_suffix
- this_revision, this_bench_value = lines[line][-1]
- if (this_revision != newest_revision or
- bench_platform_key not in expectations):
- # Skip benches without value for latest revision.
- continue
- this_min, this_max = expectations[bench_platform_key]
- if this_bench_value < this_min or this_bench_value > this_max:
- link = ''
- # For skp benches out of range, create dashboard plot link.
- if line_str.find('.skp_') > 0:
- # Extract bench and config for dashboard plot.
- bench, config = line_str.strip('_').split('.skp_')
- link = ' <a href="'
- link += 'http://go/skpdash/SkpDash.html#%s~%s~%s~%s" ' % (
- start_rev, bench, platform, config)
- link += 'target="_blank">graph</a>'
- exception = 'Bench %s value %s out of range [%s, %s].%s' % (
- bench_platform_key, this_bench_value, this_min, this_max,
- link)
- exceptions.append(exception)
- if exceptions:
- raise Exception('Bench values out of range:\n' +
- '\n'.join(exceptions))
-
- def write_to_appengine(line_data_dict, url, newest_revision, bot):
- """Writes latest bench values to appengine datastore.
- line_data_dict: dictionary from create_lines.
- url: the appengine url used to send bench values to write
- newest_revision: the latest revision that this script reads
- bot: the bot platform the bench is run on
- """
- config_data_dic = {}
- for label in line_data_dict.iterkeys():
- if not label.bench.endswith('.skp') or label.time_type:
- # filter out non-picture and non-walltime benches
- continue
- config = label.config
- rev, val = line_data_dict[label][-1]
- # This assumes that newest_revision is >= the revision of the last
- # data point we have for each line.
- if rev != newest_revision:
- continue
- if config not in config_data_dic:
- config_data_dic[config] = []
- config_data_dic[config].append(label.bench.replace('.skp', '') +
- ':%.2f' % val)
- for config in config_data_dic:
- if config_data_dic[config]:
- data = {'master': 'Skia', 'bot': bot, 'test': config,
- 'revision': newest_revision,
- 'benches': ','.join(config_data_dic[config])}
- req = urllib2.Request(appengine_url,
- urllib.urlencode({'data': json.dumps(data)}))
- try:
- urllib2.urlopen(req)
- except urllib2.HTTPError, e:
- sys.stderr.write("HTTPError for JSON data %s: %s\n" % (
- data, e))
- except urllib2.URLError, e:
- sys.stderr.write("URLError for JSON data %s: %s\n" % (
- data, e))
- except httplib.HTTPException, e:
- sys.stderr.write("HTTPException for JSON data %s: %s\n" % (
- data, e))
-
- try:
- for option, value in opts:
- if option == "-a":
- appengine_url = value
- elif option == "-b":
- bench_of_interest = value
- elif option == "-c":
- config_of_interest = value
- elif option == "-d":
- directory = value
- elif option == "-e":
- read_expectations(bench_expectations, value)
- elif option == "-f":
- regression_range = value
- elif option == "-i":
- time_to_ignore = value
- elif option == "-l":
- title = value
- elif option == "-m":
- rep = value
- elif option == "-o":
- output_path = value
- redirect_stdout(output_path)
- elif option == "-r":
- revision_range = value
- elif option == "-s":
- add_setting(settings, value)
- elif option == "-t":
- time_of_interest = value
- elif option == "-x":
- requested_width = int(value)
- elif option == "-y":
- requested_height = int(value)
- elif option == "--default-setting":
- add_setting(default_settings, value)
- else:
- usage()
- assert False, "unhandled option"
- except ValueError:
- usage()
- sys.exit(2)
-
- if directory is None:
- usage()
- sys.exit(2)
-
- if time_of_interest:
- time_to_ignore = None
-
- # The title flag (-l) provided in buildbot slave is in the format
- # Bench_Performance_for_<platform>, and we want to extract <platform>
- # for use in platform_and_alg to track matching benches later. If title flag
- # is not in this format, there may be no matching benches in the file
- # provided by the expectation_file flag (-e).
- bot = title # To store the platform as bot name
- platform_and_alg = title
- if platform_and_alg.startswith(TITLE_PREAMBLE):
- bot = platform_and_alg[TITLE_PREAMBLE_LENGTH:]
- platform_and_alg = bot + '-' + rep
- title += ' [representation: %s]' % rep
-
- latest_revision = get_latest_revision(directory)
- oldest_revision, newest_revision = parse_range(revision_range)
- oldest_regression, newest_regression = parse_range(regression_range)
-
- unfiltered_revision_data_points = parse_dir(directory
- , default_settings
- , oldest_revision
- , newest_revision
- , rep)
-
- # Filter out any data points that are utterly bogus... make sure to report
- # that we did so later!
- (allowed_revision_data_points, ignored_revision_data_points) = filter_data_points(
- unfiltered_revision_data_points)
-
- # Update oldest_revision and newest_revision based on the data we could find
- all_revision_numbers = allowed_revision_data_points.keys()
- oldest_revision = min(all_revision_numbers)
- newest_revision = max(all_revision_numbers)
-
- lines = create_lines(allowed_revision_data_points
- , settings
- , bench_of_interest
- , config_of_interest
- , time_of_interest
- , time_to_ignore)
-
- regressions = create_regressions(lines
- , oldest_regression
- , newest_regression)
-
- if output_path:
- output_xhtml(lines, oldest_revision, newest_revision,
- ignored_revision_data_points, regressions, requested_width,
- requested_height, title)
-
- if appengine_url:
- write_to_appengine(lines, appengine_url, newest_revision, bot)
-
- if bench_expectations:
- check_expectations(lines, bench_expectations, newest_revision,
- platform_and_alg)
-
-def qa(out):
- """Stringify input and quote as an xml attribute."""
- return xml.sax.saxutils.quoteattr(str(out))
-def qe(out):
- """Stringify input and escape as xml data."""
- return xml.sax.saxutils.escape(str(out))
-
-def create_select(qualifier, lines, select_id=None):
- """Output select with options showing lines which qualifier maps to it.
-
- ((Label) -> str, {Label:_}, str?) -> _"""
- options = {} #{ option : [Label]}
- for label in lines.keys():
- option = qualifier(label)
- if (option not in options):
- options[option] = []
- options[option].append(label)
- option_list = list(options.keys())
- option_list.sort()
- print '<select class="lines"',
- if select_id is not None:
- print 'id=%s' % qa(select_id)
- print 'multiple="true" size="10" onchange="updateSvg();">'
- for option in option_list:
- print '<option value=' + qa('[' +
- reduce(lambda x,y:x+json.dumps(str(y))+',',options[option],"")[0:-1]
- + ']') + '>'+qe(option)+'</option>'
- print '</select>'
-
-def output_ignored_data_points_warning(ignored_revision_data_points):
- """Write description of ignored_revision_data_points to stdout as xhtml.
- """
- num_ignored_points = 0
- description = ''
- revisions = ignored_revision_data_points.keys()
- if revisions:
- revisions.sort()
- revisions.reverse()
- for revision in revisions:
- num_ignored_points += len(ignored_revision_data_points[revision])
- points_at_this_revision = []
- for point in ignored_revision_data_points[revision]:
- points_at_this_revision.append(point.bench)
- points_at_this_revision.sort()
- description += 'r%d: %s\n' % (revision, points_at_this_revision)
- if num_ignored_points == 0:
- print 'Did not discard any data points; all were within the range [%d-%d]' % (
- MIN_REASONABLE_TIME, MAX_REASONABLE_TIME)
- else:
- print '<table width="100%" bgcolor="ff0000"><tr><td align="center">'
- print 'Discarded %d data points outside of range [%d-%d]' % (
- num_ignored_points, MIN_REASONABLE_TIME, MAX_REASONABLE_TIME)
- print '</td></tr><tr><td width="100%" align="center">'
- print ('<textarea rows="4" style="width:97%" readonly="true" wrap="off">'
- + qe(description) + '</textarea>')
- print '</td></tr></table>'
-
-def output_xhtml(lines, oldest_revision, newest_revision, ignored_revision_data_points,
- regressions, requested_width, requested_height, title):
- """Outputs an svg/xhtml view of the data."""
- print '<!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Strict//EN"',
- print '"http://www.w3.org/TR/xhtml1/DTD/xhtml1-strict.dtd">'
- print '<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en">'
- print '<head>'
- print '<title>%s</title>' % qe(title)
- print '</head>'
- print '<body>'
-
- output_svg(lines, regressions, requested_width, requested_height)
-
- #output the manipulation controls
- print """
-<script type="text/javascript">//<![CDATA[
- function getElementsByClass(node, searchClass, tag) {
- var classElements = new Array();
- var elements = node.getElementsByTagName(tag);
- var pattern = new RegExp("^|\\s"+searchClass+"\\s|$");
- for (var i = 0, elementsFound = 0; i < elements.length; ++i) {
- if (pattern.test(elements[i].className)) {
- classElements[elementsFound] = elements[i];
- ++elementsFound;
- }
- }
- return classElements;
- }
- function getAllLines() {
- var selectElem = document.getElementById('benchSelect');
- var linesObj = {};
- for (var i = 0; i < selectElem.options.length; ++i) {
- var lines = JSON.parse(selectElem.options[i].value);
- for (var j = 0; j < lines.length; ++j) {
- linesObj[lines[j]] = true;
- }
- }
- return linesObj;
- }
- function getOptions(selectElem) {
- var linesSelectedObj = {};
- for (var i = 0; i < selectElem.options.length; ++i) {
- if (!selectElem.options[i].selected) continue;
-
- var linesSelected = JSON.parse(selectElem.options[i].value);
- for (var j = 0; j < linesSelected.length; ++j) {
- linesSelectedObj[linesSelected[j]] = true;
- }
- }
- return linesSelectedObj;
- }
- function objectEmpty(obj) {
- for (var p in obj) {
- return false;
- }
- return true;
- }
- function markSelectedLines(selectElem, allLines) {
- var linesSelected = getOptions(selectElem);
- if (!objectEmpty(linesSelected)) {
- for (var line in allLines) {
- allLines[line] &= (linesSelected[line] == true);
- }
- }
- }
- function updateSvg() {
- var allLines = getAllLines();
-
- var selects = getElementsByClass(document, 'lines', 'select');
- for (var i = 0; i < selects.length; ++i) {
- markSelectedLines(selects[i], allLines);
- }
-
- for (var line in allLines) {
- var svgLine = document.getElementById(line);
- var display = (allLines[line] ? 'inline' : 'none');
- svgLine.setAttributeNS(null,'display', display);
- }
- }
-
- function mark(markerId) {
- for (var line in getAllLines()) {
- var svgLineGroup = document.getElementById(line);
- var display = svgLineGroup.getAttributeNS(null,'display');
- if (display == null || display == "" || display != "none") {
- var svgLine = document.getElementById(line+'_line');
- if (markerId == null) {
- svgLine.removeAttributeNS(null,'marker-mid');
- } else {
- svgLine.setAttributeNS(null,'marker-mid', markerId);
- }
- }
- }
- }
-//]]></script>"""
-
- all_settings = {}
- variant_settings = set()
- for label in lines.keys():
- for key, value in label.settings.items():
- if key not in all_settings:
- all_settings[key] = value
- elif all_settings[key] != value:
- variant_settings.add(key)
-
- print '<table border="0" width="%s">' % requested_width
- #output column headers
- print """
-<tr valign="top"><td width="50%">
-<table border="0" width="100%">
-<tr><td align="center"><table border="0">
-<form>
-<tr valign="bottom" align="center">
-<td width="1">Bench&nbsp;Type</td>
-<td width="1">Bitmap Config</td>
-<td width="1">Timer&nbsp;Type (Cpu/Gpu/wall)</td>
-"""
-
- for k in variant_settings:
- print '<td width="1">%s</td>' % qe(k)
-
- print '<td width="1"><!--buttons--></td></tr>'
-
- #output column contents
- print '<tr valign="top" align="center">'
- print '<td width="1">'
- create_select(lambda l: l.bench, lines, 'benchSelect')
- print '</td><td width="1">'
- create_select(lambda l: l.config, lines)
- print '</td><td width="1">'
- create_select(lambda l: l.time_type, lines)
-
- for k in variant_settings:
- print '</td><td width="1">'
- create_select(lambda l: l.settings.get(k, " "), lines)
-
- print '</td><td width="1"><button type="button"',
- print 'onclick=%s' % qa("mark('url(#circleMark)'); return false;"),
- print '>Mark Points</button>'
- print '<button type="button" onclick="mark(null);">Clear Points</button>'
- print '</td>'
- print """
-</tr>
-</form>
-</table></td></tr>
-<tr><td align="center">
-<hr />
-"""
-
- output_ignored_data_points_warning(ignored_revision_data_points)
- print '</td></tr></table>'
- print '</td><td width="2%"><!--gutter--></td>'
-
- print '<td><table border="0">'
- print '<tr><td align="center">%s<br></br>revisions r%s - r%s</td></tr>' % (
- qe(title),
- bench_util.CreateRevisionLink(oldest_revision),
- bench_util.CreateRevisionLink(newest_revision))
- print """
-<tr><td align="left">
-<p>Brighter red indicates tests that have gotten worse; brighter green
-indicates tests that have gotten better.</p>
-<p>To highlight individual tests, hold down CONTROL and mouse over
-graph lines.</p>
-<p>To highlight revision numbers, hold down SHIFT and mouse over
-the graph area.</p>
-<p>To only show certain tests on the graph, select any combination of
-tests in the selectors at left. (To show all, select all.)</p>
-<p>Use buttons at left to mark/clear points on the lines for selected
-benchmarks.</p>
-</td></tr>
-</table>
-
-</td>
-</tr>
-</table>
-</body>
-</html>"""
-
-def compute_size(requested_width, requested_height, rev_width, time_height):
- """Converts potentially empty requested size into a concrete size.
-
- (Number?, Number?) -> (Number, Number)"""
- pic_width = 0
- pic_height = 0
- if (requested_width is not None and requested_height is not None):
- pic_height = requested_height
- pic_width = requested_width
-
- elif (requested_width is not None):
- pic_width = requested_width
- pic_height = pic_width * (float(time_height) / rev_width)
-
- elif (requested_height is not None):
- pic_height = requested_height
- pic_width = pic_height * (float(rev_width) / time_height)
-
- else:
- pic_height = 800
- pic_width = max(rev_width*3
- , pic_height * (float(rev_width) / time_height))
-
- return (pic_width, pic_height)
-
-def output_svg(lines, regressions, requested_width, requested_height):
- """Outputs an svg view of the data."""
-
- (global_min_x, _), (global_max_x, global_max_y) = bounds(lines)
- max_up_slope, min_down_slope = bounds_slope(regressions)
-
- #output
- global_min_y = 0
- x = global_min_x
- y = global_min_y
- w = global_max_x - global_min_x
- h = global_max_y - global_min_y
- font_size = 16
- line_width = 2
-
- # If there is nothing to see, don't try to draw anything.
- if w == 0 or h == 0:
- return
-
- pic_width, pic_height = compute_size(requested_width, requested_height
- , w, h)
-
- def cw(w1):
- """Converts a revision difference to display width."""
- return (pic_width / float(w)) * w1
- def cx(x):
- """Converts a revision to a horizontal display position."""
- return cw(x - global_min_x)
-
- def ch(h1):
- """Converts a time difference to a display height."""
- return -(pic_height / float(h)) * h1
- def cy(y):
- """Converts a time to a vertical display position."""
- return pic_height + ch(y - global_min_y)
-
- print '<!--Picture height %.2f corresponds to bench value %.2f.-->' % (
- pic_height, h)
- print '<svg',
- print 'width=%s' % qa(str(pic_width)+'px')
- print 'height=%s' % qa(str(pic_height)+'px')
- print 'viewBox="0 0 %s %s"' % (str(pic_width), str(pic_height))
- print 'onclick=%s' % qa(
- "var event = arguments[0] || window.event;"
- " if (event.shiftKey) { highlightRevision(null); }"
- " if (event.ctrlKey) { highlight(null); }"
- " return false;")
- print 'xmlns="http://www.w3.org/2000/svg"'
- print 'xmlns:xlink="http://www.w3.org/1999/xlink">'
-
- print """
-<defs>
- <marker id="circleMark"
- viewBox="0 0 2 2" refX="1" refY="1"
- markerUnits="strokeWidth"
- markerWidth="2" markerHeight="2"
- orient="0">
- <circle cx="1" cy="1" r="1"/>
- </marker>
-</defs>"""
-
- #output the revisions
- print """
-<script type="text/javascript">//<![CDATA[
- var previousRevision;
- var previousRevisionFill;
- var previousRevisionStroke
- function highlightRevision(id) {
- if (previousRevision == id) return;
-
- document.getElementById('revision').firstChild.nodeValue = 'r' + id;
- document.getElementById('rev_link').setAttribute('xlink:href',
- 'http://code.google.com/p/skia/source/detail?r=' + id);
-
- var preRevision = document.getElementById(previousRevision);
- if (preRevision) {
- preRevision.setAttributeNS(null,'fill', previousRevisionFill);
- preRevision.setAttributeNS(null,'stroke', previousRevisionStroke);
- }
-
- var revision = document.getElementById(id);
- previousRevision = id;
- if (revision) {
- previousRevisionFill = revision.getAttributeNS(null,'fill');
- revision.setAttributeNS(null,'fill','rgb(100%, 95%, 95%)');
-
- previousRevisionStroke = revision.getAttributeNS(null,'stroke');
- revision.setAttributeNS(null,'stroke','rgb(100%, 90%, 90%)');
- }
- }
-//]]></script>"""
-
- def print_rect(x, y, w, h, revision):
- """Outputs a revision rectangle in display space,
- taking arguments in revision space."""
- disp_y = cy(y)
- disp_h = ch(h)
- if disp_h < 0:
- disp_y += disp_h
- disp_h = -disp_h
-
- print '<rect id=%s x=%s y=%s' % (qa(revision), qa(cx(x)), qa(disp_y),),
- print 'width=%s height=%s' % (qa(cw(w)), qa(disp_h),),
- print 'fill="white"',
- print 'stroke="rgb(98%%,98%%,88%%)" stroke-width=%s' % qa(line_width),
- print 'onmouseover=%s' % qa(
- "var event = arguments[0] || window.event;"
- " if (event.shiftKey) {"
- " highlightRevision('"+str(revision)+"');"
- " return false;"
- " }"),
- print ' />'
-
- xes = set()
- for line in lines.itervalues():
- for point in line:
- xes.add(point[0])
- revisions = list(xes)
- revisions.sort()
-
- left = x
- current_revision = revisions[0]
- for next_revision in revisions[1:]:
- width = (((next_revision - current_revision) / 2.0)
- + (current_revision - left))
- print_rect(left, y, width, h, current_revision)
- left += width
- current_revision = next_revision
- print_rect(left, y, x+w - left, h, current_revision)
-
- #output the lines
- print """
-<script type="text/javascript">//<![CDATA[
- var previous;
- var previousColor;
- var previousOpacity;
- function highlight(id) {
- if (previous == id) return;
-
- document.getElementById('label').firstChild.nodeValue = id;
-
- var preGroup = document.getElementById(previous);
- if (preGroup) {
- var preLine = document.getElementById(previous+'_line');
- preLine.setAttributeNS(null,'stroke', previousColor);
- preLine.setAttributeNS(null,'opacity', previousOpacity);
-
- var preSlope = document.getElementById(previous+'_linear');
- if (preSlope) {
- preSlope.setAttributeNS(null,'visibility', 'hidden');
- }
- }
-
- var group = document.getElementById(id);
- previous = id;
- if (group) {
- group.parentNode.appendChild(group);
-
- var line = document.getElementById(id+'_line');
- previousColor = line.getAttributeNS(null,'stroke');
- previousOpacity = line.getAttributeNS(null,'opacity');
- line.setAttributeNS(null,'stroke', 'blue');
- line.setAttributeNS(null,'opacity', '1');
-
- var slope = document.getElementById(id+'_linear');
- if (slope) {
- slope.setAttributeNS(null,'visibility', 'visible');
- }
- }
- }
-//]]></script>"""
-
- # Add a new element to each item in the 'lines' list: the label in string
- # form. Then use that element to sort the list.
- sorted_lines = []
- for label, line in lines.items():
- sorted_lines.append([str(label), label, line])
- sorted_lines.sort()
-
- for label_as_string, label, line in sorted_lines:
- print '<g id=%s>' % qa(label_as_string)
- r = 128
- g = 128
- b = 128
- a = .10
- if label in regressions:
- regression = regressions[label]
- min_slope = regression.find_min_slope()
- if min_slope < 0:
- d = max(0, (min_slope / min_down_slope))
- g += int(d*128)
- a += d*0.9
- elif min_slope > 0:
- d = max(0, (min_slope / max_up_slope))
- r += int(d*128)
- a += d*0.9
-
- slope = regression.slope
- intercept = regression.intercept
- min_x = regression.min_x
- max_x = regression.max_x
- print '<polyline id=%s' % qa(str(label)+'_linear'),
- print 'fill="none" stroke="yellow"',
- print 'stroke-width=%s' % qa(abs(ch(regression.serror*2))),
- print 'opacity="0.5" pointer-events="none" visibility="hidden"',
- print 'points="',
- print '%s,%s' % (str(cx(min_x)), str(cy(slope*min_x + intercept))),
- print '%s,%s' % (str(cx(max_x)), str(cy(slope*max_x + intercept))),
- print '"/>'
-
- print '<polyline id=%s' % qa(str(label)+'_line'),
- print 'onmouseover=%s' % qa(
- "var event = arguments[0] || window.event;"
- " if (event.ctrlKey) {"
- " highlight('"+str(label).replace("'", "\\'")+"');"
- " return false;"
- " }"),
- print 'fill="none" stroke="rgb(%s,%s,%s)"' % (str(r), str(g), str(b)),
- print 'stroke-width=%s' % qa(line_width),
- print 'opacity=%s' % qa(a),
- print 'points="',
- for point in line:
- print '%s,%s' % (str(cx(point[0])), str(cy(point[1]))),
- print '"/>'
-
- print '</g>'
-
- #output the labels
- print '<text id="label" x="0" y=%s' % qa(font_size),
- print 'font-size=%s> </text>' % qa(font_size)
-
- print '<a id="rev_link" xlink:href="" target="_top">'
- print '<text id="revision" x="0" y=%s style="' % qa(font_size*2)
- print 'font-size: %s; ' % qe(font_size)
- print 'stroke: #0000dd; text-decoration: underline; '
- print '"> </text></a>'
-
- print '</svg>'
-
-if __name__ == "__main__":
- main()
diff --git a/bench/check_bench_regressions.py b/bench/check_bench_regressions.py
new file mode 100644
index 0000000000..b26eb56ae0
--- /dev/null
+++ b/bench/check_bench_regressions.py
@@ -0,0 +1,218 @@
+'''
+Created on May 16, 2011
+
+@author: bungeman
+'''
+import bench_util
+import getopt
+import httplib
+import itertools
+import json
+import os
+import re
+import sys
+import urllib
+import urllib2
+import xml.sax.saxutils
+
+# Maximum expected number of characters we expect in an svn revision.
+MAX_SVN_REV_LENGTH = 5
+
+def usage():
+ """Prints simple usage information."""
+
+ print '-a <representation_alg> bench representation algorithm to use. '
+ print ' Defaults to "25th". See bench_util.py for details.'
+ print '-b <builder> name of the builder whose bench data we are checking.'
+ print '-d <dir> a directory containing bench_<revision>_<scalar> files.'
+ print '-e <file> file containing expected bench builder values/ranges.'
+ print ' Will raise exception if actual bench values are out of range.'
+ print ' See bench_expectations_<builder>.txt for data format / examples.'
+ print '-r <revision> the git commit hash or svn revision for checking '
+ print ' bench values.'
+
+
+class Label:
+ """The information in a label.
+
+ (str, str, str, str, {str:str})"""
+ def __init__(self, bench, config, time_type, settings):
+ self.bench = bench
+ self.config = config
+ self.time_type = time_type
+ self.settings = settings
+
+ def __repr__(self):
+ return "Label(%s, %s, %s, %s)" % (
+ str(self.bench),
+ str(self.config),
+ str(self.time_type),
+ str(self.settings),
+ )
+
+ def __str__(self):
+ return "%s_%s_%s_%s" % (
+ str(self.bench),
+ str(self.config),
+ str(self.time_type),
+ str(self.settings),
+ )
+
+ def __eq__(self, other):
+ return (self.bench == other.bench and
+ self.config == other.config and
+ self.time_type == other.time_type and
+ self.settings == other.settings)
+
+ def __hash__(self):
+ return (hash(self.bench) ^
+ hash(self.config) ^
+ hash(self.time_type) ^
+ hash(frozenset(self.settings.iteritems())))
+
+def parse_dir(directory, default_settings, revision, rep):
+ """Parses bench data from bench logs files.
+ revision can be either svn revision or git commit hash.
+ """
+ revision_data_points = [] # list of BenchDataPoint
+ file_list = os.listdir(directory)
+ file_list.sort()
+ for bench_file in file_list:
+ scalar_type = None
+ # Scalar type, if any, is in the bench filename after revision
+ if (len(revision) > MAX_SVN_REV_LENGTH and
+ bench_file.startswith('bench_' + revision + '_')):
+ # The revision is GIT commit hash.
+ scalar_type = bench_file[len(revision) + len('bench_') + 1:]
+ elif (bench_file.startswith('bench_r' + revision + '_') and
+ revision.isdigit()):
+ # The revision is SVN number
+ scalar_type = bench_file[len(revision) + len('bench_r') + 1:]
+ else:
+ continue
+
+ file_handle = open(directory + '/' + bench_file, 'r')
+
+ default_settings['scalar'] = scalar_type
+ revision_data_points.extend(
+ bench_util.parse(default_settings, file_handle, rep))
+ file_handle.close()
+ return revision_data_points
+
+def create_bench_dict(revision_data_points):
+ """Convert current revision data into a dictionary of line data.
+
+ Args:
+ revision_data_points: a list of bench data points
+
+ Returns:
+ a dictionary of this form:
+ keys = Label objects
+ values = the corresponding bench value
+ """
+ bench_dict = {}
+ for point in revision_data_points:
+ point_name = Label(point.bench,point.config,point.time_type,
+ point.settings)
+ if point_name not in bench_dict:
+ bench_dict[point_name] = point.time
+ else:
+ raise Exception('Duplicate expectation entry: ' + str(point_name))
+
+ return bench_dict
+
+def read_expectations(expectations, filename):
+ """Reads expectations data from file and put in expectations dict."""
+ for expectation in open(filename).readlines():
+ elements = expectation.strip().split(',')
+ if not elements[0] or elements[0].startswith('#'):
+ continue
+ if len(elements) != 5:
+ raise Exception("Invalid expectation line format: %s" %
+ expectation)
+ bench_entry = elements[0] + ',' + elements[1]
+ if bench_entry in expectations:
+ raise Exception("Dup entries for bench expectation %s" %
+ bench_entry)
+ # [<Bench_BmpConfig_TimeType>,<Platform-Alg>] -> (LB, UB)
+ expectations[bench_entry] = (float(elements[-2]),
+ float(elements[-1]))
+
+def check_expectations(lines, expectations, revision, key_suffix):
+ """Check if there are benches in the given revising out of range.
+ """
+ # The platform for this bot, to pass to the dashboard plot.
+ platform = key_suffix[ : key_suffix.rfind('-')]
+ exceptions = []
+ for line in lines:
+ line_str = str(line)
+ line_str = line_str[ : line_str.find('_{')]
+ bench_platform_key = line_str + ',' + key_suffix
+ if bench_platform_key not in expectations:
+ continue
+ this_bench_value = lines[line]
+ this_min, this_max = expectations[bench_platform_key]
+ if this_bench_value < this_min or this_bench_value > this_max:
+ exception = 'Bench %s value %s out of range [%s, %s].' % (
+ bench_platform_key, this_bench_value, this_min, this_max)
+ exceptions.append(exception)
+ if exceptions:
+ raise Exception('Bench values out of range:\n' +
+ '\n'.join(exceptions))
+
+def main():
+ """Parses command line and checks bench expectations."""
+ try:
+ opts, _ = getopt.getopt(sys.argv[1:],
+ "a:b:d:e:r:",
+ "default-setting=")
+ except getopt.GetoptError, err:
+ print str(err)
+ usage()
+ sys.exit(2)
+
+ directory = None
+ bench_expectations = {}
+ rep = '25th' # bench representation algorithm, default to 25th
+ rev = None # git commit hash or svn revision number
+ bot = None
+
+ try:
+ for option, value in opts:
+ if option == "-a":
+ rep = value
+ elif option == "-b":
+ bot = value
+ elif option == "-d":
+ directory = value
+ elif option == "-e":
+ read_expectations(bench_expectations, value)
+ elif option == "-r":
+ rev = value
+ else:
+ usage()
+ assert False, "unhandled option"
+ except ValueError:
+ usage()
+ sys.exit(2)
+
+ if directory is None or bot is None or rev is None:
+ usage()
+ sys.exit(2)
+
+ platform_and_alg = bot + '-' + rep
+
+ data_points = parse_dir(directory,
+ {}, # Sets default settings to empty.
+ rev,
+ rep)
+
+ bench_dict = create_bench_dict(data_points)
+
+ if bench_expectations:
+ check_expectations(bench_dict, bench_expectations, rev,
+ platform_and_alg)
+
+
+if __name__ == "__main__":
+ main()