aboutsummaryrefslogtreecommitdiffhomepage
path: root/bench
diff options
context:
space:
mode:
authorGravatar Mike Klein <mtklein@google.com>2014-08-08 17:28:53 -0400
committerGravatar Mike Klein <mtklein@google.com>2014-08-08 17:28:53 -0400
commite530eb370c084336b584a6dff5a9e6974d932dfa (patch)
tree9787ffcfb843b25dbf76e8242c5b12dd2a0e194a /bench
parent3b1c3d2c6160f714c4ebada05dd834f88fe5fd00 (diff)
Restore bench_util.py
Diffstat (limited to 'bench')
-rw-r--r--bench/bench_util.py356
1 files changed, 356 insertions, 0 deletions
diff --git a/bench/bench_util.py b/bench/bench_util.py
new file mode 100644
index 0000000000..b6fecb7ca8
--- /dev/null
+++ b/bench/bench_util.py
@@ -0,0 +1,356 @@
+'''
+Created on May 19, 2011
+
+@author: bungeman
+'''
+
+import os
+import re
+import math
+
+# bench representation algorithm constant names
+ALGORITHM_AVERAGE = 'avg'
+ALGORITHM_MEDIAN = 'med'
+ALGORITHM_MINIMUM = 'min'
+ALGORITHM_25TH_PERCENTILE = '25th'
+
+# Regular expressions used throughout.
+PER_SETTING_RE = '([^\s=]+)(?:=(\S+))?'
+SETTINGS_RE = 'skia bench:((?:\s+' + PER_SETTING_RE + ')*)'
+BENCH_RE = 'running bench (?:\[\d+ \d+\] )?\s*(\S+)'
+TIME_RE = '(?:(\w*)msecs = )?\s*((?:\d+\.\d+)(?:,\s*\d+\.\d+)*)'
+# non-per-tile benches have configs that don't end with ']' or '>'
+CONFIG_RE = '(\S+[^\]>]):\s+((?:' + TIME_RE + '\s+)+)'
+# per-tile bench lines are in the following format. Note that there are
+# non-averaged bench numbers in separate lines, which we ignore now due to
+# their inaccuracy.
+TILE_RE = (' tile_(\S+): tile \[\d+,\d+\] out of \[\d+,\d+\] <averaged>:'
+ ' ((?:' + TIME_RE + '\s+)+)')
+# for extracting tile layout
+TILE_LAYOUT_RE = ' out of \[(\d+),(\d+)\] <averaged>: '
+
+PER_SETTING_RE_COMPILED = re.compile(PER_SETTING_RE)
+SETTINGS_RE_COMPILED = re.compile(SETTINGS_RE)
+BENCH_RE_COMPILED = re.compile(BENCH_RE)
+TIME_RE_COMPILED = re.compile(TIME_RE)
+CONFIG_RE_COMPILED = re.compile(CONFIG_RE)
+TILE_RE_COMPILED = re.compile(TILE_RE)
+TILE_LAYOUT_RE_COMPILED = re.compile(TILE_LAYOUT_RE)
+
+class BenchDataPoint:
+ """A single data point produced by bench.
+ """
+ def __init__(self, bench, config, time_type, time, settings,
+ tile_layout='', per_tile_values=[], per_iter_time=[]):
+ # string name of the benchmark to measure
+ self.bench = bench
+ # string name of the configurations to run
+ self.config = config
+ # type of the timer in string: '' (walltime), 'c' (cpu) or 'g' (gpu)
+ self.time_type = time_type
+ # float number of the bench time value
+ self.time = time
+ # dictionary of the run settings
+ self.settings = settings
+ # how tiles cover the whole picture: '5x3' means 5 columns and 3 rows
+ self.tile_layout = tile_layout
+ # list of float for per_tile bench values, if applicable
+ self.per_tile_values = per_tile_values
+ # list of float for per-iteration bench time, if applicable
+ self.per_iter_time = per_iter_time
+
+ def __repr__(self):
+ return "BenchDataPoint(%s, %s, %s, %s, %s)" % (
+ str(self.bench),
+ str(self.config),
+ str(self.time_type),
+ str(self.time),
+ str(self.settings),
+ )
+
+class _ExtremeType(object):
+ """Instances of this class compare greater or less than other objects."""
+ def __init__(self, cmpr, rep):
+ object.__init__(self)
+ self._cmpr = cmpr
+ self._rep = rep
+
+ def __cmp__(self, other):
+ if isinstance(other, self.__class__) and other._cmpr == self._cmpr:
+ return 0
+ return self._cmpr
+
+ def __repr__(self):
+ return self._rep
+
+Max = _ExtremeType(1, "Max")
+Min = _ExtremeType(-1, "Min")
+
+class _ListAlgorithm(object):
+ """Algorithm for selecting the representation value from a given list.
+ representation is one of the ALGORITHM_XXX representation types."""
+ def __init__(self, data, representation=None):
+ if not representation:
+ representation = ALGORITHM_AVERAGE # default algorithm
+ self._data = data
+ self._len = len(data)
+ if representation == ALGORITHM_AVERAGE:
+ self._rep = sum(self._data) / self._len
+ else:
+ self._data.sort()
+ if representation == ALGORITHM_MINIMUM:
+ self._rep = self._data[0]
+ else:
+ # for percentiles, we use the value below which x% of values are
+ # found, which allows for better detection of quantum behaviors.
+ if representation == ALGORITHM_MEDIAN:
+ x = int(round(0.5 * self._len + 0.5))
+ elif representation == ALGORITHM_25TH_PERCENTILE:
+ x = int(round(0.25 * self._len + 0.5))
+ else:
+ raise Exception("invalid representation algorithm %s!" %
+ representation)
+ self._rep = self._data[x - 1]
+
+ def compute(self):
+ return self._rep
+
+def _ParseAndStoreTimes(config_re_compiled, is_per_tile, line, bench,
+ value_dic, layout_dic):
+ """Parses given bench time line with regex and adds data to value_dic.
+
+ config_re_compiled: precompiled regular expression for parsing the config
+ line.
+ is_per_tile: boolean indicating whether this is a per-tile bench.
+ If so, we add tile layout into layout_dic as well.
+ line: input string line to parse.
+ bench: name of bench for the time values.
+ value_dic: dictionary to store bench values. See bench_dic in parse() below.
+ layout_dic: dictionary to store tile layouts. See parse() for descriptions.
+ """
+
+ for config in config_re_compiled.finditer(line):
+ current_config = config.group(1)
+ tile_layout = ''
+ if is_per_tile: # per-tile bench, add name prefix
+ current_config = 'tile_' + current_config
+ layouts = TILE_LAYOUT_RE_COMPILED.search(line)
+ if layouts and len(layouts.groups()) == 2:
+ tile_layout = '%sx%s' % layouts.groups()
+ times = config.group(2)
+ for new_time in TIME_RE_COMPILED.finditer(times):
+ current_time_type = new_time.group(1)
+ iters = [float(i) for i in
+ new_time.group(2).strip().split(',')]
+ value_dic.setdefault(bench, {}).setdefault(
+ current_config, {}).setdefault(current_time_type, []).append(
+ iters)
+ layout_dic.setdefault(bench, {}).setdefault(
+ current_config, {}).setdefault(current_time_type, tile_layout)
+
+def parse_skp_bench_data(directory, revision, rep, default_settings=None):
+ """Parses all the skp bench data in the given directory.
+
+ Args:
+ directory: string of path to input data directory.
+ revision: git hash revision that matches the data to process.
+ rep: bench representation algorithm, see bench_util.py.
+ default_settings: dictionary of other run settings. See writer.option() in
+ bench/benchmain.cpp.
+
+ Returns:
+ A list of BenchDataPoint objects.
+ """
+ revision_data_points = []
+ file_list = os.listdir(directory)
+ file_list.sort()
+ for bench_file in file_list:
+ scalar_type = None
+ # Scalar type, if any, is in the bench filename after 'scalar_'.
+ if (bench_file.startswith('bench_' + revision + '_data_')):
+ if bench_file.find('scalar_') > 0:
+ components = bench_file.split('_')
+ scalar_type = components[components.index('scalar') + 1]
+ else: # Skips non skp bench files.
+ continue
+
+ with open('/'.join([directory, bench_file]), 'r') as file_handle:
+ settings = dict(default_settings or {})
+ settings['scalar'] = scalar_type
+ revision_data_points.extend(parse(settings, file_handle, rep))
+
+ return revision_data_points
+
+# TODO(bensong): switch to reading JSON output when available. This way we don't
+# need the RE complexities.
+def parse(settings, lines, representation=None):
+ """Parses bench output into a useful data structure.
+
+ ({str:str}, __iter__ -> str) -> [BenchDataPoint]
+ representation is one of the ALGORITHM_XXX types."""
+
+ benches = []
+ current_bench = None
+ # [bench][config][time_type] -> [[per-iter values]] where per-tile config
+ # has per-iter value list for each tile [[<tile1_iter1>,<tile1_iter2>,...],
+ # [<tile2_iter1>,<tile2_iter2>,...],...], while non-per-tile config only
+ # contains one list of iterations [[iter1, iter2, ...]].
+ bench_dic = {}
+ # [bench][config][time_type] -> tile_layout
+ layout_dic = {}
+
+ for line in lines:
+
+ # see if this line is a settings line
+ settingsMatch = SETTINGS_RE_COMPILED.search(line)
+ if (settingsMatch):
+ settings = dict(settings)
+ for settingMatch in PER_SETTING_RE_COMPILED.finditer(settingsMatch.group(1)):
+ if (settingMatch.group(2)):
+ settings[settingMatch.group(1)] = settingMatch.group(2)
+ else:
+ settings[settingMatch.group(1)] = True
+
+ # see if this line starts a new bench
+ new_bench = BENCH_RE_COMPILED.search(line)
+ if new_bench:
+ current_bench = new_bench.group(1)
+
+ # add configs on this line to the bench_dic
+ if current_bench:
+ if line.startswith(' tile_') :
+ _ParseAndStoreTimes(TILE_RE_COMPILED, True, line, current_bench,
+ bench_dic, layout_dic)
+ else:
+ _ParseAndStoreTimes(CONFIG_RE_COMPILED, False, line,
+ current_bench, bench_dic, layout_dic)
+
+ # append benches to list
+ for bench in bench_dic:
+ for config in bench_dic[bench]:
+ for time_type in bench_dic[bench][config]:
+ tile_layout = ''
+ per_tile_values = [] # empty for non-per-tile configs
+ per_iter_time = [] # empty for per-tile configs
+ bench_summary = None # a single final bench value
+ if len(bench_dic[bench][config][time_type]) > 1:
+ # per-tile config; compute representation for each tile
+ per_tile_values = [
+ _ListAlgorithm(iters, representation).compute()
+ for iters in bench_dic[bench][config][time_type]]
+ # use sum of each tile representation for total bench value
+ bench_summary = sum(per_tile_values)
+ # extract tile layout
+ tile_layout = layout_dic[bench][config][time_type]
+ else:
+ # get the list of per-iteration values
+ per_iter_time = bench_dic[bench][config][time_type][0]
+ bench_summary = _ListAlgorithm(
+ per_iter_time, representation).compute()
+ benches.append(BenchDataPoint(
+ bench,
+ config,
+ time_type,
+ bench_summary,
+ settings,
+ tile_layout,
+ per_tile_values,
+ per_iter_time))
+
+ return benches
+
+class LinearRegression:
+ """Linear regression data based on a set of data points.
+
+ ([(Number,Number)])
+ There must be at least two points for this to make sense."""
+ def __init__(self, points):
+ n = len(points)
+ max_x = Min
+ min_x = Max
+
+ Sx = 0.0
+ Sy = 0.0
+ Sxx = 0.0
+ Sxy = 0.0
+ Syy = 0.0
+ for point in points:
+ x = point[0]
+ y = point[1]
+ max_x = max(max_x, x)
+ min_x = min(min_x, x)
+
+ Sx += x
+ Sy += y
+ Sxx += x*x
+ Sxy += x*y
+ Syy += y*y
+
+ denom = n*Sxx - Sx*Sx
+ if (denom != 0.0):
+ B = (n*Sxy - Sx*Sy) / denom
+ else:
+ B = 0.0
+ a = (1.0/n)*(Sy - B*Sx)
+
+ se2 = 0
+ sB2 = 0
+ sa2 = 0
+ if (n >= 3 and denom != 0.0):
+ se2 = (1.0/(n*(n-2)) * (n*Syy - Sy*Sy - B*B*denom))
+ sB2 = (n*se2) / denom
+ sa2 = sB2 * (1.0/n) * Sxx
+
+
+ self.slope = B
+ self.intercept = a
+ self.serror = math.sqrt(max(0, se2))
+ self.serror_slope = math.sqrt(max(0, sB2))
+ self.serror_intercept = math.sqrt(max(0, sa2))
+ self.max_x = max_x
+ self.min_x = min_x
+
+ def __repr__(self):
+ return "LinearRegression(%s, %s, %s, %s, %s)" % (
+ str(self.slope),
+ str(self.intercept),
+ str(self.serror),
+ str(self.serror_slope),
+ str(self.serror_intercept),
+ )
+
+ def find_min_slope(self):
+ """Finds the minimal slope given one standard deviation."""
+ slope = self.slope
+ intercept = self.intercept
+ error = self.serror
+ regr_start = self.min_x
+ regr_end = self.max_x
+ regr_width = regr_end - regr_start
+
+ if slope < 0:
+ lower_left_y = slope*regr_start + intercept - error
+ upper_right_y = slope*regr_end + intercept + error
+ return min(0, (upper_right_y - lower_left_y) / regr_width)
+
+ elif slope > 0:
+ upper_left_y = slope*regr_start + intercept + error
+ lower_right_y = slope*regr_end + intercept - error
+ return max(0, (lower_right_y - upper_left_y) / regr_width)
+
+ return 0
+
+def CreateRevisionLink(revision_number):
+ """Returns HTML displaying the given revision number and linking to
+ that revision's change page at code.google.com, e.g.
+ http://code.google.com/p/skia/source/detail?r=2056
+ """
+ return '<a href="http://code.google.com/p/skia/source/detail?r=%s">%s</a>'%(
+ revision_number, revision_number)
+
+def main():
+ foo = [[0.0, 0.0], [0.0, 1.0], [0.0, 2.0], [0.0, 3.0]]
+ LinearRegression(foo)
+
+if __name__ == "__main__":
+ main()