#!/usr/bin/python ''' Copyright 2012 Google Inc. Use of this source code is governed by a BSD-style license that can be found in the LICENSE file. ''' ''' Rebaselines the given GM tests, on all bots and all configurations. Must be run from the gm-expected directory. If run from a git or SVN checkout, the files will be added to the staging area for commit. ''' # System-level imports import argparse import os import re import subprocess import sys import urllib2 # Imports from local directory import rebaseline_imagefiles # Imports from within Skia # # We need to add the 'gm' directory, so that we can import gm_json.py within # that directory. That script allows us to parse the actual-results.json file # written out by the GM tool. # Make sure that the 'gm' dir is in the PYTHONPATH, but add it at the *end* # so any dirs that are already in the PYTHONPATH will be preferred. # # This assumes that the 'gm' directory has been checked out as a sibling of # the 'tools' directory containing this script, which will be the case if # 'trunk' was checked out as a single unit. GM_DIRECTORY = os.path.realpath( os.path.join(os.path.dirname(os.path.dirname(__file__)), 'gm')) if GM_DIRECTORY not in sys.path: sys.path.append(GM_DIRECTORY) import gm_json JSON_EXPECTATIONS_FILENAME='expected-results.json' # Mapping of gm-expectations subdir (under # https://skia.googlecode.com/svn/gm-expected/ ) # to builder name (see list at http://108.170.217.252:10117/builders ) SUBDIR_MAPPING = { 'base-shuttle-win7-intel-float': 'Test-Win7-ShuttleA-HD2000-x86-Release', 'base-shuttle-win7-intel-angle': 'Test-Win7-ShuttleA-HD2000-x86-Release-ANGLE', 'base-shuttle-win7-intel-directwrite': 'Test-Win7-ShuttleA-HD2000-x86-Release-DirectWrite', 'base-shuttle_ubuntu12_ati5770': 'Test-Ubuntu12-ShuttleA-ATI5770-x86_64-Release', 'base-macmini': 'Test-Mac10.6-MacMini4.1-GeForce320M-x86-Release', 'base-macmini-lion-float': 'Test-Mac10.7-MacMini4.1-GeForce320M-x86-Release', 'base-android-galaxy-nexus': 'Test-Android-GalaxyNexus-SGX540-Arm7-Debug', 'base-android-nexus-7': 'Test-Android-Nexus7-Tegra3-Arm7-Release', 'base-android-nexus-s': 'Test-Android-NexusS-SGX540-Arm7-Release', 'base-android-xoom': 'Test-Android-Xoom-Tegra2-Arm7-Release', 'base-android-nexus-10': 'Test-Android-Nexus10-MaliT604-Arm7-Release', } class CommandFailedException(Exception): pass # Object that rebaselines a JSON expectations file (not individual image files). # # TODO(epoger): Most of this is just the code from the old ImageRebaseliner... # some of it will need to be updated in order to properly rebaseline JSON files. # There is a lot of code duplicated between here and ImageRebaseliner, but # that's fine because we will delete ImageRebaseliner soon. class JsonRebaseliner(object): # params: # expectations_root: root directory of all expectations # json_base_url: base URL from which to read json_filename # json_filename: filename (under json_base_url) from which to read a # summary of results; typically "actual-results.json" # tests: list of tests to rebaseline, or None if we should rebaseline # whatever files the JSON results summary file tells us to # configs: which configs to run for each test; this should only be # specified if the list of tests was also specified (otherwise, # the JSON file will give us test names and configs) # dry_run: if True, instead of actually downloading files or adding # files to checkout, display a list of operations that # we would normally perform # add_new: if True, add expectations for tests which don't have any yet # missing_json_is_fatal: whether to halt execution if we cannot read a # JSON actual result summary file def __init__(self, expectations_root, json_base_url, json_filename, tests=None, configs=None, dry_run=False, add_new=False, missing_json_is_fatal=False): raise ValueError('JsonRebaseliner not yet implemented') # TODO(epoger) if configs and not tests: raise ValueError('configs should only be specified if tests ' + 'were specified also') self._expectations_root = expectations_root self._tests = tests self._configs = configs self._json_base_url = json_base_url self._json_filename = json_filename self._dry_run = dry_run self._add_new = add_new self._missing_json_is_fatal = missing_json_is_fatal self._googlestorage_gm_actuals_root = ( 'http://chromium-skia-gm.commondatastorage.googleapis.com/gm') self._testname_pattern = re.compile('(\S+)_(\S+).png') self._is_svn_checkout = ( os.path.exists('.svn') or os.path.exists(os.path.join(os.pardir, '.svn'))) self._is_git_checkout = ( os.path.exists('.git') or os.path.exists(os.path.join(os.pardir, '.git'))) # If dry_run is False, execute subprocess.call(cmd). # If dry_run is True, print the command we would have otherwise run. # Raises a CommandFailedException if the command fails. def _Call(self, cmd): if self._dry_run: print '%s' % ' '.join(cmd) return if subprocess.call(cmd) != 0: raise CommandFailedException('error running command: ' + ' '.join(cmd)) # Download a single actual result from GoogleStorage, returning True if it # succeeded. def _DownloadFromGoogleStorage(self, infilename, outfilename, all_results): test_name = self._testname_pattern.match(infilename).group(1) if not test_name: print '# unable to find test_name for infilename %s' % infilename return False try: hash_type, hash_value = all_results[infilename] except KeyError: print ('# unable to find filename %s in all_results dict' % infilename) return False except ValueError as e: print '# ValueError reading filename %s from all_results dict: %s'%( infilename, e) return False url = '%s/%s/%s/%s.png' % (self._googlestorage_gm_actuals_root, hash_type, test_name, hash_value) try: self._DownloadFile(source_url=url, dest_filename=outfilename) return True except CommandFailedException: print '# Couldn\'t fetch gs_url %s' % url return False # Download a single actual result from skia-autogen, returning True if it # succeeded. def _DownloadFromAutogen(self, infilename, outfilename, expectations_subdir, builder_name): url = ('http://skia-autogen.googlecode.com/svn/gm-actual/' + expectations_subdir + '/' + builder_name + '/' + expectations_subdir + '/' + infilename) try: self._DownloadFile(source_url=url, dest_filename=outfilename) return True except CommandFailedException: print '# Couldn\'t fetch autogen_url %s' % url return False # Download a single file, raising a CommandFailedException if it fails. def _DownloadFile(self, source_url, dest_filename): # Download into a temporary file and then rename it afterwards, # so that we don't corrupt the existing file if it fails midway thru. temp_filename = os.path.join(os.path.dirname(dest_filename), '.temp-' + os.path.basename(dest_filename)) # TODO(epoger): Replace calls to "curl"/"mv" (which will only work on # Unix) with a Python HTTP library (which should work cross-platform) self._Call([ 'curl', '--fail', '--silent', source_url, '--output', temp_filename ]) self._Call([ 'mv', temp_filename, dest_filename ]) # Returns the full contents of a URL, as a single string. # # Unlike standard URL handling, we allow relative "file:" URLs; # for example, "file:one/two" resolves to the file ./one/two # (relative to current working dir) def _GetContentsOfUrl(self, url): file_prefix = 'file:' if url.startswith(file_prefix): filename = url[len(file_prefix):] return open(filename, 'r').read() else: return urllib2.urlopen(url).read() # Returns a dictionary of actual results from actual-results.json file. # # The dictionary returned has this format: # { # u'imageblur_565.png': [u'bitmap-64bitMD5', 3359963596899141322], # u'imageblur_8888.png': [u'bitmap-64bitMD5', 4217923806027861152], # u'shadertext3_8888.png': [u'bitmap-64bitMD5', 3713708307125704716] # } # # If the JSON actual result summary file cannot be loaded, the behavior # depends on self._missing_json_is_fatal: # - if true: execution will halt with an exception # - if false: we will log an error message but return an empty dictionary # # params: # json_url: URL pointing to a JSON actual result summary file # sections: a list of section names to include in the results, e.g. # [gm_json.JSONKEY_ACTUALRESULTS_FAILED, # gm_json.JSONKEY_ACTUALRESULTS_NOCOMPARISON] ; # if None, then include ALL sections. def _GetActualResults(self, json_url, sections=None): try: json_contents = self._GetContentsOfUrl(json_url) except (urllib2.HTTPError, IOError): message = 'unable to load JSON summary URL %s' % json_url if self._missing_json_is_fatal: raise ValueError(message) else: print '# %s' % message return {} json_dict = gm_json.LoadFromString(json_contents) results_to_return = {} actual_results = json_dict[gm_json.JSONKEY_ACTUALRESULTS] if not sections: sections = actual_results.keys() for section in sections: section_results = actual_results[section] if section_results: results_to_return.update(section_results) return results_to_return # Returns a list of files that require rebaselining. # # Note that this returns a list of FILES, like this: # ['imageblur_565.png', 'xfermodes_pdf.png'] # rather than a list of TESTS, like this: # ['imageblur', 'xfermodes'] # # params: # json_url: URL pointing to a JSON actual result summary file # add_new: if True, then return files listed in any of these sections: # - JSONKEY_ACTUALRESULTS_FAILED # - JSONKEY_ACTUALRESULTS_NOCOMPARISON # if False, then return files listed in these sections: # - JSONKEY_ACTUALRESULTS_FAILED # def _GetFilesToRebaseline(self, json_url, add_new): if self._dry_run: print '' print '#' print ('# Getting files to rebaseline from JSON summary URL %s ...' % json_url) sections = [gm_json.JSONKEY_ACTUALRESULTS_FAILED] if add_new: sections.append(gm_json.JSONKEY_ACTUALRESULTS_NOCOMPARISON) results_to_rebaseline = self._GetActualResults(json_url=json_url, sections=sections) files_to_rebaseline = results_to_rebaseline.keys() files_to_rebaseline.sort() print '# ... found files_to_rebaseline %s' % files_to_rebaseline if self._dry_run: print '#' return files_to_rebaseline # Rebaseline a single file. def _RebaselineOneFile(self, expectations_subdir, builder_name, infilename, outfilename, all_results): if self._dry_run: print '' print '# ' + infilename # First try to download this result image from Google Storage. # If that fails, try skia-autogen. # If that fails too, just go on to the next file. # # This not treated as a fatal failure because not all # platforms generate all configs (e.g., Android does not # generate PDF). # # TODO(epoger): Once we are downloading only files that the # actual-results.json file told us to, this should become a # fatal error. (If the actual-results.json file told us that # the test failed with XXX results, we should be able to download # those results every time.) if not self._DownloadFromGoogleStorage(infilename=infilename, outfilename=outfilename, all_results=all_results): if not self._DownloadFromAutogen(infilename=infilename, outfilename=outfilename, expectations_subdir=expectations_subdir, builder_name=builder_name): print '# Couldn\'t fetch infilename ' + infilename return # Add this file to version control (if appropriate). if self._add_new: if self._is_svn_checkout: cmd = [ 'svn', 'add', '--quiet', outfilename ] self._Call(cmd) cmd = [ 'svn', 'propset', '--quiet', 'svn:mime-type', 'image/png', outfilename ]; self._Call(cmd) elif self._is_git_checkout: cmd = [ 'git', 'add', outfilename ] self._Call(cmd) # Rebaseline the given configs for a single test. # # params: # expectations_subdir # builder_name # test: a single test to rebaseline # all_results: a dictionary of all actual results def _RebaselineOneTest(self, expectations_subdir, builder_name, test, all_results): if self._configs: configs = self._configs else: if (expectations_subdir == 'base-shuttle-win7-intel-angle'): configs = [ 'angle', 'anglemsaa16' ] else: configs = [ '565', '8888', 'gpu', 'pdf', 'mesa', 'msaa16', 'msaa4' ] if self._dry_run: print '' print '# ' + expectations_subdir + ':' for config in configs: infilename = test + '_' + config + '.png' outfilename = os.path.join(expectations_subdir, infilename); self._RebaselineOneFile(expectations_subdir=expectations_subdir, builder_name=builder_name, infilename=infilename, outfilename=outfilename, all_results=all_results) # Rebaseline all tests/types we specified in the constructor, # within this gm-expectations subdir. # # params: # subdir : e.g. 'base-shuttle-win7-intel-float' # builder : e.g. 'Test-Win7-ShuttleA-HD2000-x86-Release' def RebaselineSubdir(self, subdir, builder): json_url = '/'.join([self._json_base_url, subdir, builder, subdir, self._json_filename]) all_results = self._GetActualResults(json_url=json_url) if self._tests: for test in self._tests: self._RebaselineOneTest(expectations_subdir=subdir, builder_name=builder, test=test, all_results=all_results) else: # get the raw list of files that need rebaselining from JSON filenames = self._GetFilesToRebaseline(json_url=json_url, add_new=self._add_new) for filename in filenames: outfilename = os.path.join(subdir, filename); self._RebaselineOneFile(expectations_subdir=subdir, builder_name=builder, infilename=filename, outfilename=outfilename, all_results=all_results) # main... parser = argparse.ArgumentParser() parser.add_argument('--add-new', action='store_true', help='in addition to the standard behavior of ' + 'updating expectations for failing tests, add ' + 'expectations for tests which don\'t have expectations ' + 'yet.') parser.add_argument('--configs', metavar='CONFIG', nargs='+', help='which configurations to rebaseline, e.g. ' + '"--configs 565 8888"; if unspecified, run a default ' + 'set of configs. This should ONLY be specified if ' + '--tests has also been specified.') parser.add_argument('--dry-run', action='store_true', help='instead of actually downloading files or adding ' + 'files to checkout, display a list of operations that ' + 'we would normally perform') parser.add_argument('--expectations-root', help='root of expectations directory to update-- should ' + 'contain one or more base-* subdirectories. Defaults to ' + '%(default)s', default='.') parser.add_argument('--json-base-url', help='base URL from which to read JSON_FILENAME ' + 'files; defaults to %(default)s', default='http://skia-autogen.googlecode.com/svn/gm-actual') parser.add_argument('--json-filename', help='filename (under JSON_BASE_URL) to read a summary ' + 'of results from; defaults to %(default)s', default='actual-results.json') parser.add_argument('--subdirs', metavar='SUBDIR', nargs='+', help='which platform subdirectories to rebaseline; ' + 'if unspecified, rebaseline all subdirs, same as ' + '"--subdirs %s"' % ' '.join(sorted(SUBDIR_MAPPING.keys()))) parser.add_argument('--tests', metavar='TEST', nargs='+', help='which tests to rebaseline, e.g. ' + '"--tests aaclip bigmatrix"; if unspecified, then all ' + 'failing tests (according to the actual-results.json ' + 'file) will be rebaselined.') args = parser.parse_args() if args.subdirs: subdirs = args.subdirs missing_json_is_fatal = True else: subdirs = sorted(SUBDIR_MAPPING.keys()) missing_json_is_fatal = False for subdir in subdirs: if not subdir in SUBDIR_MAPPING.keys(): raise Exception(('unrecognized platform subdir "%s"; ' + 'should be one of %s') % ( subdir, SUBDIR_MAPPING.keys())) builder = SUBDIR_MAPPING[subdir] # We instantiate different Rebaseliner objects depending # on whether we are rebaselining an expected-results.json file, or # individual image files. Different gm-expected subdirectories may move # from individual image files to JSON-format expectations at different # times, so we need to make this determination per subdirectory. # # See https://goto.google.com/ChecksumTransitionDetail expectations_json_file = os.path.join(args.expectations_root, subdir, JSON_EXPECTATIONS_FILENAME) if os.path.isfile(expectations_json_file): sys.stderr.write('ERROR: JsonRebaseliner is not implemented yet.\n') sys.exit(1) rebaseliner = JsonRebaseliner( expectations_root=args.expectations_root, tests=args.tests, configs=args.configs, dry_run=args.dry_run, json_base_url=args.json_base_url, json_filename=args.json_filename, add_new=args.add_new, missing_json_is_fatal=missing_json_is_fatal) else: rebaseliner = rebaseline_imagefiles.ImageRebaseliner( expectations_root=args.expectations_root, tests=args.tests, configs=args.configs, dry_run=args.dry_run, json_base_url=args.json_base_url, json_filename=args.json_filename, add_new=args.add_new, missing_json_is_fatal=missing_json_is_fatal) rebaseliner.RebaselineSubdir(subdir=subdir, builder=builder)