aboutsummaryrefslogtreecommitdiffhomepage
path: root/tools/rebaseline_imagefiles.py
diff options
context:
space:
mode:
authorGravatar epoger@google.com <epoger@google.com@2bbb7eff-a529-9590-31e7-b0007b416f81>2013-06-19 18:56:59 +0000
committerGravatar epoger@google.com <epoger@google.com@2bbb7eff-a529-9590-31e7-b0007b416f81>2013-06-19 18:56:59 +0000
commit99a8ec97058a9224bdef434d1410dd5c24bffec6 (patch)
treefc4a0d71660b9d351aa7d781ea5e9a6a4e319d94 /tools/rebaseline_imagefiles.py
parent27d302581a09d55cd8b2e64e42b23fceca66e33a (diff)
rebaseline.py: split image-based rebaselining, which will go away soon, into its own script
(part of step 3 in https://goto.google.com/ChecksumTransitionDetail ) also adds new --expectations-root option R=scroggo@google.com, senorblanco@chromium.org Review URL: https://codereview.chromium.org/17379004 git-svn-id: http://skia.googlecode.com/svn/trunk@9689 2bbb7eff-a529-9590-31e7-b0007b416f81
Diffstat (limited to 'tools/rebaseline_imagefiles.py')
-rwxr-xr-xtools/rebaseline_imagefiles.py341
1 files changed, 341 insertions, 0 deletions
diff --git a/tools/rebaseline_imagefiles.py b/tools/rebaseline_imagefiles.py
new file mode 100755
index 0000000000..6c60745690
--- /dev/null
+++ b/tools/rebaseline_imagefiles.py
@@ -0,0 +1,341 @@
+#!/usr/bin/python
+
+'''
+Copyright 2013 Google Inc.
+
+Use of this source code is governed by a BSD-style license that can be
+found in the LICENSE file.
+'''
+
+'''
+Rebaselines GM test results as individual image files
+(the "old way", before https://goto.google.com/ChecksumTransitionDetail ).
+
+Once we have switched our expectations to JSON form for all platforms,
+we can delete this file.
+
+There is a lot of code duplicated between here and rebaseline.py, but
+that's fine because we will delete this file soon.
+'''
+
+# System-level imports
+import os
+import re
+import subprocess
+import sys
+import urllib2
+
+# Imports from within Skia
+#
+# We need to add the 'gm' directory, so that we can import gm_json.py within
+# that directory. That script allows us to parse the actual-results.json file
+# written out by the GM tool.
+# Make sure that the 'gm' dir is in the PYTHONPATH, but add it at the *end*
+# so any dirs that are already in the PYTHONPATH will be preferred.
+#
+# This assumes that the 'gm' directory has been checked out as a sibling of
+# the 'tools' directory containing this script, which will be the case if
+# 'trunk' was checked out as a single unit.
+GM_DIRECTORY = os.path.realpath(
+ os.path.join(os.path.dirname(os.path.dirname(__file__)), 'gm'))
+if GM_DIRECTORY not in sys.path:
+ sys.path.append(GM_DIRECTORY)
+import gm_json
+
+
+class CommandFailedException(Exception):
+ pass
+
+class ImageRebaseliner(object):
+
+ # params:
+ # expectations_root: root directory of all expectations
+ # json_base_url: base URL from which to read json_filename
+ # json_filename: filename (under json_base_url) from which to read a
+ # summary of results; typically "actual-results.json"
+ # tests: list of tests to rebaseline, or None if we should rebaseline
+ # whatever files the JSON results summary file tells us to
+ # configs: which configs to run for each test; this should only be
+ # specified if the list of tests was also specified (otherwise,
+ # the JSON file will give us test names and configs)
+ # dry_run: if True, instead of actually downloading files or adding
+ # files to checkout, display a list of operations that
+ # we would normally perform
+ # add_new: if True, add expectations for tests which don't have any yet
+ # missing_json_is_fatal: whether to halt execution if we cannot read a
+ # JSON actual result summary file
+ def __init__(self, expectations_root, json_base_url, json_filename,
+ tests=None, configs=None, dry_run=False,
+ add_new=False, missing_json_is_fatal=False):
+ if configs and not tests:
+ raise ValueError('configs should only be specified if tests ' +
+ 'were specified also')
+ self._expectations_root = expectations_root
+ self._tests = tests
+ self._configs = configs
+ self._json_base_url = json_base_url
+ self._json_filename = json_filename
+ self._dry_run = dry_run
+ self._add_new = add_new
+ self._missing_json_is_fatal = missing_json_is_fatal
+ self._googlestorage_gm_actuals_root = (
+ 'http://chromium-skia-gm.commondatastorage.googleapis.com/gm')
+ self._testname_pattern = re.compile('(\S+)_(\S+).png')
+ self._is_svn_checkout = (
+ os.path.exists('.svn') or
+ os.path.exists(os.path.join(os.pardir, '.svn')))
+ self._is_git_checkout = (
+ os.path.exists('.git') or
+ os.path.exists(os.path.join(os.pardir, '.git')))
+
+ # If dry_run is False, execute subprocess.call(cmd).
+ # If dry_run is True, print the command we would have otherwise run.
+ # Raises a CommandFailedException if the command fails.
+ def _Call(self, cmd):
+ if self._dry_run:
+ print '%s' % ' '.join(cmd)
+ return
+ if subprocess.call(cmd) != 0:
+ raise CommandFailedException('error running command: ' +
+ ' '.join(cmd))
+
+ # Download a single actual result from GoogleStorage, returning True if it
+ # succeeded.
+ def _DownloadFromGoogleStorage(self, infilename, outfilename, all_results):
+ test_name = self._testname_pattern.match(infilename).group(1)
+ if not test_name:
+ print '# unable to find test_name for infilename %s' % infilename
+ return False
+ try:
+ hash_type, hash_value = all_results[infilename]
+ except KeyError:
+ print ('# unable to find filename %s in all_results dict' %
+ infilename)
+ return False
+ except ValueError as e:
+ print '# ValueError reading filename %s from all_results dict: %s'%(
+ infilename, e)
+ return False
+ url = '%s/%s/%s/%s.png' % (self._googlestorage_gm_actuals_root,
+ hash_type, test_name, hash_value)
+ try:
+ self._DownloadFile(source_url=url, dest_filename=outfilename)
+ return True
+ except CommandFailedException:
+ print '# Couldn\'t fetch gs_url %s' % url
+ return False
+
+ # Download a single actual result from skia-autogen, returning True if it
+ # succeeded.
+ def _DownloadFromAutogen(self, infilename, outfilename,
+ expectations_subdir, builder_name):
+ url = ('http://skia-autogen.googlecode.com/svn/gm-actual/' +
+ expectations_subdir + '/' + builder_name + '/' +
+ expectations_subdir + '/' + infilename)
+ try:
+ self._DownloadFile(source_url=url, dest_filename=outfilename)
+ return True
+ except CommandFailedException:
+ print '# Couldn\'t fetch autogen_url %s' % url
+ return False
+
+ # Download a single file, raising a CommandFailedException if it fails.
+ def _DownloadFile(self, source_url, dest_filename):
+ # Download into a temporary file and then rename it afterwards,
+ # so that we don't corrupt the existing file if it fails midway thru.
+ temp_filename = os.path.join(os.path.dirname(dest_filename),
+ '.temp-' + os.path.basename(dest_filename))
+
+ # TODO(epoger): Replace calls to "curl"/"mv" (which will only work on
+ # Unix) with a Python HTTP library (which should work cross-platform)
+ self._Call([ 'curl', '--fail', '--silent', source_url,
+ '--output', temp_filename ])
+ self._Call([ 'mv', temp_filename, dest_filename ])
+
+ # Returns the full contents of a URL, as a single string.
+ #
+ # Unlike standard URL handling, we allow relative "file:" URLs;
+ # for example, "file:one/two" resolves to the file ./one/two
+ # (relative to current working dir)
+ def _GetContentsOfUrl(self, url):
+ file_prefix = 'file:'
+ if url.startswith(file_prefix):
+ filename = url[len(file_prefix):]
+ return open(filename, 'r').read()
+ else:
+ return urllib2.urlopen(url).read()
+
+ # Returns a dictionary of actual results from actual-results.json file.
+ #
+ # The dictionary returned has this format:
+ # {
+ # u'imageblur_565.png': [u'bitmap-64bitMD5', 3359963596899141322],
+ # u'imageblur_8888.png': [u'bitmap-64bitMD5', 4217923806027861152],
+ # u'shadertext3_8888.png': [u'bitmap-64bitMD5', 3713708307125704716]
+ # }
+ #
+ # If the JSON actual result summary file cannot be loaded, the behavior
+ # depends on self._missing_json_is_fatal:
+ # - if true: execution will halt with an exception
+ # - if false: we will log an error message but return an empty dictionary
+ #
+ # params:
+ # json_url: URL pointing to a JSON actual result summary file
+ # sections: a list of section names to include in the results, e.g.
+ # [gm_json.JSONKEY_ACTUALRESULTS_FAILED,
+ # gm_json.JSONKEY_ACTUALRESULTS_NOCOMPARISON] ;
+ # if None, then include ALL sections.
+ def _GetActualResults(self, json_url, sections=None):
+ try:
+ json_contents = self._GetContentsOfUrl(json_url)
+ except (urllib2.HTTPError, IOError):
+ message = 'unable to load JSON summary URL %s' % json_url
+ if self._missing_json_is_fatal:
+ raise ValueError(message)
+ else:
+ print '# %s' % message
+ return {}
+
+ json_dict = gm_json.LoadFromString(json_contents)
+ results_to_return = {}
+ actual_results = json_dict[gm_json.JSONKEY_ACTUALRESULTS]
+ if not sections:
+ sections = actual_results.keys()
+ for section in sections:
+ section_results = actual_results[section]
+ if section_results:
+ results_to_return.update(section_results)
+ return results_to_return
+
+ # Returns a list of files that require rebaselining.
+ #
+ # Note that this returns a list of FILES, like this:
+ # ['imageblur_565.png', 'xfermodes_pdf.png']
+ # rather than a list of TESTS, like this:
+ # ['imageblur', 'xfermodes']
+ #
+ # params:
+ # json_url: URL pointing to a JSON actual result summary file
+ # add_new: if True, then return files listed in any of these sections:
+ # - JSONKEY_ACTUALRESULTS_FAILED
+ # - JSONKEY_ACTUALRESULTS_NOCOMPARISON
+ # if False, then return files listed in these sections:
+ # - JSONKEY_ACTUALRESULTS_FAILED
+ #
+ def _GetFilesToRebaseline(self, json_url, add_new):
+ if self._dry_run:
+ print ''
+ print '#'
+ print ('# Getting files to rebaseline from JSON summary URL %s ...'
+ % json_url)
+ sections = [gm_json.JSONKEY_ACTUALRESULTS_FAILED]
+ if add_new:
+ sections.append(gm_json.JSONKEY_ACTUALRESULTS_NOCOMPARISON)
+ results_to_rebaseline = self._GetActualResults(json_url=json_url,
+ sections=sections)
+ files_to_rebaseline = results_to_rebaseline.keys()
+ files_to_rebaseline.sort()
+ print '# ... found files_to_rebaseline %s' % files_to_rebaseline
+ if self._dry_run:
+ print '#'
+ return files_to_rebaseline
+
+ # Rebaseline a single file.
+ def _RebaselineOneFile(self, expectations_subdir, builder_name,
+ infilename, outfilename, all_results):
+ if self._dry_run:
+ print ''
+ print '# ' + infilename
+
+ # First try to download this result image from Google Storage.
+ # If that fails, try skia-autogen.
+ # If that fails too, just go on to the next file.
+ #
+ # This not treated as a fatal failure because not all
+ # platforms generate all configs (e.g., Android does not
+ # generate PDF).
+ #
+ # TODO(epoger): Once we are downloading only files that the
+ # actual-results.json file told us to, this should become a
+ # fatal error. (If the actual-results.json file told us that
+ # the test failed with XXX results, we should be able to download
+ # those results every time.)
+ if not self._DownloadFromGoogleStorage(infilename=infilename,
+ outfilename=outfilename,
+ all_results=all_results):
+ if not self._DownloadFromAutogen(infilename=infilename,
+ outfilename=outfilename,
+ expectations_subdir=expectations_subdir,
+ builder_name=builder_name):
+ print '# Couldn\'t fetch infilename ' + infilename
+ return
+
+ # Add this file to version control (if appropriate).
+ if self._add_new:
+ if self._is_svn_checkout:
+ cmd = [ 'svn', 'add', '--quiet', outfilename ]
+ self._Call(cmd)
+ cmd = [ 'svn', 'propset', '--quiet', 'svn:mime-type',
+ 'image/png', outfilename ];
+ self._Call(cmd)
+ elif self._is_git_checkout:
+ cmd = [ 'git', 'add', outfilename ]
+ self._Call(cmd)
+
+ # Rebaseline the given configs for a single test.
+ #
+ # params:
+ # expectations_subdir
+ # builder_name
+ # test: a single test to rebaseline
+ # all_results: a dictionary of all actual results
+ def _RebaselineOneTest(self, expectations_subdir, builder_name, test,
+ all_results):
+ if self._configs:
+ configs = self._configs
+ else:
+ if (expectations_subdir == 'base-shuttle-win7-intel-angle'):
+ configs = [ 'angle', 'anglemsaa16' ]
+ else:
+ configs = [ '565', '8888', 'gpu', 'pdf', 'mesa', 'msaa16',
+ 'msaa4' ]
+ if self._dry_run:
+ print ''
+ print '# ' + expectations_subdir + ':'
+ for config in configs:
+ infilename = test + '_' + config + '.png'
+ outfilename = os.path.join(self._expectations_root,
+ expectations_subdir, infilename);
+ self._RebaselineOneFile(expectations_subdir=expectations_subdir,
+ builder_name=builder_name,
+ infilename=infilename,
+ outfilename=outfilename,
+ all_results=all_results)
+
+ # Rebaseline all tests/types we specified in the constructor,
+ # within this gm-expectations subdir.
+ #
+ # params:
+ # subdir : e.g. 'base-shuttle-win7-intel-float'
+ # builder : e.g. 'Test-Win7-ShuttleA-HD2000-x86-Release'
+ def RebaselineSubdir(self, subdir, builder):
+ json_url = '/'.join([self._json_base_url,
+ subdir, builder, subdir,
+ self._json_filename])
+ all_results = self._GetActualResults(json_url=json_url)
+
+ if self._tests:
+ for test in self._tests:
+ self._RebaselineOneTest(expectations_subdir=subdir,
+ builder_name=builder,
+ test=test, all_results=all_results)
+ else: # get the raw list of files that need rebaselining from JSON
+ filenames = self._GetFilesToRebaseline(json_url=json_url,
+ add_new=self._add_new)
+ for filename in filenames:
+ outfilename = os.path.join(subdir, filename);
+ self._RebaselineOneFile(expectations_subdir=subdir,
+ builder_name=builder,
+ infilename=filename,
+ outfilename=outfilename,
+ all_results=all_results)