#!/usr/bin/python """ Copyright 2013 Google Inc. Use of this source code is governed by a BSD-style license that can be found in the LICENSE file. Repackage expected/actual GM results as needed by our HTML rebaseline viewer. """ # System-level imports import fnmatch import os import re # Imports from within Skia import fix_pythonpath # must do this first import gm_json import imagepairset # Keys used to link an image to a particular GM test. # NOTE: Keep these in sync with static/constants.js VALUE__HEADER__SCHEMA_VERSION = 3 KEY__EXPECTATIONS__BUGS = gm_json.JSONKEY_EXPECTEDRESULTS_BUGS KEY__EXPECTATIONS__IGNOREFAILURE = gm_json.JSONKEY_EXPECTEDRESULTS_IGNOREFAILURE KEY__EXPECTATIONS__REVIEWED = gm_json.JSONKEY_EXPECTEDRESULTS_REVIEWED KEY__EXTRACOLUMNS__BUILDER = 'builder' KEY__EXTRACOLUMNS__CONFIG = 'config' KEY__EXTRACOLUMNS__RESULT_TYPE = 'resultType' KEY__EXTRACOLUMNS__TEST = 'test' KEY__HEADER__DATAHASH = 'dataHash' KEY__HEADER__IS_EDITABLE = 'isEditable' KEY__HEADER__IS_EXPORTED = 'isExported' KEY__HEADER__IS_STILL_LOADING = 'resultsStillLoading' KEY__HEADER__RESULTS_ALL = 'all' KEY__HEADER__RESULTS_FAILURES = 'failures' KEY__HEADER__SCHEMA_VERSION = 'schemaVersion' KEY__HEADER__TIME_NEXT_UPDATE_AVAILABLE = 'timeNextUpdateAvailable' KEY__HEADER__TIME_UPDATED = 'timeUpdated' KEY__HEADER__TYPE = 'type' KEY__RESULT_TYPE__FAILED = gm_json.JSONKEY_ACTUALRESULTS_FAILED KEY__RESULT_TYPE__FAILUREIGNORED = gm_json.JSONKEY_ACTUALRESULTS_FAILUREIGNORED KEY__RESULT_TYPE__NOCOMPARISON = gm_json.JSONKEY_ACTUALRESULTS_NOCOMPARISON KEY__RESULT_TYPE__SUCCEEDED = gm_json.JSONKEY_ACTUALRESULTS_SUCCEEDED IMAGE_FILENAME_RE = re.compile(gm_json.IMAGE_FILENAME_PATTERN) IMAGE_FILENAME_FORMATTER = '%s_%s.png' # pass in (testname, config) PARENT_DIRECTORY = os.path.dirname(os.path.realpath(__file__)) DEFAULT_ACTUALS_DIR = '.gm-actuals' DEFAULT_GENERATED_IMAGES_ROOT = os.path.join( PARENT_DIRECTORY, '.generated-images') # Define the default set of builders we will process expectations/actuals for. # This allows us to ignore builders for which we don't maintain expectations # (trybots, Valgrind, ASAN, TSAN), and avoid problems like # https://code.google.com/p/skia/issues/detail?id=2036 ('rebaseline_server # produces error when trying to add baselines for ASAN/TSAN builders') DEFAULT_MATCH_BUILDERS_PATTERN_LIST = ['.*'] DEFAULT_SKIP_BUILDERS_PATTERN_LIST = [ '.*-Trybot', '.*Valgrind.*', '.*TSAN.*', '.*ASAN.*'] class BaseComparisons(object): """Base class for generating summary of comparisons between two image sets. """ def get_results_of_type(self, results_type): """Return results of some/all tests (depending on 'results_type' parameter). Args: results_type: string describing which types of results to include; must be one of the RESULTS_* constants Results are returned in a dictionary as output by ImagePairSet.as_dict(). """ return self._results[results_type] def get_packaged_results_of_type(self, results_type, reload_seconds=None, is_editable=False, is_exported=True): """Package the results of some/all tests as a complete response_dict. Args: results_type: string indicating which set of results to return; must be one of the RESULTS_* constants reload_seconds: if specified, note that new results may be available once these results are reload_seconds old is_editable: whether clients are allowed to submit new baselines is_exported: whether these results are being made available to other network hosts """ response_dict = self._results[results_type] time_updated = self.get_timestamp() response_dict[imagepairset.KEY__ROOT__HEADER] = { KEY__HEADER__SCHEMA_VERSION: ( VALUE__HEADER__SCHEMA_VERSION), # Timestamps: # 1. when this data was last updated # 2. when the caller should check back for new data (if ever) KEY__HEADER__TIME_UPDATED: time_updated, KEY__HEADER__TIME_NEXT_UPDATE_AVAILABLE: ( (time_updated+reload_seconds) if reload_seconds else None), # The type we passed to get_results_of_type() KEY__HEADER__TYPE: results_type, # Hash of dataset, which the client must return with any edits-- # this ensures that the edits were made to a particular dataset. KEY__HEADER__DATAHASH: str(hash(repr( response_dict[imagepairset.KEY__ROOT__IMAGEPAIRS]))), # Whether the server will accept edits back. KEY__HEADER__IS_EDITABLE: is_editable, # Whether the service is accessible from other hosts. KEY__HEADER__IS_EXPORTED: is_exported, } return response_dict def get_timestamp(self): """Return the time at which this object was created, in seconds past epoch (UTC). """ return self._timestamp _match_builders_pattern_list = [ re.compile(p) for p in DEFAULT_MATCH_BUILDERS_PATTERN_LIST] _skip_builders_pattern_list = [ re.compile(p) for p in DEFAULT_SKIP_BUILDERS_PATTERN_LIST] def set_match_builders_pattern_list(self, pattern_list): """Override the default set of builders we should process. The default is DEFAULT_MATCH_BUILDERS_PATTERN_LIST . Note that skip_builders_pattern_list overrides this; regardless of whether a builder is in the "match" list, if it's in the "skip" list, we will skip it. Args: pattern_list: list of regex patterns; process builders that match any entry within this list """ if pattern_list == None: pattern_list = [] self._match_builders_pattern_list = [re.compile(p) for p in pattern_list] def set_skip_builders_pattern_list(self, pattern_list): """Override the default set of builders we should skip while processing. The default is DEFAULT_SKIP_BUILDERS_PATTERN_LIST . This overrides match_builders_pattern_list; regardless of whether a builder is in the "match" list, if it's in the "skip" list, we will skip it. Args: pattern_list: list of regex patterns; skip builders that match any entry within this list """ if pattern_list == None: pattern_list = [] self._skip_builders_pattern_list = [re.compile(p) for p in pattern_list] def _ignore_builder(self, builder): """Returns True if we should skip processing this builder. Args: builder: name of this builder, as a string Returns: True if we should ignore expectations and actuals for this builder. """ for pattern in self._skip_builders_pattern_list: if pattern.match(builder): return True for pattern in self._match_builders_pattern_list: if pattern.match(builder): return False return True def _read_builder_dicts_from_root(self, root, pattern='*.json'): """Read all JSON dictionaries within a directory tree. Skips any dictionaries belonging to a builder we have chosen to ignore. Args: root: path to root of directory tree pattern: which files to read within root (fnmatch-style pattern) Returns: A meta-dictionary containing all the JSON dictionaries found within the directory tree, keyed by builder name (the basename of the directory where each JSON dictionary was found). Raises: IOError if root does not refer to an existing directory """ # I considered making this call _read_dicts_from_root(), but I decided # it was better to prune out the ignored builders within the os.walk(). if not os.path.isdir(root): raise IOError('no directory found at path %s' % root) meta_dict = {} for dirpath, dirnames, filenames in os.walk(root): for matching_filename in fnmatch.filter(filenames, pattern): builder = os.path.basename(dirpath) if self._ignore_builder(builder): continue full_path = os.path.join(dirpath, matching_filename) meta_dict[builder] = gm_json.LoadFromFile(full_path) return meta_dict def _read_dicts_from_root(self, root, pattern='*.json'): """Read all JSON dictionaries within a directory tree. Args: root: path to root of directory tree pattern: which files to read within root (fnmatch-style pattern) Returns: A meta-dictionary containing all the JSON dictionaries found within the directory tree, keyed by the pathname (relative to root) of each JSON dictionary. Raises: IOError if root does not refer to an existing directory """ if not os.path.isdir(root): raise IOError('no directory found at path %s' % root) meta_dict = {} for abs_dirpath, dirnames, filenames in os.walk(root): rel_dirpath = os.path.relpath(abs_dirpath, root) for matching_filename in fnmatch.filter(filenames, pattern): abs_path = os.path.join(abs_dirpath, matching_filename) rel_path = os.path.join(rel_dirpath, matching_filename) meta_dict[rel_path] = gm_json.LoadFromFile(abs_path) return meta_dict @staticmethod def _read_noncomment_lines(path): """Return a list of all noncomment lines within a file. (A "noncomment" line is one that does not start with a '#'.) Args: path: path to file """ lines = [] with open(path, 'r') as fh: for line in fh: if not line.startswith('#'): lines.append(line.strip()) return lines @staticmethod def _create_relative_url(hashtype_and_digest, test_name): """Returns the URL for this image, relative to GM_ACTUALS_ROOT_HTTP_URL. If we don't have a record of this image, returns None. Args: hashtype_and_digest: (hash_type, hash_digest) tuple, or None if we don't have a record of this image test_name: string; name of the GM test that created this image """ if not hashtype_and_digest: return None return gm_json.CreateGmRelativeUrl( test_name=test_name, hash_type=hashtype_and_digest[0], hash_digest=hashtype_and_digest[1]) @staticmethod def combine_subdicts(input_dict): """ Flatten out a dictionary structure by one level. Input: { KEY_A1 : { KEY_B1 : VALUE_B1, }, KEY_A2 : { KEY_B2 : VALUE_B2, } } Output: { KEY_B1 : VALUE_B1, KEY_B2 : VALUE_B2, } If this would result in any repeated keys, it will raise an Exception. """ output_dict = {} for key, subdict in input_dict.iteritems(): for subdict_key, subdict_value in subdict.iteritems(): if subdict_key in output_dict: raise Exception('duplicate key %s in combine_subdicts' % subdict_key) output_dict[subdict_key] = subdict_value return output_dict @staticmethod def get_multilevel(input_dict, *keys): """ Returns input_dict[key1][key2][...], or None if any key is not found. """ for key in keys: if input_dict == None: return None input_dict = input_dict.get(key, None) return input_dict