blob: cd75f548f102abb452cc1bf13155af5d2fc9394f [file] [log] [blame]
# Copyright 2015 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import re
from recipe_engine.types import freeze
class V8TestingVariants(object):
"""Immutable class to manage the testing variant passed to v8.
There are several test-runner flags that determine the v8-side testing
variants to be used. This class manages passing those flags to the runner
and makes sure that only one such flag is passed.
Infra test configurations might specify the variants on different levels,
e.g. per test, per step or per builder. This class makes sure that the
most specific variants are used.
The variants have the following allowed transitions:
exhaustive variants -> no exhaustive variants -> one specific variant
"""
def __init__(self):
self.test_args = []
def __eq__(self, other):
assert isinstance(other, V8TestingVariants)
return self.test_args == other.test_args
def __add__(self, right):
"""Use + to specify variants with the more specific one on the right-hand
side.
"""
assert isinstance(right, V8TestingVariants)
return right._specify(self)
def _specify(self, previous): # pragma: no cover
raise NotImplementedError()
class V8ExhaustiveVariants(V8TestingVariants):
def __init__(self):
self.test_args = ['--exhaustive-variants']
def _specify(self, previous):
# Keep the previous as it's either exhaustive already or more specific.
return previous
class V8NoExhaustiveVariants(V8TestingVariants):
def _specify(self, previous):
# This is used to remove the default exhaustive variants on some bots.
if isinstance(previous, V8ExhaustiveVariants):
return self
else:
return previous
class V8Variant(V8TestingVariants):
def __init__(self, name):
self.test_args = ['--variants=' + name]
def _specify(self, previous):
# A specific variant cannot be replaced by a different one. E.g. if a
# builder is specified to run with the default it can't have a test step
# that runs ignition only.
assert not isinstance(previous, V8Variant) or self == previous
return self
class _V8VariantNeutral(V8TestingVariants):
"""Convenience null object to specify effectless default values."""
def _specify(self, previous):
return previous
V8VariantNeutral = _V8VariantNeutral()
TEST_CONFIGS = freeze({
'benchmarks': {
'name': 'Benchmarks',
'tests': ['benchmarks'],
'test_args': ['--download-data'],
},
'deopt': {
'tool': 'run-deopt-fuzzer',
'isolated_target': 'run-deopt-fuzzer',
},
'jsfunfuzz': {
'tool': 'jsfunfuzz',
'isolated_target': 'jsfunfuzz',
},
'gcmole': {
'tool': 'run-gcmole',
'isolated_target': 'run-gcmole',
},
'mjsunit': {
'name': 'Mjsunit',
'tests': ['mjsunit'],
},
'mjsunit_extra': {
'name': 'Mjsunit - extra',
'tests': ['mjsunit'],
'variants': V8Variant('extra'),
},
'mjsunit_sp_frame_access': {
'name': 'Mjsunit - sp frame access',
'tests': ['mjsunit'],
'test_args': ['--extra-flags=--turbo_sp_frame_access'],
'variants': V8Variant('turbofan'),
},
'mozilla': {
'name': 'Mozilla',
'tests': ['mozilla'],
},
'optimize_for_size': {
'name': 'OptimizeForSize',
'tests': ['optimize_for_size'],
'suite_mapping': ['mjsunit', 'cctest', 'webkit', 'intl'],
'test_args': ['--extra-flags=--optimize-for-size'],
'variants': V8Variant('default'),
},
'simpleleak': {
'tool': 'run-valgrind',
'isolated_target': 'run-valgrind',
},
'test262': {
'name': 'Test262 - no variants',
'tests': ['test262'],
'test_args': ['--download-data'],
'variants': V8Variant('default'),
},
'test262_extra': {
'name': 'Test262 - extra',
'tests': ['test262'],
'test_args': ['--download-data'],
'variants': V8Variant('extra'),
},
'test262_variants': {
'name': 'Test262',
'tests': ['test262'],
'test_args': ['--download-data'],
},
'unittests': {
'name': 'Unittests',
'tests': ['unittests'],
},
'v8initializers': {
'tool': 'check-static-initializers',
'isolated_target': 'check-static-initializers',
},
'v8testing': {
'name': 'Check',
'tests': ['bot_default'],
'suite_mapping': [
'mjsunit', 'cctest', 'webkit', 'message', 'preparser', 'intl'],
},
'v8testing_extra': {
'name': 'Check - extra',
'tests': ['bot_default'],
'suite_mapping': [
'mjsunit', 'cctest', 'webkit', 'message', 'preparser', 'intl'],
'variants': V8Variant('extra'),
},
'webkit': {
'name': 'Webkit',
'tests': ['webkit'],
},
})
class NullCoverageContext(object):
"""Null object to represent testing without collecting coverage."""
def get_test_runner_args(self):
return []
def get_swarming_collect_args(self):
return []
def setup(self):
pass
def post_run(self):
pass
def maybe_upload(self):
pass
NULL_COVERAGE = NullCoverageContext()
class SanitizerCoverageContext(object):
"""Context during testing to collect coverage data.
Only testing on swarming is supported.
"""
def __init__(self, api, v8):
self.api = api
self.v8 = v8
self.coverage_dir = api.path.mkdtemp('coverage_output')
def get_test_runner_args(self):
"""Returns the test runner arguments for collecting coverage data."""
return ['--sancov-dir', '${ISOLATED_OUTDIR}']
def get_swarming_collect_args(self):
"""Returns the swarming collect step's arguments for merging."""
return [
'--coverage-dir', self.coverage_dir,
'--sancov-merger', self.api.path['checkout'].join(
'tools', 'sanitizers', 'sancov_merger.py'),
]
def setup(self):
"""Build data file with initial zero coverage data.
To be called before any coverage data from testing is merged in.
"""
self.api.python(
'Initialize coverage data',
self.api.path['checkout'].join(
'tools', 'sanitizers', 'sancov_formatter.py'),
[
'all',
'--json-output', self.coverage_dir.join('data.json'),
],
)
def post_run(self):
"""Merge coverage data from one test run.
To be called after every test step. Requires existing initial zero
coverage data, obtained by calling setup().
"""
self.api.python(
'Merge coverage data',
self.api.path['checkout'].join(
'tools', 'sanitizers', 'sancov_formatter.py'),
[
'merge',
'--json-input', self.coverage_dir.join('data.json'),
'--json-output', self.coverage_dir.join('data.json'),
'--coverage-dir', self.coverage_dir,
],
)
self.api.python.inline(
'Purge sancov files',
"""
import glob
import os
for f in glob.glob('%s'):
os.remove(f)
""" % self.coverage_dir.join('*.sancov'),
)
def maybe_upload(self):
"""Uploads coverage data to google storage if on tryserver."""
if self.api.tryserver.is_tryserver:
assert self.api.properties['issue']
assert self.api.properties['patchset']
results_path = '/'.join([
'tryserver',
'sanitizer_coverage',
str(self.api.properties['issue']),
str(self.api.properties['patchset']),
self.v8.bot_config.get('sanitizer_coverage_folder'),
])
self.api.gsutil.upload(
self.coverage_dir.join('data.json'),
'chromium-v8',
results_path + '/data.json',
)
data_dir = self.api.path.mkdtemp('coverage_data')
self.api.python(
'Split coverage data',
self.api.path['checkout'].join(
'tools', 'sanitizers', 'sancov_formatter.py'),
[
'split',
'--json-input', self.coverage_dir.join('data.json'),
'--output-dir', data_dir,
],
# Allow to work with older v8 revisions that don't have the split
# function in which case the directory will stay empty.
# TODO(machenbach): Remove this when v8's passed CP 34834 + 1000.
ok_ret='any',
)
self.api.gsutil(
[
'-m', 'cp', '-a', 'public-read', '-R', data_dir.join('*'),
'gs://chromium-v8/%s/' % results_path,
],
'coverage data',
# Same as in the step above.
ok_ret='any',
)
class BaseTest(object):
def __init__(self, test_step_config, api, v8):
self.test_step_config = test_step_config
self.name = test_step_config.name
self.api = api
self.v8 = v8
def _get_isolated_hash(self, test):
isolated = test.get('isolated_target')
if not isolated:
# Normally we run only one test and the isolate name is the same as the
# test name.
assert len(test['tests']) == 1
isolated = test['tests'][0]
isolated_hash = self.v8.isolated_tests.get(isolated)
# TODO(machenbach): Maybe this is too hard. Implement a more forgiving
# solution.
assert isolated_hash
return isolated_hash
@property
def uses_swarming(self):
"""Returns true if the test uses swarming."""
return False
def apply_filter(self):
# Run all tests by default.
return True
def pre_run(self, test=None, coverage_context=NULL_COVERAGE, **kwargs):
pass # pragma: no cover
def run(self, test=None, coverage_context=NULL_COVERAGE, **kwargs):
raise NotImplementedError() # pragma: no cover
def rerun(self, failure_dict, **kwargs): # pragma: no cover
raise NotImplementedError()
class V8Test(BaseTest):
def apply_filter(self):
self.applied_test_filter = self.v8._applied_test_filter(
TEST_CONFIGS[self.name])
if self.v8.test_filter and not self.applied_test_filter:
self.api.step(TEST_CONFIGS[self.name]['name'] + ' - skipped', cmd=None)
return False
return True
def run(self, test=None, coverage_context=NULL_COVERAGE, **kwargs):
test = test or TEST_CONFIGS[self.name]
full_args, env = self.v8._setup_test_runner(
test, self.applied_test_filter, self.test_step_config)
if self.v8.c.testing.may_shard and self.v8.c.testing.SHARD_COUNT > 1:
full_args += [
'--shard-count=%d' % self.v8.c.testing.SHARD_COUNT,
'--shard-run=%d' % self.v8.c.testing.SHARD_RUN,
]
full_args += [
'--json-test-results',
self.api.json.output(add_json_log=False),
]
self.api.python(
test['name'] + self.test_step_config.suffix,
self.api.path['checkout'].join('tools', 'run-tests.py'),
full_args,
cwd=self.api.path['checkout'],
env=env,
step_test_data=lambda: self.v8.test_api.output_json(),
**kwargs
)
return self.post_run(test)
def post_run(self, test, coverage_context=NULL_COVERAGE):
# The active step was either a local test run or the swarming collect step.
step_result = self.api.step.active_result
json_output = step_result.json.output
# Log used test filters.
if self.applied_test_filter:
step_result.presentation.logs['test filter'] = self.applied_test_filter
# The output is expected to be a list of architecture dicts that
# each contain a results list. On buildbot, there is only one
# architecture.
assert len(json_output) == 1
self.v8._update_durations(json_output[0], step_result.presentation)
failure_factory=Failure.factory_func(self.test_step_config)
failure_log, failures, flake_log, flakes = (
self.v8._get_failure_logs(json_output[0], failure_factory))
self.v8._update_failure_presentation(
failure_log, failures, step_result.presentation)
if failure_log and failures:
# Mark the test step as failure only if there were real failures (i.e.
# non-flakes) present.
step_result.presentation.status = self.api.step.FAILURE
if flake_log and flakes:
# Emit a separate step to show flakes from the previous step
# to not close the tree.
step_result = self.api.step(
test['name'] + self.test_step_config.suffix + ' (flakes)', cmd=None)
step_result.presentation.status = self.api.step.WARNING
self.v8._update_failure_presentation(
flake_log, flakes, step_result.presentation)
coverage_context.post_run()
return TestResults(failures, flakes, [])
def _setup_rerun_config(self, failure_dict):
"""Return: A test config that reproduces a specific failure."""
# Make sure bisection is only activated on builders that give enough
# information to retry.
assert failure_dict.get('variant')
assert failure_dict.get('random_seed')
orig_config = TEST_CONFIGS[self.name]
# If not specified, the isolated target is the same as the first test of
# the original list. We need to set it explicitly now, as the tests
# parameter changes on rerun, but the isolated target is still the same.
isolated_target = orig_config.get(
'isolated_target', orig_config['tests'][0])
test_args = list(orig_config.get('test_args', [])) + [
'--random-seed', failure_dict['random_seed'],
]
rerun_config = {
'name': 'Retry',
'isolated_target': isolated_target,
'tests': [failure_dict['name']],
'test_args': test_args,
'variants': V8Variant(failure_dict['variant'])
}
# Switch off test filters on rerun.
self.applied_test_filter = None
return rerun_config
def rerun(self, failure_dict, **kwargs):
return self.run(test=self._setup_rerun_config(failure_dict), **kwargs)
class V8SwarmingTest(V8Test):
@property
def uses_swarming(self):
"""Returns true if the test uses swarming."""
return True
def _v8_collect_step(self, task, coverage_context=NULL_COVERAGE, **kwargs):
"""Produces a step that collects and processes a result of a v8 task."""
# Placeholder for the merged json output.
json_output = self.api.json.output(add_json_log=False)
# Shim script's own arguments.
args = [
'--swarming-client-dir', self.api.swarming_client.path,
'--temp-root-dir', self.api.path['tmp_base'],
'--merged-test-output', json_output,
] + coverage_context.get_swarming_collect_args()
# Arguments for actual 'collect' command.
args.append('--')
args.extend(self.api.swarming.get_collect_cmd_args(task))
return self.api.python(
name=self.test['name'] + self.test_step_config.suffix,
script=self.v8.resource('collect_v8_task.py'),
args=args,
allow_subannotations=True,
infra_step=True,
step_test_data=kwargs.pop('step_test_data', None),
**kwargs)
def pre_run(self, test=None, coverage_context=NULL_COVERAGE, **kwargs):
# Set up arguments for test runner.
self.test = test or TEST_CONFIGS[self.name]
extra_args, _ = self.v8._setup_test_runner(
self.test, self.applied_test_filter, self.test_step_config)
# Let json results be stored in swarming's output folder. The collect
# step will copy the folder's contents back to the client.
extra_args += [
'--swarming',
'--json-test-results',
'${ISOLATED_OUTDIR}/output.json',
] + coverage_context.get_test_runner_args()
# Initialize number of shards, either per test or per builder.
shards = 1
if self.v8.c.testing.may_shard:
shards = self.test_step_config.shards
if self.v8.c.testing.SHARD_COUNT > 1: # pragma: no cover
shards = self.v8.c.testing.SHARD_COUNT
# Initialize swarming task with custom data-collection step for v8
# test-runner output.
self.task = self.api.swarming.task(
title=self.test['name'] + self.test_step_config.suffix,
isolated_hash=self._get_isolated_hash(self.test),
shards=shards,
extra_args=extra_args,
)
self.task.collect_step = lambda task, **kw: (
self._v8_collect_step(task, coverage_context, **kw))
# Add custom dimensions.
if self.v8.bot_config.get('swarming_dimensions'):
self.task.dimensions.update(self.v8.bot_config['swarming_dimensions'])
self.api.swarming.trigger_task(self.task)
def run(self, coverage_context=NULL_COVERAGE, **kwargs):
# TODO(machenbach): Soften this when softening 'assert isolated_hash'
# above.
assert self.task
result = TestResults.empty()
try:
# Collect swarming results. Use the same test simulation data for the
# swarming collect step like for local testing.
self.api.swarming.collect_task(
self.task,
step_test_data=lambda: self.v8.test_api.output_json(),
)
except self.api.step.InfraFailure as e:
result += TestResults.infra_failure(e)
return result + self.post_run(self.test, coverage_context)
def rerun(self, failure_dict, **kwargs):
self.pre_run(test=self._setup_rerun_config(failure_dict), **kwargs)
return self.run(**kwargs)
class V8Presubmit(BaseTest):
def run(self, **kwargs):
self.api.python(
'Presubmit',
self.api.path['checkout'].join('tools', 'presubmit.py'),
cwd=self.api.path['checkout'],
)
return TestResults.empty()
class V8GenericSwarmingTest(BaseTest):
def __init__(self, test_step_config, api, v8,
title='Generic test', extra_args=None):
super(V8GenericSwarmingTest, self).__init__(test_step_config, api, v8)
self._extra_args = extra_args or []
self._title = title
@property
def title(self):
return self._title # pragma: no cover
@property
def extra_args(self):
return self._extra_args # pragma: no cover
@property
def task_output_dir(self):
return None # pragma: no cover
@property
def uses_swarming(self):
"""Returns true if the test uses swarming."""
return True
def pre_run(self, test=None, **kwargs):
self.test = test or TEST_CONFIGS[self.name]
self.task = self.api.swarming.task(
title=self.title,
isolated_hash=self._get_isolated_hash(self.test),
extra_args=self.extra_args,
task_output_dir=self.task_output_dir,
)
self.api.swarming.trigger_task(self.task)
def run(self, **kwargs):
assert self.task
self.api.swarming.collect_task(self.task)
return TestResults.empty()
class V8CompositeSwarmingTest(BaseTest):
@property
def composite_tests(self):
"""Returns: An iterable of V8GenericSwarmingTest instances."""
raise NotImplementedError() # pragma: no cover
@property
def uses_swarming(self):
"""Returns true if the test uses swarming."""
return True
def pre_run(self, test=None, **kwargs):
self.composites = list(self.composite_tests)
for c in self.composites:
c.pre_run(test, **kwargs)
def run(self, **kwargs):
for c in self.composites:
c.run(**kwargs)
return TestResults.empty()
class V8CheckInitializers(V8GenericSwarmingTest):
@property
def title(self):
return 'Static-Initializers'
@property
def extra_args(self):
return [self.v8.relative_path_to_d8]
class V8Fuzzer(V8GenericSwarmingTest):
def __init__(self, test_step_config, api, v8,
title='Generic test', extra_args=None):
self.output_dir = api.path.mkdtemp('swarming_output')
self.archive = 'fuzz-results-%s.tar.bz2' % (
api.properties['parent_got_revision'])
super(V8Fuzzer, self).__init__(
test_step_config, api, v8,
title='Fuzz',
extra_args=[
v8.relative_path_to_d8,
'${ISOLATED_OUTDIR}/%s' % self.archive,
],
)
@property
def task_output_dir(self):
return self.output_dir
def run(self, **kwargs):
try:
super(V8Fuzzer, self).run(**kwargs)
except self.api.step.StepFailure as e:
self.api.gsutil.upload(
self.output_dir.join('0', self.archive),
'chromium-v8',
self.api.path.join('fuzzer-archives', self.archive),
)
raise e
return TestResults.empty()
class V8DeoptFuzzer(V8GenericSwarmingTest):
@property
def title(self):
return 'Deopt Fuzz'
@property
def extra_args(self):
return [
'--mode', self.api.chromium.c.build_config_fs,
'--arch', self.api.chromium.c.gyp_env.GYP_DEFINES['v8_target_arch'],
'--progress', 'verbose',
'--buildbot',
] + self.v8.c.testing.test_args
class V8GCMole(V8CompositeSwarmingTest):
@property
def composite_tests(self):
return [
V8GenericSwarmingTest(
self.test_step_config, self.api, self.v8,
title='GCMole %s' % arch,
extra_args=[arch],
) for arch in ['ia32', 'x64', 'arm', 'arm64']
]
class V8SimpleLeakCheck(V8GenericSwarmingTest):
@property
def title(self):
return 'Simple Leak Check'
@property
def extra_args(self):
return [self.v8.relative_path_to_d8, '-e', 'print(1+2)']
V8_NON_STANDARD_TESTS = freeze({
'presubmit': V8Presubmit,
})
TOOL_TO_TEST = freeze({
'run-tests': V8Test,
})
TOOL_TO_TEST_SWARMING = freeze({
'check-static-initializers': V8CheckInitializers,
'jsfunfuzz': V8Fuzzer,
'run-deopt-fuzzer': V8DeoptFuzzer,
'run-gcmole': V8GCMole,
'run-valgrind': V8SimpleLeakCheck,
'run-tests': V8SwarmingTest,
})
class Failure(object):
def __init__(self, test_step_config, failure_dict, duration):
self.test_step_config = test_step_config
self.failure_dict = failure_dict
self.duration = duration
@staticmethod
def factory_func(test_step_config):
def create(failure_dict, duration):
return Failure(test_step_config, failure_dict, duration)
return create
class TestResults(object):
def __init__(self, failures, flakes, infra_failures):
self.failures = failures
self.flakes = flakes
self.infra_failures = infra_failures
@staticmethod
def empty():
return TestResults([], [], [])
@staticmethod
def infra_failure(exception):
return TestResults([], [], [exception])
@property
def is_negative(self):
return bool(self.failures or self.flakes or self.infra_failures)
@property
def has_failures(self):
return bool(self.failures or self.infra_failures)
def __add__(self, other):
return TestResults(
self.failures + other.failures,
self.flakes + other.flakes,
self.infra_failures + other.infra_failures,
)
def create_test(test_step_config, api, v8_api):
test_cls = V8_NON_STANDARD_TESTS.get(test_step_config.name)
if not test_cls:
# TODO(machenbach): Implement swarming for non-standard tests.
if v8_api.bot_config.get('enable_swarming') and test_step_config.swarming:
tools_mapping = TOOL_TO_TEST_SWARMING
else:
tools_mapping = TOOL_TO_TEST
# The tool the test is going to use. Default: V8 test runner (run-tests).
tool = TEST_CONFIGS[test_step_config.name].get('tool', 'run-tests')
test_cls = tools_mapping[tool]
return test_cls(test_step_config, api, v8_api)