| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350 | #!/usr/bin/env python2.7# Copyright 2015, Google Inc.# All rights reserved.## Redistribution and use in source and binary forms, with or without# modification, are permitted provided that the following conditions are# met:##     * Redistributions of source code must retain the above copyright# notice, this list of conditions and the following disclaimer.#     * Redistributions in binary form must reproduce the above# copyright notice, this list of conditions and the following disclaimer# in the documentation and/or other materials provided with the# distribution.#     * Neither the name of Google Inc. nor the names of its# contributors may be used to endorse or promote products derived from# this software without specific prior written permission.## THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS# "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT# LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR# A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT# OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,# SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT# LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE# OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE."""Run test matrix."""import argparseimport jobsetimport multiprocessingimport osimport report_utilsimport sysfrom filter_pull_request_tests import filter_tests_ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..'))os.chdir(_ROOT)# Set the timeout high to allow enough time for sanitizers and pre-building# clang docker._RUNTESTS_TIMEOUT = 4*60*60# Number of jobs assigned to each run_tests.py instance_DEFAULT_INNER_JOBS = 2def _docker_jobspec(name, runtests_args=[], inner_jobs=_DEFAULT_INNER_JOBS):  """Run a single instance of run_tests.py in a docker container"""  test_job = jobset.JobSpec(          cmdline=['python', 'tools/run_tests/run_tests.py',                   '--use_docker',                   '-t',                   '-j', str(inner_jobs),                   '-x', 'report_%s.xml' % name,                   '--report_suite_name', '%s' % name] + runtests_args,          shortname='run_tests_%s' % name,          timeout_seconds=_RUNTESTS_TIMEOUT)  return test_jobdef _workspace_jobspec(name, runtests_args=[], workspace_name=None, inner_jobs=_DEFAULT_INNER_JOBS):  """Run a single instance of run_tests.py in a separate workspace"""  if not workspace_name:    workspace_name = 'workspace_%s' % name  env = {'WORKSPACE_NAME': workspace_name}  test_job = jobset.JobSpec(          cmdline=['tools/run_tests/run_tests_in_workspace.sh',                   '-t',                   '-j', str(inner_jobs),                   '-x', '../report_%s.xml' % name,                   '--report_suite_name', '%s' % name] + runtests_args,          environ=env,          shortname='run_tests_%s' % name,          timeout_seconds=_RUNTESTS_TIMEOUT)  return test_jobdef _generate_jobs(languages, configs, platforms,                  arch=None, compiler=None,                  labels=[], extra_args=[],                  inner_jobs=_DEFAULT_INNER_JOBS):  result = []  for language in languages:    for platform in platforms:      for config in configs:        name = '%s_%s_%s' % (language, platform, config)        runtests_args = ['-l', language,                         '-c', config]        if arch or compiler:          name += '_%s_%s' % (arch, compiler)          runtests_args += ['--arch', arch,                            '--compiler', compiler]        runtests_args += extra_args        if platform == 'linux':          job = _docker_jobspec(name=name, runtests_args=runtests_args, inner_jobs=inner_jobs)        else:          job = _workspace_jobspec(name=name, runtests_args=runtests_args, inner_jobs=inner_jobs)        job.labels = [platform, config, language] + labels        result.append(job)  return resultdef _create_test_jobs(extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS):  test_jobs = []  # supported on linux only  test_jobs += _generate_jobs(languages=['sanity', 'php7'],                             configs=['dbg', 'opt'],                             platforms=['linux'],                             labels=['basictests'],                             extra_args=extra_args,                             inner_jobs=inner_jobs)  # supported on all platforms.  test_jobs += _generate_jobs(languages=['c', 'csharp', 'node', 'python'],                             configs=['dbg', 'opt'],                             platforms=['linux', 'macos', 'windows'],                             labels=['basictests'],                             extra_args=extra_args,                             inner_jobs=inner_jobs)  # supported on linux and mac.  test_jobs += _generate_jobs(languages=['c++', 'ruby', 'php'],                              configs=['dbg', 'opt'],                              platforms=['linux', 'macos'],                              labels=['basictests'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  # supported on mac only.  test_jobs += _generate_jobs(languages=['objc'],                              configs=['dbg', 'opt'],                              platforms=['macos'],                              labels=['basictests'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  # sanitizers  test_jobs += _generate_jobs(languages=['c'],                              configs=['msan', 'asan', 'tsan'],                              platforms=['linux'],                              labels=['sanitizers'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  test_jobs += _generate_jobs(languages=['c++'],                              configs=['asan', 'tsan'],                              platforms=['linux'],                              labels=['sanitizers'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  # libuv tests  test_jobs += _generate_jobs(languages=['c'],                              configs=['dbg', 'opt'],                              platforms=['linux'],                              labels=['libuv'],                              extra_args=extra_args + ['--iomgr_platform=uv'],                              inner_jobs=inner_jobs)  return test_jobsdef _create_portability_test_jobs(extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS):  test_jobs = []  # portability C x86  test_jobs += _generate_jobs(languages=['c'],                              configs=['dbg'],                              platforms=['linux'],                              arch='x86',                              compiler='default',                              labels=['portability'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  # portability C and C++ on x64  for compiler in ['gcc4.4', 'gcc4.6', 'gcc5.3',                   'clang3.5', 'clang3.6', 'clang3.7']:    test_jobs += _generate_jobs(languages=['c'],                                configs=['dbg'],                                platforms=['linux'],                                arch='x64',                                compiler=compiler,                                labels=['portability'],                                extra_args=extra_args,                                inner_jobs=inner_jobs)  for compiler in ['gcc4.8', 'gcc5.3',                   'clang3.5', 'clang3.6', 'clang3.7']:    test_jobs += _generate_jobs(languages=['c++'],                                configs=['dbg'],                                platforms=['linux'],                                arch='x64',                                compiler=compiler,                                labels=['portability'],                                extra_args=extra_args,                                inner_jobs=inner_jobs)  # portability C on Windows  for arch in ['x86', 'x64']:    for compiler in ['vs2013', 'vs2015']:      test_jobs += _generate_jobs(languages=['c'],                                  configs=['dbg'],                                  platforms=['windows'],                                  arch=arch,                                  compiler=compiler,                                  labels=['portability'],                                  extra_args=extra_args,                                  inner_jobs=inner_jobs)  test_jobs += _generate_jobs(languages=['python'],                              configs=['dbg'],                              platforms=['linux'],                              arch='default',                              compiler='python3.4',                              labels=['portability'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  test_jobs += _generate_jobs(languages=['csharp'],                              configs=['dbg'],                              platforms=['linux'],                              arch='default',                              compiler='coreclr',                              labels=['portability'],                              extra_args=extra_args,                              inner_jobs=inner_jobs)  return test_jobsdef _allowed_labels():  """Returns a list of existing job labels."""  all_labels = set()  for job in _create_test_jobs() + _create_portability_test_jobs():    for label in job.labels:      all_labels.add(label)  return sorted(all_labels)if __name__ == "__main__":  argp = argparse.ArgumentParser(description='Run a matrix of run_tests.py tests.')  argp.add_argument('-j', '--jobs',                    default=multiprocessing.cpu_count()/_DEFAULT_INNER_JOBS,                    type=int,                    help='Number of concurrent run_tests.py instances.')  argp.add_argument('-f', '--filter',                    choices=_allowed_labels(),                    nargs='+',                    default=[],                    help='Filter targets to run by label with AND semantics.')  argp.add_argument('--build_only',                    default=False,                    action='store_const',                    const=True,                    help='Pass --build_only flag to run_tests.py instances.')  argp.add_argument('--force_default_poller', default=False, action='store_const', const=True,                    help='Pass --force_default_poller to run_tests.py instances.')  argp.add_argument('--dry_run',                    default=False,                    action='store_const',                    const=True,                    help='Only print what would be run.')  argp.add_argument('--filter_pr_tests',                    default=False,                    action='store_const',                    const=True,                    help='Filters out tests irrelavant to pull request changes.')  argp.add_argument('--base_branch',                    default='origin/master',                    type=str,                    help='Branch that pull request is requesting to merge into')  argp.add_argument('--inner_jobs',                    default=_DEFAULT_INNER_JOBS,                    type=int,                    help='Number of jobs in each run_tests.py instance')  args = argp.parse_args()  extra_args = []  if args.build_only:    extra_args.append('--build_only')  if args.force_default_poller:    extra_args.append('--force_default_poller')  all_jobs = _create_test_jobs(extra_args=extra_args, inner_jobs=args.inner_jobs) + \             _create_portability_test_jobs(extra_args=extra_args, inner_jobs=args.inner_jobs)  jobs = []  for job in all_jobs:    if not args.filter or all(filter in job.labels for filter in args.filter):      jobs.append(job)  if not jobs:    jobset.message('FAILED', 'No test suites match given criteria.',                   do_newline=True)    sys.exit(1)  print('IMPORTANT: The changes you are testing need to be locally committed')  print('because only the committed changes in the current branch will be')  print('copied to the docker environment or into subworkspaces.')  skipped_jobs = []  if args.filter_pr_tests:    print('Looking for irrelevant tests to skip...')    relevant_jobs = filter_tests(jobs, args.base_branch)    if len(relevant_jobs) == len(jobs):      print('No tests will be skipped.')    else:      print('These tests will be skipped:')      skipped_jobs = [job for job in jobs if job not in relevant_jobs]      for job in list(skipped_jobs):        print('  %s' % job.shortname)    jobs = relevant_jobs  print('Will run these tests:')  for job in jobs:    if args.dry_run:      print('  %s: "%s"' % (job.shortname, ' '.join(job.cmdline)))    else:      print('  %s' % job.shortname)  print  if args.dry_run:    print('--dry_run was used, exiting')    sys.exit(1)  jobset.message('START', 'Running test matrix.', do_newline=True)  num_failures, resultset = jobset.run(jobs,                                       newline_on_success=True,                                       travis=True,                                       maxjobs=args.jobs)  # Merge skipped tests into results to show skipped tests on report.xml  if skipped_jobs:    skipped_results = jobset.run(skipped_jobs,                                 skip_jobs=True)    resultset.update(skipped_results)  report_utils.render_junit_xml_report(resultset, 'report.xml',                                       suite_name='aggregate_tests')  if num_failures == 0:    jobset.message('SUCCESS', 'All run_tests.py instance finished successfully.',                   do_newline=True)  else:    jobset.message('FAILED', 'Some run_tests.py instance have failed.',                   do_newline=True)    sys.exit(1)
 |