| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610 | #!/usr/bin/env python# Copyright 2015 gRPC authors.## Licensed under the Apache License, Version 2.0 (the "License");# you may not use this file except in compliance with the License.# You may obtain a copy of the License at##     http://www.apache.org/licenses/LICENSE-2.0## Unless required by applicable law or agreed to in writing, software# distributed under the License is distributed on an "AS IS" BASIS,# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.# See the License for the specific language governing permissions and# limitations under the License."""Run test matrix."""from __future__ import print_functionimport argparseimport multiprocessingimport osimport sysimport python_utils.jobset as jobsetimport python_utils.report_utils as report_utilsfrom python_utils.filter_pull_request_tests import filter_tests_ROOT = os.path.abspath(os.path.join(os.path.dirname(sys.argv[0]), '../..'))os.chdir(_ROOT)_DEFAULT_RUNTESTS_TIMEOUT = 1 * 60 * 60# Set the timeout high to allow enough time for sanitizers and pre-building# clang docker._CPP_RUNTESTS_TIMEOUT = 4 * 60 * 60# C++ TSAN takes longer than other sanitizers_CPP_TSAN_RUNTESTS_TIMEOUT = 8 * 60 * 60# Set timeout high for ObjC for Cocoapods to install pods_OBJC_RUNTESTS_TIMEOUT = 90 * 60# Number of jobs assigned to each run_tests.py instance_DEFAULT_INNER_JOBS = 2def _safe_report_name(name):    """Reports with '+' in target name won't show correctly in ResultStore"""    return name.replace('+', 'p')def _report_filename(name):    """Generates report file name with directory structure that leads to better presentation by internal CI"""    # 'sponge_log.xml' suffix must be there for results to get recognized by kokoro.    return '%s/%s' % (_safe_report_name(name), 'sponge_log.xml')def _report_logfilename(name):    """Generates log file name that corresponds to name generated by _report_filename"""    # 'sponge_log.log' suffix must be there for log to get recognized as "target log"    # for the corresponding 'sponge_log.xml' report.    return '%s/%s' % (_safe_report_name(name), 'sponge_log.log')def _docker_jobspec(name,                    runtests_args=[],                    runtests_envs={},                    inner_jobs=_DEFAULT_INNER_JOBS,                    timeout_seconds=None):    """Run a single instance of run_tests.py in a docker container"""    if not timeout_seconds:        timeout_seconds = _DEFAULT_RUNTESTS_TIMEOUT    test_job = jobset.JobSpec(        cmdline=[            'python', 'tools/run_tests/run_tests.py', '--use_docker', '-t',            '-j',            str(inner_jobs), '-x',            _report_filename(name), '--report_suite_name',            '%s' % _safe_report_name(name)        ] + runtests_args,        environ=runtests_envs,        shortname='run_tests_%s' % name,        timeout_seconds=timeout_seconds,        logfilename=_report_logfilename(name))    return test_jobdef _workspace_jobspec(name,                       runtests_args=[],                       workspace_name=None,                       runtests_envs={},                       inner_jobs=_DEFAULT_INNER_JOBS,                       timeout_seconds=None):    """Run a single instance of run_tests.py in a separate workspace"""    if not workspace_name:        workspace_name = 'workspace_%s' % name    if not timeout_seconds:        timeout_seconds = _DEFAULT_RUNTESTS_TIMEOUT    env = {'WORKSPACE_NAME': workspace_name}    env.update(runtests_envs)    test_job = jobset.JobSpec(        cmdline=[            'bash', 'tools/run_tests/helper_scripts/run_tests_in_workspace.sh',            '-t', '-j',            str(inner_jobs), '-x',            '../%s' % _report_filename(name), '--report_suite_name',            '%s' % _safe_report_name(name)        ] + runtests_args,        environ=env,        shortname='run_tests_%s' % name,        timeout_seconds=timeout_seconds,        logfilename=_report_logfilename(name))    return test_jobdef _generate_jobs(languages,                   configs,                   platforms,                   iomgr_platforms=['native'],                   arch=None,                   compiler=None,                   labels=[],                   extra_args=[],                   extra_envs={},                   inner_jobs=_DEFAULT_INNER_JOBS,                   timeout_seconds=None):    result = []    for language in languages:        for platform in platforms:            for iomgr_platform in iomgr_platforms:                for config in configs:                    name = '%s_%s_%s_%s' % (language, platform, config,                                            iomgr_platform)                    runtests_args = [                        '-l', language, '-c', config, '--iomgr_platform',                        iomgr_platform                    ]                    if arch or compiler:                        name += '_%s_%s' % (arch, compiler)                        runtests_args += [                            '--arch', arch, '--compiler', compiler                        ]                    if '--build_only' in extra_args:                        name += '_buildonly'                    for extra_env in extra_envs:                        name += '_%s_%s' % (extra_env, extra_envs[extra_env])                    runtests_args += extra_args                    if platform == 'linux':                        job = _docker_jobspec(                            name=name,                            runtests_args=runtests_args,                            runtests_envs=extra_envs,                            inner_jobs=inner_jobs,                            timeout_seconds=timeout_seconds)                    else:                        job = _workspace_jobspec(                            name=name,                            runtests_args=runtests_args,                            runtests_envs=extra_envs,                            inner_jobs=inner_jobs,                            timeout_seconds=timeout_seconds)                    job.labels = [platform, config, language, iomgr_platform                                 ] + labels                    result.append(job)    return resultdef _create_test_jobs(extra_args=[], inner_jobs=_DEFAULT_INNER_JOBS):    test_jobs = []    # sanity tests    test_jobs += _generate_jobs(        languages=['sanity'],        configs=['dbg', 'opt'],        platforms=['linux'],        labels=['basictests'],        extra_args=extra_args,        inner_jobs=inner_jobs)    # supported on linux only    test_jobs += _generate_jobs(        languages=['php7'],        configs=['dbg', 'opt'],        platforms=['linux'],        labels=['basictests', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    # supported on all platforms.    test_jobs += _generate_jobs(        languages=['c'],        configs=['dbg', 'opt'],        platforms=['linux', 'macos', 'windows'],        labels=['basictests', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    test_jobs += _generate_jobs(        languages=['csharp'],        configs=['dbg', 'opt'],        platforms=['linux', 'macos', 'windows'],        labels=['basictests', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    test_jobs += _generate_jobs(        languages=['python'],        configs=['opt'],        platforms=['linux', 'macos', 'windows'],        iomgr_platforms=['native', 'gevent'],        labels=['basictests', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    # supported on linux and mac.    test_jobs += _generate_jobs(        languages=['c++'],        configs=['dbg', 'opt'],        platforms=['linux', 'macos'],        labels=['basictests', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    test_jobs += _generate_jobs(        languages=['grpc-node', 'ruby', 'php'],        configs=['dbg', 'opt'],        platforms=['linux', 'macos'],        labels=['basictests', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    # supported on mac only.    test_jobs += _generate_jobs(        languages=['objc'],        configs=['dbg', 'opt'],        platforms=['macos'],        labels=['basictests', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_OBJC_RUNTESTS_TIMEOUT)    # sanitizers    test_jobs += _generate_jobs(        languages=['c'],        configs=['msan', 'asan', 'tsan', 'ubsan'],        platforms=['linux'],        arch='x64',        compiler='clang7.0',        labels=['sanitizers', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    test_jobs += _generate_jobs(        languages=['c++'],        configs=['asan'],        platforms=['linux'],        arch='x64',        compiler='clang7.0',        labels=['sanitizers', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    test_jobs += _generate_jobs(        languages=['c++'],        configs=['tsan'],        platforms=['linux'],        arch='x64',        compiler='clang7.0',        labels=['sanitizers', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_CPP_TSAN_RUNTESTS_TIMEOUT)    return test_jobsdef _create_portability_test_jobs(extra_args=[],                                  inner_jobs=_DEFAULT_INNER_JOBS):    test_jobs = []    # portability C x86    test_jobs += _generate_jobs(        languages=['c'],        configs=['dbg'],        platforms=['linux'],        arch='x86',        compiler='default',        labels=['portability', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    # portability C and C++ on x64    for compiler in [            'gcc4.8', 'gcc5.3', 'gcc7.2', 'gcc_musl', 'clang3.5', 'clang3.6',            'clang3.7', 'clang7.0'    ]:        test_jobs += _generate_jobs(            languages=['c', 'c++'],            configs=['dbg'],            platforms=['linux'],            arch='x64',            compiler=compiler,            labels=['portability', 'corelang'],            extra_args=extra_args,            inner_jobs=inner_jobs,            timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    # portability C on Windows 64-bit (x86 is the default)    test_jobs += _generate_jobs(        languages=['c'],        configs=['dbg'],        platforms=['windows'],        arch='x64',        compiler='default',        labels=['portability', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    # portability C++ on Windows    # TODO(jtattermusch): some of the tests are failing, so we force --build_only    test_jobs += _generate_jobs(        languages=['c++'],        configs=['dbg'],        platforms=['windows'],        arch='default',        compiler='default',        labels=['portability', 'corelang'],        extra_args=extra_args + ['--build_only'],        inner_jobs=inner_jobs)    # portability C and C++ on Windows using VS2017 (build only)    # TODO(jtattermusch): some of the tests are failing, so we force --build_only    test_jobs += _generate_jobs(        languages=['c', 'c++'],        configs=['dbg'],        platforms=['windows'],        arch='x64',        compiler='cmake_vs2017',        labels=['portability', 'corelang'],        extra_args=extra_args + ['--build_only'],        inner_jobs=inner_jobs)    # C and C++ with the c-ares DNS resolver on Linux    test_jobs += _generate_jobs(        languages=['c', 'c++'],        configs=['dbg'],        platforms=['linux'],        labels=['portability', 'corelang'],        extra_args=extra_args,        extra_envs={'GRPC_DNS_RESOLVER': 'ares'},        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    # C and C++ with no-exceptions on Linux    test_jobs += _generate_jobs(        languages=['c', 'c++'],        configs=['noexcept'],        platforms=['linux'],        labels=['portability', 'corelang'],        extra_args=extra_args,        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    # TODO(zyc): Turn on this test after adding c-ares support on windows.    # C with the c-ares DNS resolver on Windows    # test_jobs += _generate_jobs(languages=['c'],    #                             configs=['dbg'], platforms=['windows'],    #                             labels=['portability', 'corelang'],    #                             extra_args=extra_args,    #                             extra_envs={'GRPC_DNS_RESOLVER': 'ares'})    # C and C++ build with cmake on Linux    # TODO(jtattermusch): some of the tests are failing, so we force --build_only    # to make sure it's buildable at least.    test_jobs += _generate_jobs(        languages=['c', 'c++'],        configs=['dbg'],        platforms=['linux'],        arch='default',        compiler='cmake',        labels=['portability', 'corelang'],        extra_args=extra_args + ['--build_only'],        inner_jobs=inner_jobs)    test_jobs += _generate_jobs(        languages=['python'],        configs=['dbg'],        platforms=['linux'],        arch='default',        compiler='python_alpine',        labels=['portability', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    test_jobs += _generate_jobs(        languages=['csharp'],        configs=['dbg'],        platforms=['linux'],        arch='default',        compiler='coreclr',        labels=['portability', 'multilang'],        extra_args=extra_args,        inner_jobs=inner_jobs)    test_jobs += _generate_jobs(        languages=['c'],        configs=['dbg'],        platforms=['linux'],        iomgr_platforms=['uv'],        labels=['portability', 'corelang'],        extra_args=extra_args,        inner_jobs=inner_jobs,        timeout_seconds=_CPP_RUNTESTS_TIMEOUT)    return test_jobsdef _allowed_labels():    """Returns a list of existing job labels."""    all_labels = set()    for job in _create_test_jobs() + _create_portability_test_jobs():        for label in job.labels:            all_labels.add(label)    return sorted(all_labels)def _runs_per_test_type(arg_str):    """Auxiliary function to parse the "runs_per_test" flag."""    try:        n = int(arg_str)        if n <= 0: raise ValueError        return n    except:        msg = '\'{}\' is not a positive integer'.format(arg_str)        raise argparse.ArgumentTypeError(msg)if __name__ == "__main__":    argp = argparse.ArgumentParser(        description='Run a matrix of run_tests.py tests.')    argp.add_argument(        '-j',        '--jobs',        default=multiprocessing.cpu_count() / _DEFAULT_INNER_JOBS,        type=int,        help='Number of concurrent run_tests.py instances.')    argp.add_argument(        '-f',        '--filter',        choices=_allowed_labels(),        nargs='+',        default=[],        help='Filter targets to run by label with AND semantics.')    argp.add_argument(        '--exclude',        choices=_allowed_labels(),        nargs='+',        default=[],        help='Exclude targets with any of given labels.')    argp.add_argument(        '--build_only',        default=False,        action='store_const',        const=True,        help='Pass --build_only flag to run_tests.py instances.')    argp.add_argument(        '--force_default_poller',        default=False,        action='store_const',        const=True,        help='Pass --force_default_poller to run_tests.py instances.')    argp.add_argument(        '--dry_run',        default=False,        action='store_const',        const=True,        help='Only print what would be run.')    argp.add_argument(        '--filter_pr_tests',        default=False,        action='store_const',        const=True,        help='Filters out tests irrelevant to pull request changes.')    argp.add_argument(        '--base_branch',        default='origin/master',        type=str,        help='Branch that pull request is requesting to merge into')    argp.add_argument(        '--inner_jobs',        default=_DEFAULT_INNER_JOBS,        type=int,        help='Number of jobs in each run_tests.py instance')    argp.add_argument(        '-n',        '--runs_per_test',        default=1,        type=_runs_per_test_type,        help='How many times to run each tests. >1 runs implies ' +        'omitting passing test from the output & reports.')    argp.add_argument(        '--max_time',        default=-1,        type=int,        help='Maximum amount of time to run tests for' +        '(other tests will be skipped)')    argp.add_argument(        '--internal_ci',        default=False,        action='store_const',        const=True,        help=        '(Deprecated, has no effect) Put reports into subdirectories to improve presentation of '        'results by Kokoro.')    argp.add_argument(        '--bq_result_table',        default='',        type=str,        nargs='?',        help='Upload test results to a specified BQ table.')    args = argp.parse_args()    extra_args = []    if args.build_only:        extra_args.append('--build_only')    if args.force_default_poller:        extra_args.append('--force_default_poller')    if args.runs_per_test > 1:        extra_args.append('-n')        extra_args.append('%s' % args.runs_per_test)        extra_args.append('--quiet_success')    if args.max_time > 0:        extra_args.extend(('--max_time', '%d' % args.max_time))    if args.bq_result_table:        extra_args.append('--bq_result_table')        extra_args.append('%s' % args.bq_result_table)        extra_args.append('--measure_cpu_costs')    all_jobs = _create_test_jobs(extra_args=extra_args, inner_jobs=args.inner_jobs) + \               _create_portability_test_jobs(extra_args=extra_args, inner_jobs=args.inner_jobs)    jobs = []    for job in all_jobs:        if not args.filter or all(                filter in job.labels for filter in args.filter):            if not any(exclude_label in job.labels                       for exclude_label in args.exclude):                jobs.append(job)    if not jobs:        jobset.message(            'FAILED', 'No test suites match given criteria.', do_newline=True)        sys.exit(1)    print('IMPORTANT: The changes you are testing need to be locally committed')    print('because only the committed changes in the current branch will be')    print('copied to the docker environment or into subworkspaces.')    skipped_jobs = []    if args.filter_pr_tests:        print('Looking for irrelevant tests to skip...')        relevant_jobs = filter_tests(jobs, args.base_branch)        if len(relevant_jobs) == len(jobs):            print('No tests will be skipped.')        else:            print('These tests will be skipped:')            skipped_jobs = list(set(jobs) - set(relevant_jobs))            # Sort by shortnames to make printing of skipped tests consistent            skipped_jobs.sort(key=lambda job: job.shortname)            for job in list(skipped_jobs):                print('  %s' % job.shortname)        jobs = relevant_jobs    print('Will run these tests:')    for job in jobs:        if args.dry_run:            print('  %s: "%s"' % (job.shortname, ' '.join(job.cmdline)))        else:            print('  %s' % job.shortname)    print    if args.dry_run:        print('--dry_run was used, exiting')        sys.exit(1)    jobset.message('START', 'Running test matrix.', do_newline=True)    num_failures, resultset = jobset.run(        jobs, newline_on_success=True, travis=True, maxjobs=args.jobs)    # Merge skipped tests into results to show skipped tests on report.xml    if skipped_jobs:        ignored_num_skipped_failures, skipped_results = jobset.run(            skipped_jobs, skip_jobs=True)        resultset.update(skipped_results)    report_utils.render_junit_xml_report(        resultset,        _report_filename('aggregate_tests'),        suite_name='aggregate_tests')    if num_failures == 0:        jobset.message(            'SUCCESS',            'All run_tests.py instance finished successfully.',            do_newline=True)    else:        jobset.message(            'FAILED',            'Some run_tests.py instance have failed.',            do_newline=True)        sys.exit(1)
 |