blob: 23c403995953224f97282d1639d12c9c49c6600c [file] [log] [blame]
Ben Murdoch097c5b22016-05-18 11:27:45 +01001#!/usr/bin/env python
2#
3# Copyright 2013 The Chromium Authors. All rights reserved.
4# Use of this source code is governed by a BSD-style license that can be
5# found in the LICENSE file.
6
7"""Runs all types of tests from one unified interface."""
8
9import argparse
10import collections
11import itertools
12import logging
13import os
14import signal
15import sys
16import threading
17import unittest
18
19import devil_chromium
20from devil import base_error
21from devil import devil_env
22from devil.android import device_blacklist
23from devil.android import device_errors
24from devil.android import device_utils
25from devil.android import forwarder
26from devil.android import ports
27from devil.utils import reraiser_thread
28from devil.utils import run_tests_helper
29
30from pylib import constants
31from pylib.constants import host_paths
32from pylib.base import base_test_result
33from pylib.base import environment_factory
34from pylib.base import test_dispatcher
35from pylib.base import test_instance_factory
36from pylib.base import test_run_factory
37from pylib.linker import setup as linker_setup
38from pylib.junit import setup as junit_setup
39from pylib.junit import test_dispatcher as junit_dispatcher
40from pylib.monkey import setup as monkey_setup
41from pylib.monkey import test_options as monkey_test_options
42from pylib.perf import setup as perf_setup
43from pylib.perf import test_options as perf_test_options
44from pylib.perf import test_runner as perf_test_runner
45from pylib.results import json_results
46from pylib.results import report_results
47
48
49_DEVIL_STATIC_CONFIG_FILE = os.path.abspath(os.path.join(
50 host_paths.DIR_SOURCE_ROOT, 'build', 'android', 'devil_config.json'))
51
52
53def AddCommonOptions(parser):
54 """Adds all common options to |parser|."""
55
56 group = parser.add_argument_group('Common Options')
57
58 default_build_type = os.environ.get('BUILDTYPE', 'Debug')
59
60 debug_or_release_group = group.add_mutually_exclusive_group()
61 debug_or_release_group.add_argument(
62 '--debug', action='store_const', const='Debug', dest='build_type',
63 default=default_build_type,
64 help=('If set, run test suites under out/Debug. '
65 'Default is env var BUILDTYPE or Debug.'))
66 debug_or_release_group.add_argument(
67 '--release', action='store_const', const='Release', dest='build_type',
68 help=('If set, run test suites under out/Release. '
69 'Default is env var BUILDTYPE or Debug.'))
70
71 group.add_argument('--build-directory', dest='build_directory',
72 help=('Path to the directory in which build files are'
73 ' located (should not include build type)'))
74 group.add_argument('--output-directory', dest='output_directory',
75 help=('Path to the directory in which build files are'
76 ' located (must include build type). This will take'
77 ' precedence over --debug, --release and'
78 ' --build-directory'))
79 group.add_argument('--num_retries', '--num-retries', dest='num_retries',
80 type=int, default=2,
81 help=('Number of retries for a test before '
82 'giving up (default: %(default)s).'))
83 group.add_argument('-v',
84 '--verbose',
85 dest='verbose_count',
86 default=0,
87 action='count',
88 help='Verbose level (multiple times for more)')
89 group.add_argument('--flakiness-dashboard-server',
90 dest='flakiness_dashboard_server',
91 help=('Address of the server that is hosting the '
92 'Chrome for Android flakiness dashboard.'))
93 group.add_argument('--enable-platform-mode', action='store_true',
94 help=('Run the test scripts in platform mode, which '
95 'conceptually separates the test runner from the '
96 '"device" (local or remote, real or emulated) on '
97 'which the tests are running. [experimental]'))
98 group.add_argument('-e', '--environment', default='local',
99 choices=constants.VALID_ENVIRONMENTS,
100 help='Test environment to run in (default: %(default)s).')
101 group.add_argument('--adb-path',
102 help=('Specify the absolute path of the adb binary that '
103 'should be used.'))
104 group.add_argument('--json-results-file', '--test-launcher-summary-output',
105 dest='json_results_file',
106 help='If set, will dump results in JSON form '
107 'to specified file.')
108
109 logcat_output_group = group.add_mutually_exclusive_group()
110 logcat_output_group.add_argument(
111 '--logcat-output-dir',
112 help='If set, will dump logcats recorded during test run to directory. '
113 'File names will be the device ids with timestamps.')
114 logcat_output_group.add_argument(
115 '--logcat-output-file',
116 help='If set, will merge logcats recorded during test run and dump them '
117 'to the specified file.')
118
119 class FastLocalDevAction(argparse.Action):
120 def __call__(self, parser, namespace, values, option_string=None):
121 namespace.verbose_count = max(namespace.verbose_count, 1)
122 namespace.num_retries = 0
123 namespace.enable_device_cache = True
124 namespace.enable_concurrent_adb = True
125 namespace.skip_clear_data = True
126 namespace.extract_test_list_from_filter = True
127
128 group.add_argument('--fast-local-dev', type=bool, nargs=0,
129 action=FastLocalDevAction,
130 help='Alias for: --verbose --num-retries=0 '
131 '--enable-device-cache --enable-concurrent-adb '
132 '--skip-clear-data --extract-test-list-from-filter')
133
134def ProcessCommonOptions(args):
135 """Processes and handles all common options."""
136 run_tests_helper.SetLogLevel(args.verbose_count)
137 constants.SetBuildType(args.build_type)
138 if args.build_directory:
139 constants.SetBuildDirectory(args.build_directory)
140 if args.output_directory:
141 constants.SetOutputDirectory(args.output_directory)
142
143 devil_custom_deps = None
144 if args.adb_path:
145 devil_custom_deps = {
146 'adb': {
147 devil_env.GetPlatform(): [args.adb_path]
148 }
149 }
150
151 devil_chromium.Initialize(
152 output_directory=constants.GetOutDirectory(),
153 custom_deps=devil_custom_deps)
154
155 # Some things such as Forwarder require ADB to be in the environment path.
156 adb_dir = os.path.dirname(constants.GetAdbPath())
157 if adb_dir and adb_dir not in os.environ['PATH'].split(os.pathsep):
158 os.environ['PATH'] = adb_dir + os.pathsep + os.environ['PATH']
159
160
161def AddRemoteDeviceOptions(parser):
162 group = parser.add_argument_group('Remote Device Options')
163
164 group.add_argument('--trigger',
165 help=('Only triggers the test if set. Stores test_run_id '
166 'in given file path. '))
167 group.add_argument('--collect',
168 help=('Only collects the test results if set. '
169 'Gets test_run_id from given file path.'))
170 group.add_argument('--remote-device', action='append',
171 help='Device type to run test on.')
172 group.add_argument('--results-path',
173 help='File path to download results to.')
174 group.add_argument('--api-protocol',
175 help='HTTP protocol to use. (http or https)')
176 group.add_argument('--api-address',
177 help='Address to send HTTP requests.')
178 group.add_argument('--api-port',
179 help='Port to send HTTP requests to.')
180 group.add_argument('--runner-type',
181 help='Type of test to run as.')
182 group.add_argument('--runner-package',
183 help='Package name of test.')
184 group.add_argument('--device-type',
185 choices=constants.VALID_DEVICE_TYPES,
186 help=('Type of device to run on. iOS or android'))
187 group.add_argument('--device-oem', action='append',
188 help='Device OEM to run on.')
189 group.add_argument('--remote-device-file',
190 help=('File with JSON to select remote device. '
191 'Overrides all other flags.'))
192 group.add_argument('--remote-device-timeout', type=int,
193 help='Times to retry finding remote device')
194 group.add_argument('--network-config', type=int,
195 help='Integer that specifies the network environment '
196 'that the tests will be run in.')
197 group.add_argument('--test-timeout', type=int,
198 help='Test run timeout in seconds.')
199
200 device_os_group = group.add_mutually_exclusive_group()
201 device_os_group.add_argument('--remote-device-minimum-os',
202 help='Minimum OS on device.')
203 device_os_group.add_argument('--remote-device-os', action='append',
204 help='OS to have on the device.')
205
206 api_secret_group = group.add_mutually_exclusive_group()
207 api_secret_group.add_argument('--api-secret', default='',
208 help='API secret for remote devices.')
209 api_secret_group.add_argument('--api-secret-file', default='',
210 help='Path to file that contains API secret.')
211
212 api_key_group = group.add_mutually_exclusive_group()
213 api_key_group.add_argument('--api-key', default='',
214 help='API key for remote devices.')
215 api_key_group.add_argument('--api-key-file', default='',
216 help='Path to file that contains API key.')
217
218
219def AddDeviceOptions(parser):
220 """Adds device options to |parser|."""
221 group = parser.add_argument_group(title='Device Options')
222 group.add_argument('--tool',
223 dest='tool',
224 help=('Run the test under a tool '
225 '(use --tool help to list them)'))
226 group.add_argument('-d', '--device', dest='test_device',
227 help=('Target device for the test suite '
228 'to run on.'))
229 group.add_argument('--blacklist-file', help='Device blacklist file.')
230 group.add_argument('--enable-device-cache', action='store_true',
231 help='Cache device state to disk between runs')
232 group.add_argument('--enable-concurrent-adb', action='store_true',
233 help='Run multiple adb commands at the same time, even '
234 'for the same device.')
235 group.add_argument('--skip-clear-data', action='store_true',
236 help='Do not wipe app data between tests. Use this to '
237 'speed up local development and never on bots '
238 '(increases flakiness)')
239
240
241def AddGTestOptions(parser):
242 """Adds gtest options to |parser|."""
243
244 group = parser.add_argument_group('GTest Options')
245 group.add_argument('-s', '--suite', dest='suite_name',
246 nargs='+', metavar='SUITE_NAME', required=True,
247 help='Executable name of the test suite to run.')
248 group.add_argument('--executable-dist-dir',
249 help="Path to executable's dist directory for native"
250 " (non-apk) tests.")
251 group.add_argument('--test-apk-incremental-install-script',
252 help='Path to install script for the test apk.')
253 group.add_argument('--gtest_also_run_disabled_tests',
254 '--gtest-also-run-disabled-tests',
255 dest='run_disabled', action='store_true',
256 help='Also run disabled tests if applicable.')
257 group.add_argument('-a', '--test-arguments', dest='test_arguments',
258 default='',
259 help='Additional arguments to pass to the test.')
260 group.add_argument('-t', '--shard-timeout',
261 dest='shard_timeout', type=int, default=120,
262 help='Timeout to wait for each test '
263 '(default: %(default)s).')
264 group.add_argument('--isolate_file_path',
265 '--isolate-file-path',
266 dest='isolate_file_path',
267 help='.isolate file path to override the default '
268 'path')
269 group.add_argument('--app-data-file', action='append', dest='app_data_files',
270 help='A file path relative to the app data directory '
271 'that should be saved to the host.')
272 group.add_argument('--app-data-file-dir',
273 help='Host directory to which app data files will be'
274 ' saved. Used with --app-data-file.')
275 group.add_argument('--delete-stale-data', dest='delete_stale_data',
276 action='store_true',
277 help='Delete stale test data on the device.')
278 group.add_argument('--repeat', '--gtest_repeat', '--gtest-repeat',
279 dest='repeat', type=int, default=0,
280 help='Number of times to repeat the specified set of '
281 'tests.')
282 group.add_argument('--break-on-failure', '--break_on_failure',
283 dest='break_on_failure', action='store_true',
284 help='Whether to break on failure.')
285 group.add_argument('--extract-test-list-from-filter',
286 action='store_true',
287 help='When a test filter is specified, and the list of '
288 'tests can be determined from it, skip querying the '
289 'device for the list of all tests. Speeds up local '
290 'development, but is not safe to use on bots ('
291 'http://crbug.com/549214')
292
293 filter_group = group.add_mutually_exclusive_group()
294 filter_group.add_argument('-f', '--gtest_filter', '--gtest-filter',
295 dest='test_filter',
296 help='googletest-style filter string.')
297 filter_group.add_argument('--gtest-filter-file', dest='test_filter_file',
298 help='Path to file that contains googletest-style '
299 'filter strings. (Lines will be joined with '
300 '":" to create a single filter string.)')
301
302 AddDeviceOptions(parser)
303 AddCommonOptions(parser)
304 AddRemoteDeviceOptions(parser)
305
306
307def AddLinkerTestOptions(parser):
308 group = parser.add_argument_group('Linker Test Options')
309 group.add_argument('-f', '--gtest-filter', dest='test_filter',
310 help='googletest-style filter string.')
311 AddCommonOptions(parser)
312 AddDeviceOptions(parser)
313
314
315def AddJavaTestOptions(argument_group):
316 """Adds the Java test options to |option_parser|."""
317
318 argument_group.add_argument(
319 '-f', '--test-filter', '--gtest_filter', '--gtest-filter',
320 dest='test_filter',
321 help=('Test filter (if not fully qualified, will run all matches).'))
322 argument_group.add_argument(
323 '--repeat', dest='repeat', type=int, default=0,
324 help='Number of times to repeat the specified set of tests.')
325 argument_group.add_argument(
326 '--break-on-failure', '--break_on_failure',
327 dest='break_on_failure', action='store_true',
328 help='Whether to break on failure.')
329 argument_group.add_argument(
330 '-A', '--annotation', dest='annotation_str',
331 help=('Comma-separated list of annotations. Run only tests with any of '
332 'the given annotations. An annotation can be either a key or a '
333 'key-values pair. A test that has no annotation is considered '
334 '"SmallTest".'))
335 argument_group.add_argument(
336 '-E', '--exclude-annotation', dest='exclude_annotation_str',
337 help=('Comma-separated list of annotations. Exclude tests with these '
338 'annotations.'))
339 argument_group.add_argument(
340 '--screenshot', dest='screenshot_failures', action='store_true',
341 help='Capture screenshots of test failures')
342 argument_group.add_argument(
343 '--save-perf-json', action='store_true',
344 help='Saves the JSON file for each UI Perf test.')
345 argument_group.add_argument(
346 '--official-build', action='store_true', help='Run official build tests.')
347 argument_group.add_argument(
348 '--test_data', '--test-data', action='append', default=[],
349 help=('Each instance defines a directory of test data that should be '
350 'copied to the target(s) before running the tests. The argument '
351 'should be of the form <target>:<source>, <target> is relative to '
352 'the device data directory, and <source> is relative to the '
353 'chromium build directory.'))
354 argument_group.add_argument(
355 '--disable-dalvik-asserts', dest='set_asserts', action='store_false',
356 default=True, help='Removes the dalvik.vm.enableassertions property')
357
358
359
360def ProcessJavaTestOptions(args):
361 """Processes options/arguments and populates |options| with defaults."""
362
363 # TODO(jbudorick): Handle most of this function in argparse.
364 if args.annotation_str:
365 args.annotations = args.annotation_str.split(',')
366 elif args.test_filter:
367 args.annotations = []
368 else:
369 args.annotations = ['Smoke', 'SmallTest', 'MediumTest', 'LargeTest',
370 'EnormousTest', 'IntegrationTest']
371
372 if args.exclude_annotation_str:
373 args.exclude_annotations = args.exclude_annotation_str.split(',')
374 else:
375 args.exclude_annotations = []
376
377
378def AddInstrumentationTestOptions(parser):
379 """Adds Instrumentation test options to |parser|."""
380
381 parser.usage = '%(prog)s [options]'
382
383 group = parser.add_argument_group('Instrumentation Test Options')
384 AddJavaTestOptions(group)
385
386 java_or_python_group = group.add_mutually_exclusive_group()
387 java_or_python_group.add_argument(
388 '-j', '--java-only', action='store_false',
389 dest='run_python_tests', default=True, help='Run only the Java tests.')
390 java_or_python_group.add_argument(
391 '-p', '--python-only', action='store_false',
392 dest='run_java_tests', default=True,
393 help='DEPRECATED')
394
395 group.add_argument('--host-driven-root',
396 help='DEPRECATED')
397 group.add_argument('-w', '--wait_debugger', dest='wait_for_debugger',
398 action='store_true',
399 help='Wait for debugger.')
400 group.add_argument('--apk-under-test',
401 help='Path or name of the apk under test.')
402 group.add_argument('--apk-under-test-incremental-install-script',
403 help='Path to install script for the --apk-under-test.')
404 group.add_argument('--test-apk', required=True,
405 help='Path or name of the apk containing the tests '
406 '(name is without the .apk extension; '
407 'e.g. "ContentShellTest").')
408 group.add_argument('--test-apk-incremental-install-script',
409 help='Path to install script for the --test-apk.')
410 group.add_argument('--additional-apk', action='append',
411 dest='additional_apks', default=[],
412 help='Additional apk that must be installed on '
413 'the device when the tests are run')
414 group.add_argument('--coverage-dir',
415 help=('Directory in which to place all generated '
416 'EMMA coverage files.'))
417 group.add_argument('--device-flags', dest='device_flags', default='',
418 help='The relative filepath to a file containing '
419 'command-line flags to set on the device')
420 group.add_argument('--device-flags-file', default='',
421 help='The relative filepath to a file containing '
422 'command-line flags to set on the device')
423 group.add_argument('--isolate_file_path',
424 '--isolate-file-path',
425 dest='isolate_file_path',
426 help='.isolate file path to override the default '
427 'path')
428 group.add_argument('--delete-stale-data', dest='delete_stale_data',
429 action='store_true',
430 help='Delete stale test data on the device.')
431 group.add_argument('--timeout-scale', type=float,
432 help='Factor by which timeouts should be scaled.')
433 group.add_argument('--strict-mode', dest='strict_mode', default='testing',
434 help='StrictMode command-line flag set on the device, '
435 'death/testing to kill the process, off to stop '
436 'checking, flash to flash only. Default testing.')
437
438 AddCommonOptions(parser)
439 AddDeviceOptions(parser)
440 AddRemoteDeviceOptions(parser)
441
442
443def AddJUnitTestOptions(parser):
444 """Adds junit test options to |parser|."""
445
446 group = parser.add_argument_group('JUnit Test Options')
447 group.add_argument(
448 '-s', '--test-suite', dest='test_suite', required=True,
449 help=('JUnit test suite to run.'))
450 group.add_argument(
451 '-f', '--test-filter', dest='test_filter',
452 help='Filters tests googletest-style.')
453 group.add_argument(
454 '--package-filter', dest='package_filter',
455 help='Filters tests by package.')
456 group.add_argument(
457 '--runner-filter', dest='runner_filter',
458 help='Filters tests by runner class. Must be fully qualified.')
459 group.add_argument(
460 '--sdk-version', dest='sdk_version', type=int,
461 help='The Android SDK version.')
462 AddCommonOptions(parser)
463
464
465def AddMonkeyTestOptions(parser):
466 """Adds monkey test options to |parser|."""
467
468 group = parser.add_argument_group('Monkey Test Options')
469 group.add_argument(
470 '--package', required=True, choices=constants.PACKAGE_INFO.keys(),
471 metavar='PACKAGE', help='Package under test.')
472 group.add_argument(
473 '--event-count', default=10000, type=int,
474 help='Number of events to generate (default: %(default)s).')
475 group.add_argument(
476 '--category', default='',
477 help='A list of allowed categories.')
478 group.add_argument(
479 '--throttle', default=100, type=int,
480 help='Delay between events (ms) (default: %(default)s). ')
481 group.add_argument(
482 '--seed', type=int,
483 help=('Seed value for pseudo-random generator. Same seed value generates '
484 'the same sequence of events. Seed is randomized by default.'))
485 group.add_argument(
486 '--extra-args', default='',
487 help=('String of other args to pass to the command verbatim.'))
488
489 AddCommonOptions(parser)
490 AddDeviceOptions(parser)
491
492def ProcessMonkeyTestOptions(args):
493 """Processes all monkey test options.
494
495 Args:
496 args: argparse.Namespace object.
497
498 Returns:
499 A MonkeyOptions named tuple which contains all options relevant to
500 monkey tests.
501 """
502 # TODO(jbudorick): Handle this directly in argparse with nargs='+'
503 category = args.category
504 if category:
505 category = args.category.split(',')
506
507 # TODO(jbudorick): Get rid of MonkeyOptions.
508 return monkey_test_options.MonkeyOptions(
509 args.verbose_count,
510 args.package,
511 args.event_count,
512 category,
513 args.throttle,
514 args.seed,
515 args.extra_args)
516
517def AddUirobotTestOptions(parser):
518 """Adds uirobot test options to |option_parser|."""
519 group = parser.add_argument_group('Uirobot Test Options')
520
521 group.add_argument('--app-under-test', required=True,
522 help='APK to run tests on.')
523 group.add_argument(
524 '--repeat', dest='repeat', type=int, default=0,
525 help='Number of times to repeat the uirobot test.')
526 group.add_argument(
527 '--minutes', default=5, type=int,
528 help='Number of minutes to run uirobot test [default: %(default)s].')
529
530 AddCommonOptions(parser)
531 AddDeviceOptions(parser)
532 AddRemoteDeviceOptions(parser)
533
534def AddPerfTestOptions(parser):
535 """Adds perf test options to |parser|."""
536
537 group = parser.add_argument_group('Perf Test Options')
538
539 class SingleStepAction(argparse.Action):
540 def __call__(self, parser, namespace, values, option_string=None):
541 if values and not namespace.single_step:
542 parser.error('single step command provided, '
543 'but --single-step not specified.')
544 elif namespace.single_step and not values:
545 parser.error('--single-step specified, '
546 'but no single step command provided.')
547 setattr(namespace, self.dest, values)
548
549 step_group = group.add_mutually_exclusive_group(required=True)
550 # TODO(jbudorick): Revise --single-step to use argparse.REMAINDER.
551 # This requires removing "--" from client calls.
552 step_group.add_argument(
553 '--single-step', action='store_true',
554 help='Execute the given command with retries, but only print the result '
555 'for the "most successful" round.')
556 step_group.add_argument(
557 '--steps',
558 help='JSON file containing the list of commands to run.')
559 step_group.add_argument(
560 '--print-step',
561 help='The name of a previously executed perf step to print.')
562
563 group.add_argument(
564 '--output-json-list',
565 help='Write a simple list of names from --steps into the given file.')
566 group.add_argument(
567 '--collect-chartjson-data',
568 action='store_true',
569 help='Cache the chartjson output from each step for later use.')
570 group.add_argument(
571 '--output-chartjson-data',
572 default='',
573 help='Write out chartjson into the given file.')
574 group.add_argument(
575 '--get-output-dir-archive', metavar='FILENAME',
576 help='Write the chached output directory archived by a step into the'
577 ' given ZIP file.')
578 group.add_argument(
579 '--flaky-steps',
580 help=('A JSON file containing steps that are flaky '
581 'and will have its exit code ignored.'))
582 group.add_argument(
583 '--no-timeout', action='store_true',
584 help=('Do not impose a timeout. Each perf step is responsible for '
585 'implementing the timeout logic.'))
586 group.add_argument(
587 '-f', '--test-filter',
588 help=('Test filter (will match against the names listed in --steps).'))
589 group.add_argument(
590 '--dry-run', action='store_true',
591 help='Just print the steps without executing.')
592 # Uses 0.1 degrees C because that's what Android does.
593 group.add_argument(
594 '--max-battery-temp', type=int,
595 help='Only start tests when the battery is at or below the given '
596 'temperature (0.1 C)')
597 group.add_argument('single_step_command', nargs='*', action=SingleStepAction,
598 help='If --single-step is specified, the command to run.')
599 group.add_argument('--min-battery-level', type=int,
600 help='Only starts tests when the battery is charged above '
601 'given level.')
602 group.add_argument('--known-devices-file', help='Path to known device list.')
603 AddCommonOptions(parser)
604 AddDeviceOptions(parser)
605
606
607def ProcessPerfTestOptions(args):
608 """Processes all perf test options.
609
610 Args:
611 args: argparse.Namespace object.
612
613 Returns:
614 A PerfOptions named tuple which contains all options relevant to
615 perf tests.
616 """
617 # TODO(jbudorick): Move single_step handling down into the perf tests.
618 if args.single_step:
619 args.single_step = ' '.join(args.single_step_command)
620 # TODO(jbudorick): Get rid of PerfOptions.
621 return perf_test_options.PerfOptions(
622 args.steps, args.flaky_steps, args.output_json_list,
623 args.print_step, args.no_timeout, args.test_filter,
624 args.dry_run, args.single_step, args.collect_chartjson_data,
625 args.output_chartjson_data, args.get_output_dir_archive,
626 args.max_battery_temp, args.min_battery_level,
627 args.known_devices_file)
628
629
630def AddPythonTestOptions(parser):
631 group = parser.add_argument_group('Python Test Options')
632 group.add_argument(
633 '-s', '--suite', dest='suite_name', metavar='SUITE_NAME',
634 choices=constants.PYTHON_UNIT_TEST_SUITES.keys(),
635 help='Name of the test suite to run.')
636 AddCommonOptions(parser)
637
638
639def _RunLinkerTests(args, devices):
640 """Subcommand of RunTestsCommands which runs linker tests."""
641 runner_factory, tests = linker_setup.Setup(args, devices)
642
643 results, exit_code = test_dispatcher.RunTests(
644 tests, runner_factory, devices, shard=True, test_timeout=60,
645 num_retries=args.num_retries)
646
647 report_results.LogFull(
648 results=results,
649 test_type='Linker test',
650 test_package='ChromiumLinkerTest')
651
652 if args.json_results_file:
653 json_results.GenerateJsonResultsFile([results], args.json_results_file)
654
655 return exit_code
656
657
658def _RunJUnitTests(args):
659 """Subcommand of RunTestsCommand which runs junit tests."""
660 runner_factory, tests = junit_setup.Setup(args)
661 results, exit_code = junit_dispatcher.RunTests(tests, runner_factory)
662
663 report_results.LogFull(
664 results=results,
665 test_type='JUnit',
666 test_package=args.test_suite)
667
668 if args.json_results_file:
669 json_results.GenerateJsonResultsFile([results], args.json_results_file)
670
671 return exit_code
672
673
674def _RunMonkeyTests(args, devices):
675 """Subcommand of RunTestsCommands which runs monkey tests."""
676 monkey_options = ProcessMonkeyTestOptions(args)
677
678 runner_factory, tests = monkey_setup.Setup(monkey_options)
679
680 results, exit_code = test_dispatcher.RunTests(
681 tests, runner_factory, devices, shard=False, test_timeout=None,
682 num_retries=args.num_retries)
683
684 report_results.LogFull(
685 results=results,
686 test_type='Monkey',
687 test_package='Monkey')
688
689 if args.json_results_file:
690 json_results.GenerateJsonResultsFile([results], args.json_results_file)
691
692 return exit_code
693
694
695def _RunPerfTests(args, active_devices):
696 """Subcommand of RunTestsCommands which runs perf tests."""
697 perf_options = ProcessPerfTestOptions(args)
698
699 # Just save a simple json with a list of test names.
700 if perf_options.output_json_list:
701 return perf_test_runner.OutputJsonList(
702 perf_options.steps, perf_options.output_json_list)
703
704 # Just print the results from a single previously executed step.
705 if perf_options.print_step:
706 return perf_test_runner.PrintTestOutput(
707 perf_options.print_step, perf_options.output_chartjson_data,
708 perf_options.get_output_dir_archive)
709
710 runner_factory, tests, devices = perf_setup.Setup(
711 perf_options, active_devices)
712
713 # shard=False means that each device will get the full list of tests
714 # and then each one will decide their own affinity.
715 # shard=True means each device will pop the next test available from a queue,
716 # which increases throughput but have no affinity.
717 results, _ = test_dispatcher.RunTests(
718 tests, runner_factory, devices, shard=False, test_timeout=None,
719 num_retries=args.num_retries)
720
721 report_results.LogFull(
722 results=results,
723 test_type='Perf',
724 test_package='Perf')
725
726 if args.json_results_file:
727 json_results.GenerateJsonResultsFile([results], args.json_results_file)
728
729 if perf_options.single_step:
730 return perf_test_runner.PrintTestOutput('single_step')
731
732 perf_test_runner.PrintSummary(tests)
733
734 # Always return 0 on the sharding stage. Individual tests exit_code
735 # will be returned on the print_step stage.
736 return 0
737
738
739def _RunPythonTests(args):
740 """Subcommand of RunTestsCommand which runs python unit tests."""
741 suite_vars = constants.PYTHON_UNIT_TEST_SUITES[args.suite_name]
742 suite_path = suite_vars['path']
743 suite_test_modules = suite_vars['test_modules']
744
745 sys.path = [suite_path] + sys.path
746 try:
747 suite = unittest.TestSuite()
748 suite.addTests(unittest.defaultTestLoader.loadTestsFromName(m)
749 for m in suite_test_modules)
750 runner = unittest.TextTestRunner(verbosity=1+args.verbose_count)
751 return 0 if runner.run(suite).wasSuccessful() else 1
752 finally:
753 sys.path = sys.path[1:]
754
755
756def _GetAttachedDevices(blacklist_file, test_device, enable_cache, num_retries):
757 """Get all attached devices.
758
759 Args:
760 blacklist_file: Path to device blacklist.
761 test_device: Name of a specific device to use.
762 enable_cache: Whether to enable checksum caching.
763
764 Returns:
765 A list of attached devices.
766 """
767 blacklist = (device_blacklist.Blacklist(blacklist_file)
768 if blacklist_file
769 else None)
770
771 attached_devices = device_utils.DeviceUtils.HealthyDevices(
772 blacklist, enable_device_files_cache=enable_cache,
773 default_retries=num_retries)
774 if test_device:
775 test_device = [d for d in attached_devices if d == test_device]
776 if not test_device:
777 raise device_errors.DeviceUnreachableError(
778 'Did not find device %s among attached device. Attached devices: %s'
779 % (test_device, ', '.join(attached_devices)))
780 return test_device
781
782 else:
783 if not attached_devices:
784 raise device_errors.NoDevicesError()
785 return sorted(attached_devices)
786
787
788def RunTestsCommand(args): # pylint: disable=too-many-return-statements
789 """Checks test type and dispatches to the appropriate function.
790
791 Args:
792 args: argparse.Namespace object.
793
794 Returns:
795 Integer indicated exit code.
796
797 Raises:
798 Exception: Unknown command name passed in, or an exception from an
799 individual test runner.
800 """
801 command = args.command
802
803 ProcessCommonOptions(args)
804 logging.info('command: %s', ' '.join(sys.argv))
805
806 if args.enable_platform_mode or command in ('gtest', 'instrumentation'):
807 return RunTestsInPlatformMode(args)
808
809 forwarder.Forwarder.RemoveHostLog()
810 if not ports.ResetTestServerPortAllocation():
811 raise Exception('Failed to reset test server port.')
812
813 def get_devices():
814 return _GetAttachedDevices(args.blacklist_file, args.test_device,
815 args.enable_device_cache, args.num_retries)
816
817 if command == 'linker':
818 return _RunLinkerTests(args, get_devices())
819 elif command == 'junit':
820 return _RunJUnitTests(args)
821 elif command == 'monkey':
822 return _RunMonkeyTests(args, get_devices())
823 elif command == 'perf':
824 return _RunPerfTests(args, get_devices())
825 elif command == 'python':
826 return _RunPythonTests(args)
827 else:
828 raise Exception('Unknown test type.')
829
830
831_SUPPORTED_IN_PLATFORM_MODE = [
832 # TODO(jbudorick): Add support for more test types.
833 'gtest',
834 'instrumentation',
835 'uirobot',
836]
837
838
839def RunTestsInPlatformMode(args):
840
841 def infra_error(message):
842 logging.fatal(message)
843 sys.exit(constants.INFRA_EXIT_CODE)
844
845 if args.command not in _SUPPORTED_IN_PLATFORM_MODE:
846 infra_error('%s is not yet supported in platform mode' % args.command)
847
848 with environment_factory.CreateEnvironment(args, infra_error) as env:
849 with test_instance_factory.CreateTestInstance(args, infra_error) as test:
850 with test_run_factory.CreateTestRun(
851 args, env, test, infra_error) as test_run:
852 results = []
853 repetitions = (xrange(args.repeat + 1) if args.repeat >= 0
854 else itertools.count())
855 result_counts = collections.defaultdict(
856 lambda: collections.defaultdict(int))
857 iteration_count = 0
858 for _ in repetitions:
859 iteration_results = test_run.RunTests()
860 if iteration_results is not None:
861 iteration_count += 1
862 results.append(iteration_results)
863 for r in iteration_results.GetAll():
864 result_counts[r.GetName()][r.GetType()] += 1
865 report_results.LogFull(
866 results=iteration_results,
867 test_type=test.TestType(),
868 test_package=test_run.TestPackage(),
869 annotation=getattr(args, 'annotations', None),
870 flakiness_server=getattr(args, 'flakiness_dashboard_server',
871 None))
872 if args.break_on_failure and not iteration_results.DidRunPass():
873 break
874
875 if iteration_count > 1:
876 # display summary results
877 # only display results for a test if at least one test did not pass
878 all_pass = 0
879 tot_tests = 0
880 for test_name in result_counts:
881 tot_tests += 1
882 if any(result_counts[test_name][x] for x in (
883 base_test_result.ResultType.FAIL,
884 base_test_result.ResultType.CRASH,
885 base_test_result.ResultType.TIMEOUT,
886 base_test_result.ResultType.UNKNOWN)):
887 logging.critical(
888 '%s: %s',
889 test_name,
890 ', '.join('%s %s' % (str(result_counts[test_name][i]), i)
891 for i in base_test_result.ResultType.GetTypes()))
892 else:
893 all_pass += 1
894
895 logging.critical('%s of %s tests passed in all %s runs',
896 str(all_pass),
897 str(tot_tests),
898 str(iteration_count))
899
900 if args.json_results_file:
901 json_results.GenerateJsonResultsFile(
902 results, args.json_results_file)
903
904 return (0 if all(r.DidRunPass() for r in results)
905 else constants.ERROR_EXIT_CODE)
906
907
908CommandConfigTuple = collections.namedtuple(
909 'CommandConfigTuple',
910 ['add_options_func', 'help_txt'])
911VALID_COMMANDS = {
912 'gtest': CommandConfigTuple(
913 AddGTestOptions,
914 'googletest-based C++ tests'),
915 'instrumentation': CommandConfigTuple(
916 AddInstrumentationTestOptions,
917 'InstrumentationTestCase-based Java tests'),
918 'junit': CommandConfigTuple(
919 AddJUnitTestOptions,
920 'JUnit4-based Java tests'),
921 'monkey': CommandConfigTuple(
922 AddMonkeyTestOptions,
923 "Tests based on Android's monkey"),
924 'perf': CommandConfigTuple(
925 AddPerfTestOptions,
926 'Performance tests'),
927 'python': CommandConfigTuple(
928 AddPythonTestOptions,
929 'Python tests based on unittest.TestCase'),
930 'linker': CommandConfigTuple(
931 AddLinkerTestOptions,
932 'Linker tests'),
933 'uirobot': CommandConfigTuple(
934 AddUirobotTestOptions,
935 'Uirobot test'),
936}
937
938
939def DumpThreadStacks(_signal, _frame):
940 for thread in threading.enumerate():
941 reraiser_thread.LogThreadStack(thread)
942
943
944def main():
945 signal.signal(signal.SIGUSR1, DumpThreadStacks)
946
947 parser = argparse.ArgumentParser()
948 command_parsers = parser.add_subparsers(title='test types',
949 dest='command')
950
951 for test_type, config in sorted(VALID_COMMANDS.iteritems(),
952 key=lambda x: x[0]):
953 subparser = command_parsers.add_parser(
954 test_type, usage='%(prog)s [options]', help=config.help_txt)
955 config.add_options_func(subparser)
956
957 args = parser.parse_args()
958
959 try:
960 return RunTestsCommand(args)
961 except base_error.BaseError as e:
962 logging.exception('Error occurred.')
963 if e.is_infra_error:
964 return constants.INFRA_EXIT_CODE
965 return constants.ERROR_EXIT_CODE
966 except: # pylint: disable=W0702
967 logging.exception('Unrecognized error occurred.')
968 return constants.ERROR_EXIT_CODE
969
970
971if __name__ == '__main__':
972 sys.exit(main())