| # |
| # Copyright 2015 Google Inc. |
| # |
| # Use of this source code is governed by a BSD-style license that can be |
| # found in the LICENSE file. |
| # |
| |
| #!/usr/bin/env python |
| |
| usage = ''' |
| Write extra flags to outfile for nanobench based on the bot name: |
| $ python nanobench_flags.py outfile Perf-Android-GCC-GalaxyS3-GPU-Mali400-Arm7-Release |
| Or run self-tests: |
| $ python nanobench_flags.py test |
| ''' |
| |
| import inspect |
| import json |
| import os |
| import sys |
| |
| |
| def lineno(): |
| caller = inspect.stack()[1] # Up one level to our caller. |
| return inspect.getframeinfo(caller[0]).lineno |
| |
| |
| cov_start = lineno()+1 # We care about coverage starting just past this def. |
| def get_args(bot): |
| args = ['--pre_log'] |
| |
| if 'GPU' in bot: |
| args.append('--images') |
| args.extend(['--gpuStatsDump', 'true']) |
| |
| if 'Android' in bot and 'GPU' in bot: |
| args.extend(['--useThermalManager', '1,1,10,1000']) |
| |
| if 'Appurify' not in bot: |
| args.extend(['--scales', '1.0', '1.1']) |
| |
| if 'iOS' in bot: |
| args.extend(['--skps', 'ignore_skps']) |
| |
| if 'Appurify' not in bot: |
| config = ['565', '8888', 'gpu', 'nonrendering', 'angle', 'hwui'] |
| config.extend(['srgb', 'f16']) |
| # The S4 crashes and the NP produces a long error stream when we run with |
| # MSAA. |
| if ('GalaxyS4' not in bot and |
| 'NexusPlayer' not in bot): |
| if 'Android' in bot: |
| config.extend(['msaa4', 'nvprmsaa4']) |
| else: |
| config.extend(['msaa16', 'nvprmsaa16']) |
| args.append('--config') |
| args.extend(config) |
| |
| if 'Valgrind' in bot: |
| # Don't care about Valgrind performance. |
| args.extend(['--loops', '1']) |
| args.extend(['--samples', '1']) |
| # Ensure that the bot framework does not think we have timed out. |
| args.extend(['--keepAlive', 'true']) |
| |
| if 'HD2000' in bot: |
| args.extend(['--GPUbenchTileW', '256']) |
| args.extend(['--GPUbenchTileH', '256']) |
| |
| match = [] |
| if 'Android' in bot: |
| # Segfaults when run as GPU bench. Very large texture? |
| match.append('~blurroundrect') |
| match.append('~patch_grid') # skia:2847 |
| match.append('~desk_carsvg') |
| if 'HD2000' in bot: |
| match.extend(['~gradient', '~etc1bitmap']) # skia:2895 |
| if 'NexusPlayer' in bot: |
| match.append('~desk_unicodetable') |
| if 'GalaxyS4' in bot: |
| match.append('~GLInstancedArraysBench') # skia:4371 |
| |
| if 'iOS' in bot: |
| match.append('~blurroundrect') |
| match.append('~patch_grid') # skia:2847 |
| match.append('~desk_carsvg') |
| match.append('~keymobi') |
| match.append('~path_hairline') |
| match.append('~GLInstancedArraysBench') # skia:4714 |
| |
| # the 32-bit GCE bots run out of memory in DM when running these large images |
| # so defensively disable them in nanobench, too. |
| # FIXME (scroggo): This may have just been due to SkImageDecoder's |
| # buildTileIndex leaking memory (https://bug.skia.org/4360). That is |
| # disabled by default for nanobench, so we may not need this. |
| # FIXME (scroggo): Share image blacklists between dm and nanobench? |
| if 'x86' in bot and not 'x86-64' in bot: |
| match.append('~interlaced1.png') |
| match.append('~interlaced2.png') |
| match.append('~interlaced3.png') |
| |
| # We do not need or want to benchmark the decodes of incomplete images. |
| # In fact, in nanobench we assert that the full image decode succeeds. |
| match.append('~inc0.gif') |
| match.append('~inc1.gif') |
| match.append('~incInterlaced.gif') |
| match.append('~inc0.jpg') |
| match.append('~incGray.jpg') |
| match.append('~inc0.wbmp') |
| match.append('~inc1.wbmp') |
| match.append('~inc0.webp') |
| match.append('~inc1.webp') |
| match.append('~inc0.ico') |
| match.append('~inc1.ico') |
| match.append('~inc0.png') |
| match.append('~inc1.png') |
| match.append('~inc2.png') |
| match.append('~inc12.png') |
| match.append('~inc13.png') |
| match.append('~inc14.png') |
| match.append('~inc0.webp') |
| match.append('~inc1.webp') |
| |
| # As an experiment, skip nanobench on Debug trybots. |
| if 'Debug' in bot and 'CPU' in bot and 'Trybot' in bot: |
| match = ['nothing_will_match_this'] |
| |
| if match: |
| args.append('--match') |
| args.extend(match) |
| |
| return args |
| cov_end = lineno() # Don't care about code coverage past here. |
| |
| |
| def self_test(): |
| import coverage # This way the bots don't need coverage.py to be installed. |
| args = {} |
| cases = [ |
| 'Perf-Android-GCC-Nexus6-GPU-Adreno420-Arm7-Release', |
| 'Perf-Android-Nexus7-Tegra3-Arm7-Release', |
| 'Perf-Android-GCC-NexusPlayer-GPU-PowerVR-x86-Release', |
| 'Test-Ubuntu-GCC-ShuttleA-GPU-GTX550Ti-x86_64-Release-Valgrind', |
| 'Test-Win7-MSVC-ShuttleA-GPU-HD2000-x86-Debug-ANGLE', |
| 'Test-iOS-Clang-iPad4-GPU-SGX554-Arm7-Debug', |
| 'Test-Android-GCC-GalaxyS4-GPU-SGX544-Arm7-Release', |
| 'Test-Ubuntu-GCC-GCE-CPU-AVX2-x86_64-Debug-Trybot', |
| ] |
| |
| cov = coverage.coverage() |
| cov.start() |
| for case in cases: |
| args[case] = get_args(case) |
| cov.stop() |
| |
| this_file = os.path.basename(__file__) |
| _, _, not_run, _ = cov.analysis(this_file) |
| filtered = [line for line in not_run if line > cov_start and line < cov_end] |
| if filtered: |
| print 'Lines not covered by test cases: ', filtered |
| sys.exit(1) |
| |
| golden = this_file.replace('.py', '.json') |
| with open(os.path.join(os.path.dirname(__file__), golden), 'w') as f: |
| json.dump(args, f, indent=2, sort_keys=True) |
| |
| |
| if __name__ == '__main__': |
| if len(sys.argv) == 2 and sys.argv[1] == 'test': |
| self_test() |
| sys.exit(0) |
| |
| if len(sys.argv) != 3: |
| print usage |
| sys.exit(1) |
| |
| with open(sys.argv[1], 'w') as out: |
| json.dump(get_args(sys.argv[2]), out) |