| #!/usr/bin/env python3 | 
 | # Copyright (c) 2015-2016 The Khronos Group Inc. | 
 | # Copyright (c) 2015-2016 Valve Corporation | 
 | # Copyright (c) 2015-2016 LunarG, Inc. | 
 | # Copyright (c) 2015-2016 Google Inc. | 
 | # | 
 | # Licensed under the Apache License, Version 2.0 (the "License"); | 
 | # you may not use this file except in compliance with the License. | 
 | # You may obtain a copy of the License at | 
 | # | 
 | #     http://www.apache.org/licenses/LICENSE-2.0 | 
 | # | 
 | # Unless required by applicable law or agreed to in writing, software | 
 | # distributed under the License is distributed on an "AS IS" BASIS, | 
 | # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
 | # See the License for the specific language governing permissions and | 
 | # limitations under the License. | 
 | # | 
 | # Author: Tobin Ehlis <tobin@lunarg.com> | 
 |  | 
 | import argparse | 
 | import os | 
 | import sys | 
 | import vulkan | 
 | import platform | 
 |  | 
 | # vk_layer_documentation_generate.py overview | 
 | # This script is intended to generate documentation based on vulkan layers | 
 | #  It parses known validation layer headers for details of the validation checks | 
 | #  It parses validation layer source files for specific code where checks are implemented | 
 | #  structs in a human-readable txt format, as well as utility functions | 
 | #  to print enum values as strings | 
 |  | 
 | # NOTE : Initially the script is performing validation of a hand-written document | 
 | #  Right now it does 3 checks: | 
 | #  1. Verify ENUM codes declared in source are documented | 
 | #  2. Verify ENUM codes in document are declared in source | 
 | #  3. Verify API function names in document are in the actual API header (vulkan.py) | 
 | # Currently script will flag errors in all of these cases | 
 |  | 
 | # TODO : Need a formal specification of the syntax for doc generation | 
 | #  Initially, these are the basics: | 
 | #  1. Validation checks have unique ENUM values defined in validation layer header | 
 | #  2. ENUM includes comments for 1-line overview of check and more detailed description | 
 | #  3. Actual code implementing checks includes ENUM value in callback | 
 | #  4. Code to test checks should include reference to ENUM | 
 |  | 
 |  | 
 | # TODO : Need list of known validation layers to use as default input | 
 | #  Just a couple of flat lists right now, but may need to make this input file | 
 | #  or at least a more dynamic data structure | 
 | layer_inputs = { 'draw_state' : {'header' : 'layers/core_validation_error_enums.h', | 
 |                                  'source' : 'layers/core_validation.cpp', | 
 |                                  'generated' : False, | 
 |                                  'error_enum' : 'DRAW_STATE_ERROR'}, | 
 |                  'shader_checker' : {'header' : 'layers/core_validation_error_enums.h', | 
 |                                  'source' : 'layers/core_validation.cpp', | 
 |                                  'generated' : False, | 
 |                                  'error_enum' : 'SHADER_CHECKER_ERROR'}, | 
 |                  'mem_tracker' : {'header' : 'layers/core_validation_error_enums.h', | 
 |                                   'source' : 'layers/core_validation.cpp', | 
 |                                   'generated' : False, | 
 |                                   'error_enum' : 'MEM_TRACK_ERROR'}, | 
 |                  'threading' : {'header' : 'layers/threading.h', | 
 |                                 'source' : 'dbuild/layers/threading.cpp', | 
 |                                 'generated' : True, | 
 |                                 'error_enum' : 'THREADING_CHECKER_ERROR'}, | 
 |                  'object_tracker' : {'header' : 'layers/object_tracker.h', | 
 |                                 'source' : 'dbuild/layers/object_tracker.cpp', | 
 |                                 'generated' : True, | 
 |                                 'error_enum' : 'OBJECT_TRACK_ERROR',}, | 
 |                  'device_limits' : {'header' : 'layers/device_limits.h', | 
 |                                     'source' : 'layers/device_limits.cpp', | 
 |                                     'generated' : False, | 
 |                                     'error_enum' : 'DEV_LIMITS_ERROR',}, | 
 |                  'image' : {'header' : 'layers/image.h', | 
 |                             'source' : 'layers/image.cpp', | 
 |                             'generated' : False, | 
 |                             'error_enum' : 'IMAGE_ERROR',}, | 
 |                  'swapchain' : {'header' : 'layers/swapchain.h', | 
 |                             'source' : 'layers/swapchain.cpp', | 
 |                             'generated' : False, | 
 |                             'error_enum' : 'SWAPCHAIN_ERROR',}, | 
 |                  'parameter_validation' : {'header' : 'layers/parameter_validation_utils.h', | 
 |                                            'source' : 'layers/parameter_validation.cpp', | 
 |                                            'generated' : False, | 
 |                                            'error_enum' : 'ErrorCode',}, | 
 |     } | 
 |  | 
 | builtin_headers = [layer_inputs[ln]['header'] for ln in layer_inputs] | 
 | builtin_source = [layer_inputs[ln]['source'] for ln in layer_inputs] | 
 | builtin_tests = ['tests/layer_validation_tests.cpp', ] | 
 |  | 
 | # List of extensions in layers that are included in documentation, but not in vulkan.py API set | 
 | layer_extension_functions = ['objTrackGetObjects', 'objTrackGetObjectsOfType'] | 
 |  | 
 | def handle_args(): | 
 |     parser = argparse.ArgumentParser(description='Generate layer documenation from source.') | 
 |     parser.add_argument('--in_headers', required=False, default=builtin_headers, help='The input layer header files from which code will be generated.') | 
 |     parser.add_argument('--in_source', required=False, default=builtin_source, help='The input layer source files from which code will be generated.') | 
 |     parser.add_argument('--test_source', required=False, default=builtin_tests, help='The input test source files from which code will be generated.') | 
 |     parser.add_argument('--layer_doc', required=False, default='layers/vk_validation_layer_details.md', help='Existing layer document to be validated against actual layers.') | 
 |     parser.add_argument('--validate', action='store_true', default=False, help='Validate that there are no mismatches between layer documentation and source. This includes cross-checking the validation checks, and making sure documented Vulkan API calls exist.') | 
 |     parser.add_argument('--print_structs', action='store_true', default=False, help='Primarily a debug option that prints out internal data structs used to generate layer docs.') | 
 |     parser.add_argument('--print_doc_checks', action='store_true', default=False, help='Primarily a debug option that prints out all of the checks that are documented.') | 
 |     return parser.parse_args() | 
 |  | 
 | # Little helper class for coloring cmd line output | 
 | class bcolors: | 
 |  | 
 |     def __init__(self): | 
 |         self.GREEN = '\033[0;32m' | 
 |         self.RED = '\033[0;31m' | 
 |         self.YELLOW = '\033[1;33m' | 
 |         self.ENDC = '\033[0m' | 
 |         if 'Linux' != platform.system(): | 
 |             self.GREEN = '' | 
 |             self.RED = '' | 
 |             self.YELLOW = '' | 
 |             self.ENDC = '' | 
 |  | 
 |     def green(self): | 
 |         return self.GREEN | 
 |  | 
 |     def red(self): | 
 |         return self.RED | 
 |  | 
 |     def yellow(self): | 
 |         return self.YELLOW | 
 |  | 
 |     def endc(self): | 
 |         return self.ENDC | 
 |  | 
 | # Class to parse the validation layer test source and store testnames | 
 | class TestParser: | 
 |     def __init__(self, test_file_list, test_group_name=['VkLayerTest', 'VkWsiEnabledLayerTest']): | 
 |         self.test_files = test_file_list | 
 |         self.tests_set = set() | 
 |         self.test_trigger_txt_list = [] | 
 |         for tg in test_group_name: | 
 |             self.test_trigger_txt_list.append('TEST_F(%s' % tg) | 
 |             #print('Test trigger test list: %s' % (self.test_trigger_txt_list)) | 
 |  | 
 |     # Parse test files into internal data struct | 
 |     def parse(self): | 
 |         # For each test file, parse test names into set | 
 |         grab_next_line = False # handle testname on separate line than wildcard | 
 |         for test_file in self.test_files: | 
 |             with open(test_file) as tf: | 
 |                 for line in tf: | 
 |                     if True in [line.strip().startswith(comment) for comment in ['//', '/*']]: | 
 |                         continue | 
 |  | 
 |                     if True in [ttt in line for ttt in self.test_trigger_txt_list]: | 
 |                         #print('Test wildcard in line: %s' % (line)) | 
 |                         testname = line.split(',')[-1] | 
 |                         testname = testname.strip().strip(' {)') | 
 |                         #print('Inserting test: "%s"' % (testname)) | 
 |                         if ('' == testname): | 
 |                             grab_next_line = True | 
 |                             continue | 
 |                         self.tests_set.add(testname) | 
 |                     if grab_next_line: # test name on its own line | 
 |                         grab_next_line = False | 
 |                         testname = testname.strip().strip(' {)') | 
 |                         self.tests_set.add(testname) | 
 |  | 
 | # Class to parse the layer source code and store details in internal data structs | 
 | class LayerParser: | 
 |     def __init__(self, header_file_list, source_file_list): | 
 |         self.header_files = header_file_list | 
 |         self.source_files = source_file_list | 
 |         self.layer_dict = {} | 
 |         self.api_dict = {} | 
 |  | 
 |     # Parse layer header files into internal dict data structs | 
 |     def parse(self): | 
 |         # For each header file, parse details into dicts | 
 |         # TODO : Should have a global dict element to track overall list of checks | 
 |         store_enum = False | 
 |         for layer_name in layer_inputs: | 
 |             hf = layer_inputs[layer_name]['header'] | 
 |             self.layer_dict[layer_name] = {} # initialize a new dict for this layer | 
 |             self.layer_dict[layer_name]['CHECKS'] = [] # enum of checks is stored in a list | 
 |             #print('Parsing header file %s as layer name %s' % (hf, layer_name)) | 
 |             with open(hf) as f: | 
 |                 for line in f: | 
 |                     if True in [line.strip().startswith(comment) for comment in ['//', '/*']]: | 
 |                         #print("Skipping comment line: %s" % line) | 
 |                         # For now skipping lines starting w/ comment, may use these to capture | 
 |                         #  documentation in the future | 
 |                         continue | 
 |                     # Find enums | 
 |                     if store_enum: | 
 |                         if '}' in line: # we're done with enum definition | 
 |                             store_enum = False | 
 |                             continue | 
 |                         # grab the enum name as a unique check | 
 |                         if ',' in line: | 
 |                             # TODO : When documentation for a check is contained in the source, | 
 |                             #  this is where we should also capture that documentation so that | 
 |                             #  it can then be transformed into desired doc format | 
 |                             enum_name = line.split(',')[0].strip() | 
 |                             # Flag an error if we have already seen this enum | 
 |                             if enum_name in self.layer_dict[layer_name]['CHECKS']: | 
 |                                 print('ERROR : % layer has duplicate error enum: %s' % (layer_name, enum_name)) | 
 |                             self.layer_dict[layer_name]['CHECKS'].append(enum_name) | 
 |                     # If the line includes 'enum' and the expected enum name, start capturing enums | 
 |                     if False not in [ex in line for ex in ['enum', layer_inputs[layer_name]['error_enum']]]: | 
 |                         store_enum = True | 
 |  | 
 |         # For each source file, parse into dicts | 
 |         for sf in self.source_files: | 
 |             #print('Parsing source file %s' % sf) | 
 |             pass | 
 |             # TODO : In the source file we want to see where checks actually occur | 
 |             #  Need to build function tree of checks so that we know all of the | 
 |             #  checks that occur under a top-level Vulkan API call | 
 |             #  Eventually in the validation we can flag ENUMs that aren't being | 
 |             #  used in the source, and we can document source code lines as well | 
 |             #  as Vulkan API calls where each specific ENUM check is made | 
 |  | 
 |     def print_structs(self): | 
 |         print('This is where I print the data structs') | 
 |         for layer in self.layer_dict: | 
 |             print('Layer %s has %i checks:\n%s' % (layer, len(self.layer_dict[layer]['CHECKS'])-1, "\n\t".join(self.layer_dict[layer]['CHECKS']))) | 
 |  | 
 | # Class to parse hand-written md layer documentation into a dict and then validate its contents | 
 | class LayerDoc: | 
 |     def __init__(self, source_file): | 
 |         self.layer_doc_filename = source_file | 
 |         self.txt_color = bcolors() | 
 |         # Main data struct to store info from layer doc | 
 |         self.layer_doc_dict = {} | 
 |         # Comprehensive list of all validation checks recorded in doc | 
 |         self.enum_list = [] | 
 |  | 
 |     # Parse the contents of doc into data struct | 
 |     def parse(self): | 
 |         layer_name = 'INIT' | 
 |         parse_layer_details = False | 
 |         detail_trigger = '| Check | ' | 
 |         parse_pending_work = False | 
 |         pending_trigger = ' Pending Work' | 
 |         parse_overview = False | 
 |         overview_trigger = ' Overview' | 
 |         enum_prefix = '' | 
 |  | 
 |         with open(self.layer_doc_filename) as f: | 
 |             for line in f: | 
 |                 if parse_pending_work: | 
 |                     if '.' in line and line.strip()[0].isdigit(): | 
 |                         todo_item = line.split('.')[1].strip() | 
 |                         self.layer_doc_dict[layer_name]['pending'].append(todo_item) | 
 |                 if pending_trigger in line and '##' in line: | 
 |                     parse_layer_details = False | 
 |                     parse_pending_work = True | 
 |                     parse_overview = False | 
 |                     self.layer_doc_dict[layer_name]['pending'] = [] | 
 |                 if parse_layer_details: | 
 |                     # Grab details but skip the fomat line with a bunch of '-' chars | 
 |                     if '|' in line and line.count('-') < 20: | 
 |                         detail_sections = line.split('|') | 
 |                         #print("Details elements from line %s: %s" % (line, detail_sections)) | 
 |                         check_name = '%s%s' % (enum_prefix, detail_sections[3].strip()) | 
 |                         if '_NA' in check_name: | 
 |                             # TODO : Should clean up these NA checks in the doc, skipping them for now | 
 |                             continue | 
 |                         self.enum_list.append(check_name) | 
 |                         self.layer_doc_dict[layer_name][check_name] = {} | 
 |                         self.layer_doc_dict[layer_name][check_name]['summary_txt'] = detail_sections[1].strip() | 
 |                         self.layer_doc_dict[layer_name][check_name]['details_txt'] = detail_sections[2].strip() | 
 |                         self.layer_doc_dict[layer_name][check_name]['api_list'] = detail_sections[4].split() | 
 |                         self.layer_doc_dict[layer_name][check_name]['tests'] = detail_sections[5].split() | 
 |                         self.layer_doc_dict[layer_name][check_name]['notes'] = detail_sections[6].strip() | 
 |                         # strip any unwanted commas from api and test names | 
 |                         self.layer_doc_dict[layer_name][check_name]['api_list'] = [a.strip(',') for a in self.layer_doc_dict[layer_name][check_name]['api_list']] | 
 |                         test_list = [a.strip(',') for a in self.layer_doc_dict[layer_name][check_name]['tests']] | 
 |                         self.layer_doc_dict[layer_name][check_name]['tests'] = [a.split('.')[-1] for a in test_list] | 
 |                 # Trigger details parsing when we have table header | 
 |                 if detail_trigger in line: | 
 |                     parse_layer_details = True | 
 |                     parse_pending_work = False | 
 |                     parse_overview = False | 
 |                     enum_txt = line.split('|')[3] | 
 |                     if '*' in enum_txt: | 
 |                         enum_prefix = enum_txt.split()[-1].strip('*').strip() | 
 |                         #print('prefix: %s' % enum_prefix) | 
 |                 if parse_overview: | 
 |                     self.layer_doc_dict[layer_name]['overview'] += line | 
 |                 if overview_trigger in line and '##' in line: | 
 |                     parse_layer_details = False | 
 |                     parse_pending_work = False | 
 |                     parse_overview = True | 
 |                     layer_name = line.split()[1] | 
 |                     self.layer_doc_dict[layer_name] = {} | 
 |                     self.layer_doc_dict[layer_name]['overview'] = '' | 
 |  | 
 |     # Verify that checks, tests and api references in layer doc match reality | 
 |     #  Report API calls from doc that are not found in API | 
 |     #  Report checks from doc that are not in actual layers | 
 |     #  Report checks from layers that are not captured in doc | 
 |     #  Report checks from doc that do not have a valid test | 
 |     def validate(self, layer_dict, tests_set): | 
 |         #print("tests_set: %s" % (tests_set)) | 
 |         # Count number of errors found and return it | 
 |         errors_found = 0 | 
 |         warnings_found = 0 | 
 |         # First we'll go through the doc datastructures and flag any issues | 
 |         for chk in self.enum_list: | 
 |             doc_layer_found = False | 
 |             for real_layer in layer_dict: | 
 |                 if chk in layer_dict[real_layer]['CHECKS']: | 
 |                     #print('Found actual layer check %s in doc' % (chk)) | 
 |                     doc_layer_found = True | 
 |                     continue | 
 |             if not doc_layer_found: | 
 |                 print(self.txt_color.red() + 'Actual layers do not contain documented check: %s' % (chk) + self.txt_color.endc()) | 
 |                 errors_found += 1 | 
 |  | 
 |         # Now go through API names in doc and verify they're real | 
 |         # First we're going to transform proto names from vulkan.py into single list | 
 |         core_api_names = [p.name for p in vulkan.core.protos] | 
 |         wsi_s_names = [p.name for p in vulkan.ext_khr_surface.protos] | 
 |         wsi_ds_names = [p.name for p in vulkan.ext_khr_device_swapchain.protos] | 
 |         dbg_rpt_names = [p.name for p in vulkan.lunarg_debug_report.protos] | 
 |         api_names = core_api_names + wsi_s_names + wsi_ds_names + dbg_rpt_names | 
 |         for ln in self.layer_doc_dict: | 
 |             for chk in self.layer_doc_dict[ln]: | 
 |                 if chk in ['overview', 'pending']: | 
 |                     continue | 
 |                 for api in self.layer_doc_dict[ln][chk]['api_list']: | 
 |                     if api[2:] not in api_names and api not in layer_extension_functions: | 
 |                         print(self.txt_color.red() + 'Doc references invalid function: %s' % (api) + self.txt_color.endc()) | 
 |                         errors_found += 1 | 
 |                 # For now warn on missing or invalid tests | 
 |                 for test in self.layer_doc_dict[ln][chk]['tests']: | 
 |                     if '*' in test: | 
 |                         # naive way to handle wildcards, just make sure we have matches on parts | 
 |                         test_parts = test.split('*') | 
 |                         for part in test_parts: | 
 |                             part_found = False | 
 |                             for t in tests_set: | 
 |                                 if part in t: | 
 |                                     part_found = True | 
 |                                     break | 
 |                             if not part_found: | 
 |                                 print(self.txt_color.red() + 'Validation check %s has missing or invalid test : %s' % (chk, test)) | 
 |                                 errors_found += 1 | 
 |                                 break | 
 |                     elif test not in tests_set and not chk.endswith('_NONE'): | 
 |                         if test == 'TODO': | 
 |                             warnings_found += 1 | 
 |                         else: | 
 |                             print(self.txt_color.red() + 'Validation check %s has missing or invalid test : %s' % (chk, test)) | 
 |                             errors_found += 1 | 
 |         # Now go through all of the actual checks in the layers and make sure they're covered in the doc | 
 |         for ln in layer_dict: | 
 |             for chk in layer_dict[ln]['CHECKS']: | 
 |                 if chk not in self.enum_list: | 
 |                     print(self.txt_color.red() + 'Doc is missing check: %s' % (chk) + self.txt_color.endc()) | 
 |                     errors_found += 1 | 
 |  | 
 |         return (errors_found, warnings_found) | 
 |  | 
 |     # Print all of the checks captured in the doc | 
 |     def print_checks(self): | 
 |         print('Checks captured in doc:\n%s' % ('\n\t'.join(self.enum_list))) | 
 |  | 
 | def main(argv=None): | 
 |     # Parse args | 
 |     opts = handle_args() | 
 |     # Create parser for layer files | 
 |     layer_parser = LayerParser(opts.in_headers, opts.in_source) | 
 |     # Parse files into internal data structs | 
 |     layer_parser.parse() | 
 |     # Parse test files | 
 |     test_parser = TestParser(opts.test_source) | 
 |     test_parser.parse() | 
 |  | 
 |     # Generate requested types of output | 
 |     if opts.print_structs: # Print details of internal data structs | 
 |         layer_parser.print_structs() | 
 |  | 
 |     layer_doc = LayerDoc(opts.layer_doc) | 
 |     layer_doc.parse() | 
 |     if opts.print_doc_checks: | 
 |         layer_doc.print_checks() | 
 |  | 
 |     if opts.validate: | 
 |         (num_errors, num_warnings) = layer_doc.validate(layer_parser.layer_dict, test_parser.tests_set) | 
 |         txt_color = bcolors() | 
 |         if (0 == num_warnings): | 
 |             print(txt_color.green() + 'No warning cases found between %s and implementation' % (os.path.basename(opts.layer_doc)) + txt_color.endc()) | 
 |         else: | 
 |             print(txt_color.yellow() + 'Found %s warnings due to missing tests. Missing tests are labeled as "TODO" in "%s."' % (num_warnings, opts.layer_doc)) | 
 |         if (0 == num_errors): | 
 |             print(txt_color.green() + 'No mismatches found between %s and implementation' % (os.path.basename(opts.layer_doc)) + txt_color.endc()) | 
 |         else: | 
 |             return num_errors | 
 |     return 0 | 
 |  | 
 | if __name__ == "__main__": | 
 |     sys.exit(main()) | 
 |  |