blob: e72426eafbfcc149de9eaf9c6b3cc0e09a30f7f3 [file] [log] [blame]
Manuel Klimek95a58d22012-08-27 18:49:12 +00001#!/usr/bin/env python
2# A tool to parse ASTMatchers.h and update the documentation in
3# ../LibASTMatchersReference.html automatically. Run from the
4# directory in which this file is located to update the docs.
5
6import collections
7import re
8import urllib2
9
10MATCHERS_FILE = '../../include/clang/ASTMatchers/ASTMatchers.h'
11
12# Each matcher is documented in one row of the form:
13# result | name | argA
14# The subsequent row contains the documentation and is hidden by default,
15# becoming visible via javascript when the user clicks the matcher name.
16TD_TEMPLATE="""
Manuel Klimek67619ff2012-09-07 13:10:32 +000017<tr><td>%(result)s</td><td class="name" onclick="toggle('%(id)s')"><a name="%(id)sAnchor">%(name)s</a></td><td>%(args)s</td></tr>
Manuel Klimek95a58d22012-08-27 18:49:12 +000018<tr><td colspan="4" class="doc" id="%(id)s"><pre>%(comment)s</pre></td></tr>
19"""
20
21# We categorize the matchers into these three categories in the reference:
22node_matchers = {}
23narrowing_matchers = {}
24traversal_matchers = {}
25
26# We output multiple rows per matcher if the matcher can be used on multiple
27# node types. Thus, we need a new id per row to control the documentation
28# pop-up. ids[name] keeps track of those ids.
29ids = collections.defaultdict(int)
30
31# Cache for doxygen urls we have already verified.
32doxygen_probes = {}
33
34def esc(text):
35 """Escape any html in the given text."""
36 text = re.sub(r'&', '&amp;', text)
37 text = re.sub(r'<', '&lt;', text)
38 text = re.sub(r'>', '&gt;', text)
39 def link_if_exists(m):
40 name = m.group(1)
41 url = 'http://clang.llvm.org/doxygen/classclang_1_1%s.html' % name
42 if url not in doxygen_probes:
43 try:
44 print 'Probing %s...' % url
45 urllib2.urlopen(url)
46 doxygen_probes[url] = True
47 except:
48 doxygen_probes[url] = False
49 if doxygen_probes[url]:
50 return r'Matcher&lt<a href="%s">%s</a>&gt;' % (url, name)
51 else:
52 return m.group(0)
53 text = re.sub(
54 r'Matcher&lt;([^\*&]+)&gt;', link_if_exists, text)
55 return text
56
57def extract_result_types(comment):
58 """Extracts a list of result types from the given comment.
59
60 We allow annotations in the comment of the matcher to specify what
61 nodes a matcher can match on. Those comments have the form:
62 Usable as: Any Matcher | (Matcher<T1>[, Matcher<t2>[, ...]])
63
64 Returns ['*'] in case of 'Any Matcher', or ['T1', 'T2', ...].
65 Returns the empty list if no 'Usable as' specification could be
66 parsed.
67 """
68 result_types = []
69 m = re.search(r'Usable as: Any Matcher[\s\n]*$', comment, re.S)
70 if m:
71 return ['*']
72 while True:
73 m = re.match(r'^(.*)Matcher<([^>]+)>\s*,?[\s\n]*$', comment, re.S)
74 if not m:
75 if re.search(r'Usable as:\s*$', comment):
76 return result_types
77 else:
78 return None
79 result_types += [m.group(2)]
80 comment = m.group(1)
81
82def strip_doxygen(comment):
83 """Returns the given comment without \-escaped words."""
84 # If there is only a doxygen keyword in the line, delete the whole line.
85 comment = re.sub(r'^\\[^\s]+\n', r'', comment, flags=re.M)
86 # Delete the doxygen command and the following whitespace.
87 comment = re.sub(r'\\[^\s]+\s+', r'', comment)
88 return comment
89
90def unify_arguments(args):
91 """Gets rid of anything the user doesn't care about in the argument list."""
92 args = re.sub(r'internal::', r'', args)
93 args = re.sub(r'const\s+', r'', args)
94 args = re.sub(r'&', r' ', args)
95 args = re.sub(r'(^|\s)M\d?(\s)', r'\1Matcher<*>\2', args)
96 return args
97
98def add_matcher(result_type, name, args, comment, is_dyncast=False):
99 """Adds a matcher to one of our categories."""
100 if name == 'id':
101 # FIXME: Figure out whether we want to support the 'id' matcher.
102 return
103 matcher_id = '%s%d' % (name, ids[name])
104 ids[name] += 1
105 args = unify_arguments(args)
106 matcher_html = TD_TEMPLATE % {
107 'result': esc('Matcher<%s>' % result_type),
108 'name': name,
109 'args': esc(args),
110 'comment': esc(strip_doxygen(comment)),
111 'id': matcher_id,
112 }
113 if is_dyncast:
114 node_matchers[result_type + name] = matcher_html
115 # Use a heuristic to figure out whether a matcher is a narrowing or
116 # traversal matcher. By default, matchers that take other matchers as
117 # arguments (and are not node matchers) do traversal. We specifically
118 # exclude known narrowing matchers that also take other matchers as
119 # arguments.
120 elif ('Matcher<' not in args or
121 name in ['allOf', 'anyOf', 'anything', 'unless']):
122 narrowing_matchers[result_type + name] = matcher_html
123 else:
124 traversal_matchers[result_type + name] = matcher_html
125
126def act_on_decl(declaration, comment, allowed_types):
127 """Parse the matcher out of the given declaration and comment.
128
129 If 'allowed_types' is set, it contains a list of node types the matcher
130 can match on, as extracted from the static type asserts in the matcher
131 definition.
132 """
133 if declaration.strip():
134 # Node matchers are defined by writing:
135 # VariadicDynCastAllOfMatcher<ResultType, ArgumentType> name;
Manuel Klimek41df16e2013-01-09 09:38:21 +0000136 m = re.match(r""".*Variadic(?:DynCast)?AllOfMatcher\s*<
137 \s*([^\s,]+)\s*(?:,
138 \s*([^\s>]+)\s*)?>
Manuel Klimek95a58d22012-08-27 18:49:12 +0000139 \s*([^\s;]+)\s*;\s*$""", declaration, flags=re.X)
140 if m:
141 result, inner, name = m.groups()
Manuel Klimek41df16e2013-01-09 09:38:21 +0000142 if not inner:
143 inner = result
Manuel Klimek95a58d22012-08-27 18:49:12 +0000144 add_matcher(result, name, 'Matcher<%s>...' % inner,
145 comment, is_dyncast=True)
146 return
147
148 # Parse the various matcher definition macros.
Manuel Klimek41df16e2013-01-09 09:38:21 +0000149 m = re.match(""".*AST_TYPE_MATCHER\(
150 \s*([^\s,]+\s*),
151 \s*([^\s,]+\s*)
152 \)\s*;\s*$""", declaration, flags=re.X)
153 if m:
154 inner, name = m.groups()
155 add_matcher('Type', name, 'Matcher<%s>...' % inner,
156 comment, is_dyncast=True)
Manuel Klimek2c4b2e42013-07-25 06:05:50 +0000157 # FIXME: re-enable once we have implemented casting on the TypeLoc
158 # hierarchy.
159 # add_matcher('TypeLoc', '%sLoc' % name, 'Matcher<%sLoc>...' % inner,
160 # comment, is_dyncast=True)
Manuel Klimek41df16e2013-01-09 09:38:21 +0000161 return
162
163 m = re.match(""".*AST_TYPE(LOC)?_TRAVERSE_MATCHER\(
164 \s*([^\s,]+\s*),
Samuel Benzaquen3f84bb32013-07-15 19:25:06 +0000165 \s*(?:[^\s,]+\s*),
166 \s*AST_POLYMORPHIC_SUPPORTED_TYPES_([^(]*)\(([^)]*)\)
Manuel Klimek41df16e2013-01-09 09:38:21 +0000167 \)\s*;\s*$""", declaration, flags=re.X)
168 if m:
Samuel Benzaquen3f84bb32013-07-15 19:25:06 +0000169 loc, name, n_results, results = m.groups()[0:4]
170 result_types = [r.strip() for r in results.split(',')]
171
172 comment_result_types = extract_result_types(comment)
173 if (comment_result_types and
174 sorted(result_types) != sorted(comment_result_types)):
175 raise Exception('Inconsistent documentation for: %s' % name)
Manuel Klimek41df16e2013-01-09 09:38:21 +0000176 for result_type in result_types:
177 add_matcher(result_type, name, 'Matcher<Type>', comment)
178 if loc:
179 add_matcher('%sLoc' % result_type, '%sLoc' % name, 'Matcher<TypeLoc>',
180 comment)
181 return
182
Samuel Benzaquenef7eb022013-06-21 15:51:31 +0000183 m = re.match(r"""^\s*AST_POLYMORPHIC_MATCHER(_P)?(.?)(?:_OVERLOAD)?\(
184 \s*([^\s,]+)\s*,
185 \s*AST_POLYMORPHIC_SUPPORTED_TYPES_([^(]*)\(([^)]*)\)
186 (?:,\s*([^\s,]+)\s*
187 ,\s*([^\s,]+)\s*)?
188 (?:,\s*([^\s,]+)\s*
189 ,\s*([^\s,]+)\s*)?
190 (?:,\s*\d+\s*)?
191 \)\s*{\s*$""", declaration, flags=re.X)
192
193 if m:
194 p, n, name, n_results, results = m.groups()[0:5]
195 args = m.groups()[5:]
196 result_types = [r.strip() for r in results.split(',')]
197 if allowed_types and allowed_types != result_types:
198 raise Exception('Inconsistent documentation for: %s' % name)
199 if n not in ['', '2']:
200 raise Exception('Cannot parse "%s"' % declaration)
201 args = ', '.join('%s %s' % (args[i], args[i+1])
202 for i in range(0, len(args), 2) if args[i])
203 for result_type in result_types:
204 add_matcher(result_type, name, args, comment)
205 return
206
207 m = re.match(r"""^\s*AST_MATCHER(_P)?(.?)(?:_OVERLOAD)?\(
Manuel Klimek95a58d22012-08-27 18:49:12 +0000208 (?:\s*([^\s,]+)\s*,)?
209 \s*([^\s,]+)\s*
210 (?:,\s*([^\s,]+)\s*
211 ,\s*([^\s,]+)\s*)?
212 (?:,\s*([^\s,]+)\s*
213 ,\s*([^\s,]+)\s*)?
Manuel Klimek415514d2013-02-06 20:36:22 +0000214 (?:,\s*\d+\s*)?
Manuel Klimek95a58d22012-08-27 18:49:12 +0000215 \)\s*{\s*$""", declaration, flags=re.X)
216 if m:
Samuel Benzaquenef7eb022013-06-21 15:51:31 +0000217 p, n, result, name = m.groups()[0:4]
218 args = m.groups()[4:]
Manuel Klimek95a58d22012-08-27 18:49:12 +0000219 if not result:
220 if not allowed_types:
221 raise Exception('Did not find allowed result types for: %s' % name)
222 result_types = allowed_types
223 else:
224 result_types = [result]
225 if n not in ['', '2']:
226 raise Exception('Cannot parse "%s"' % declaration)
227 args = ', '.join('%s %s' % (args[i], args[i+1])
228 for i in range(0, len(args), 2) if args[i])
229 for result_type in result_types:
230 add_matcher(result_type, name, args, comment)
231 return
232
Samuel Benzaquenee0da952013-08-16 16:19:42 +0000233 # Parse ArgumentAdapting matchers.
234 m = re.match(
235 r"""^.*ArgumentAdaptingMatcherFunc<.*>\s*([a-zA-Z]*)\s*=\s*{};$""",
236 declaration, flags=re.X)
237 if m:
238 name = m.groups()[0]
239 add_matcher('*', name, 'Matcher<*>', comment)
240 return
241
242
Manuel Klimek95a58d22012-08-27 18:49:12 +0000243 # Parse free standing matcher functions, like:
244 # Matcher<ResultType> Name(Matcher<ArgumentType> InnerMatcher) {
245 m = re.match(r"""^\s*(.*)\s+
246 ([^\s\(]+)\s*\(
247 (.*)
248 \)\s*{""", declaration, re.X)
249 if m:
250 result, name, args = m.groups()
251 args = ', '.join(p.strip() for p in args.split(','))
Manuel Klimek41df16e2013-01-09 09:38:21 +0000252 m = re.match(r'.*\s+internal::(Bindable)?Matcher<([^>]+)>$', result)
Manuel Klimek95a58d22012-08-27 18:49:12 +0000253 if m:
Manuel Klimek41df16e2013-01-09 09:38:21 +0000254 result_types = [m.group(2)]
Manuel Klimek95a58d22012-08-27 18:49:12 +0000255 else:
256 result_types = extract_result_types(comment)
257 if not result_types:
258 if not comment:
259 # Only overloads don't have their own doxygen comments; ignore those.
260 print 'Ignoring "%s"' % name
261 else:
262 print 'Cannot determine result type for "%s"' % name
263 else:
264 for result_type in result_types:
265 add_matcher(result_type, name, args, comment)
266 else:
267 print '*** Unparsable: "' + declaration + '" ***'
268
269def sort_table(matcher_type, matcher_map):
270 """Returns the sorted html table for the given row map."""
271 table = ''
272 for key in sorted(matcher_map.keys()):
273 table += matcher_map[key] + '\n'
274 return ('<!-- START_%(type)s_MATCHERS -->\n' +
275 '%(table)s' +
276 '<!--END_%(type)s_MATCHERS -->') % {
277 'type': matcher_type,
278 'table': table,
279 }
280
281# Parse the ast matchers.
282# We alternate between two modes:
283# body = True: We parse the definition of a matcher. We need
284# to parse the full definition before adding a matcher, as the
285# definition might contain static asserts that specify the result
286# type.
287# body = False: We parse the comments and declaration of the matcher.
288comment = ''
289declaration = ''
290allowed_types = []
291body = False
292for line in open(MATCHERS_FILE).read().splitlines():
293 if body:
294 if line.strip() and line[0] == '}':
295 if declaration:
296 act_on_decl(declaration, comment, allowed_types)
297 comment = ''
298 declaration = ''
299 allowed_types = []
300 body = False
301 else:
302 m = re.search(r'is_base_of<([^,]+), NodeType>', line)
303 if m and m.group(1):
304 allowed_types += [m.group(1)]
305 continue
306 if line.strip() and line.lstrip()[0] == '/':
307 comment += re.sub(r'/+\s?', '', line) + '\n'
308 else:
309 declaration += ' ' + line
310 if ((not line.strip()) or
311 line.rstrip()[-1] == ';' or
312 line.rstrip()[-1] == '{'):
313 if line.strip() and line.rstrip()[-1] == '{':
314 body = True
315 else:
316 act_on_decl(declaration, comment, allowed_types)
317 comment = ''
318 declaration = ''
319 allowed_types = []
320
321node_matcher_table = sort_table('DECL', node_matchers)
322narrowing_matcher_table = sort_table('NARROWING', narrowing_matchers)
323traversal_matcher_table = sort_table('TRAVERSAL', traversal_matchers)
324
325reference = open('../LibASTMatchersReference.html').read()
326reference = re.sub(r'<!-- START_DECL_MATCHERS.*END_DECL_MATCHERS -->',
327 '%s', reference, flags=re.S) % node_matcher_table
328reference = re.sub(r'<!-- START_NARROWING_MATCHERS.*END_NARROWING_MATCHERS -->',
329 '%s', reference, flags=re.S) % narrowing_matcher_table
330reference = re.sub(r'<!-- START_TRAVERSAL_MATCHERS.*END_TRAVERSAL_MATCHERS -->',
331 '%s', reference, flags=re.S) % traversal_matcher_table
332
333with open('../LibASTMatchersReference.html', 'w') as output:
334 output.write(reference)
335