|  | //===--- CommentLexer.cpp -------------------------------------------------===// | 
|  | // | 
|  | // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. | 
|  | // See https://llvm.org/LICENSE.txt for license information. | 
|  | // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception | 
|  | // | 
|  | //===----------------------------------------------------------------------===// | 
|  |  | 
|  | #include "clang/AST/CommentLexer.h" | 
|  | #include "clang/AST/CommentCommandTraits.h" | 
|  | #include "clang/AST/CommentDiagnostic.h" | 
|  | #include "clang/Basic/CharInfo.h" | 
|  | #include "llvm/ADT/StringExtras.h" | 
|  | #include "llvm/ADT/StringSwitch.h" | 
|  | #include "llvm/Support/ConvertUTF.h" | 
|  | #include "llvm/Support/ErrorHandling.h" | 
|  |  | 
|  | namespace clang { | 
|  | namespace comments { | 
|  |  | 
|  | void Token::dump(const Lexer &L, const SourceManager &SM) const { | 
|  | llvm::errs() << "comments::Token Kind=" << Kind << " "; | 
|  | Loc.print(llvm::errs(), SM); | 
|  | llvm::errs() << " " << Length << " \"" << L.getSpelling(*this, SM) << "\"\n"; | 
|  | } | 
|  |  | 
|  | static inline bool isHTMLNamedCharacterReferenceCharacter(char C) { | 
|  | return isLetter(C); | 
|  | } | 
|  |  | 
|  | static inline bool isHTMLDecimalCharacterReferenceCharacter(char C) { | 
|  | return isDigit(C); | 
|  | } | 
|  |  | 
|  | static inline bool isHTMLHexCharacterReferenceCharacter(char C) { | 
|  | return isHexDigit(C); | 
|  | } | 
|  |  | 
|  | static inline StringRef convertCodePointToUTF8( | 
|  | llvm::BumpPtrAllocator &Allocator, | 
|  | unsigned CodePoint) { | 
|  | char *Resolved = Allocator.Allocate<char>(UNI_MAX_UTF8_BYTES_PER_CODE_POINT); | 
|  | char *ResolvedPtr = Resolved; | 
|  | if (llvm::ConvertCodePointToUTF8(CodePoint, ResolvedPtr)) | 
|  | return StringRef(Resolved, ResolvedPtr - Resolved); | 
|  | else | 
|  | return StringRef(); | 
|  | } | 
|  |  | 
|  | namespace { | 
|  |  | 
|  | #include "clang/AST/CommentHTMLTags.inc" | 
|  | #include "clang/AST/CommentHTMLNamedCharacterReferences.inc" | 
|  |  | 
|  | } // end anonymous namespace | 
|  |  | 
|  | StringRef Lexer::resolveHTMLNamedCharacterReference(StringRef Name) const { | 
|  | // Fast path, first check a few most widely used named character references. | 
|  | return llvm::StringSwitch<StringRef>(Name) | 
|  | .Case("amp", "&") | 
|  | .Case("lt", "<") | 
|  | .Case("gt", ">") | 
|  | .Case("quot", "\"") | 
|  | .Case("apos", "\'") | 
|  | // Slow path. | 
|  | .Default(translateHTMLNamedCharacterReferenceToUTF8(Name)); | 
|  | } | 
|  |  | 
|  | StringRef Lexer::resolveHTMLDecimalCharacterReference(StringRef Name) const { | 
|  | unsigned CodePoint = 0; | 
|  | for (unsigned i = 0, e = Name.size(); i != e; ++i) { | 
|  | assert(isHTMLDecimalCharacterReferenceCharacter(Name[i])); | 
|  | CodePoint *= 10; | 
|  | CodePoint += Name[i] - '0'; | 
|  | } | 
|  | return convertCodePointToUTF8(Allocator, CodePoint); | 
|  | } | 
|  |  | 
|  | StringRef Lexer::resolveHTMLHexCharacterReference(StringRef Name) const { | 
|  | unsigned CodePoint = 0; | 
|  | for (unsigned i = 0, e = Name.size(); i != e; ++i) { | 
|  | CodePoint *= 16; | 
|  | const char C = Name[i]; | 
|  | assert(isHTMLHexCharacterReferenceCharacter(C)); | 
|  | CodePoint += llvm::hexDigitValue(C); | 
|  | } | 
|  | return convertCodePointToUTF8(Allocator, CodePoint); | 
|  | } | 
|  |  | 
|  | void Lexer::skipLineStartingDecorations() { | 
|  | // This function should be called only for C comments | 
|  | assert(CommentState == LCS_InsideCComment); | 
|  |  | 
|  | if (BufferPtr == CommentEnd) | 
|  | return; | 
|  |  | 
|  | switch (*BufferPtr) { | 
|  | case ' ': | 
|  | case '\t': | 
|  | case '\f': | 
|  | case '\v': { | 
|  | const char *NewBufferPtr = BufferPtr; | 
|  | NewBufferPtr++; | 
|  | if (NewBufferPtr == CommentEnd) | 
|  | return; | 
|  |  | 
|  | char C = *NewBufferPtr; | 
|  | while (isHorizontalWhitespace(C)) { | 
|  | NewBufferPtr++; | 
|  | if (NewBufferPtr == CommentEnd) | 
|  | return; | 
|  | C = *NewBufferPtr; | 
|  | } | 
|  | if (C == '*') | 
|  | BufferPtr = NewBufferPtr + 1; | 
|  | break; | 
|  | } | 
|  | case '*': | 
|  | BufferPtr++; | 
|  | break; | 
|  | } | 
|  | } | 
|  |  | 
|  | namespace { | 
|  | /// Returns pointer to the first newline character in the string. | 
|  | const char *findNewline(const char *BufferPtr, const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (isVerticalWhitespace(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | const char *skipNewline(const char *BufferPtr, const char *BufferEnd) { | 
|  | if (BufferPtr == BufferEnd) | 
|  | return BufferPtr; | 
|  |  | 
|  | if (*BufferPtr == '\n') | 
|  | BufferPtr++; | 
|  | else { | 
|  | assert(*BufferPtr == '\r'); | 
|  | BufferPtr++; | 
|  | if (BufferPtr != BufferEnd && *BufferPtr == '\n') | 
|  | BufferPtr++; | 
|  | } | 
|  | return BufferPtr; | 
|  | } | 
|  |  | 
|  | const char *skipNamedCharacterReference(const char *BufferPtr, | 
|  | const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (!isHTMLNamedCharacterReferenceCharacter(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | const char *skipDecimalCharacterReference(const char *BufferPtr, | 
|  | const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (!isHTMLDecimalCharacterReferenceCharacter(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | const char *skipHexCharacterReference(const char *BufferPtr, | 
|  | const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (!isHTMLHexCharacterReferenceCharacter(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | bool isHTMLIdentifierStartingCharacter(char C) { | 
|  | return isLetter(C); | 
|  | } | 
|  |  | 
|  | bool isHTMLIdentifierCharacter(char C) { | 
|  | return isAlphanumeric(C); | 
|  | } | 
|  |  | 
|  | const char *skipHTMLIdentifier(const char *BufferPtr, const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (!isHTMLIdentifierCharacter(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | /// Skip HTML string quoted in single or double quotes.  Escaping quotes inside | 
|  | /// string allowed. | 
|  | /// | 
|  | /// Returns pointer to closing quote. | 
|  | const char *skipHTMLQuotedString(const char *BufferPtr, const char *BufferEnd) | 
|  | { | 
|  | const char Quote = *BufferPtr; | 
|  | assert(Quote == '\"' || Quote == '\''); | 
|  |  | 
|  | BufferPtr++; | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | const char C = *BufferPtr; | 
|  | if (C == Quote && BufferPtr[-1] != '\\') | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | const char *skipWhitespace(const char *BufferPtr, const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (!isWhitespace(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | bool isWhitespace(const char *BufferPtr, const char *BufferEnd) { | 
|  | return skipWhitespace(BufferPtr, BufferEnd) == BufferEnd; | 
|  | } | 
|  |  | 
|  | bool isCommandNameStartCharacter(char C) { | 
|  | return isLetter(C); | 
|  | } | 
|  |  | 
|  | bool isCommandNameCharacter(char C) { | 
|  | return isAlphanumeric(C); | 
|  | } | 
|  |  | 
|  | const char *skipCommandName(const char *BufferPtr, const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (!isCommandNameCharacter(*BufferPtr)) | 
|  | return BufferPtr; | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | /// Return the one past end pointer for BCPL comments. | 
|  | /// Handles newlines escaped with backslash or trigraph for backslahs. | 
|  | const char *findBCPLCommentEnd(const char *BufferPtr, const char *BufferEnd) { | 
|  | const char *CurPtr = BufferPtr; | 
|  | while (CurPtr != BufferEnd) { | 
|  | while (!isVerticalWhitespace(*CurPtr)) { | 
|  | CurPtr++; | 
|  | if (CurPtr == BufferEnd) | 
|  | return BufferEnd; | 
|  | } | 
|  | // We found a newline, check if it is escaped. | 
|  | const char *EscapePtr = CurPtr - 1; | 
|  | while(isHorizontalWhitespace(*EscapePtr)) | 
|  | EscapePtr--; | 
|  |  | 
|  | if (*EscapePtr == '\\' || | 
|  | (EscapePtr - 2 >= BufferPtr && EscapePtr[0] == '/' && | 
|  | EscapePtr[-1] == '?' && EscapePtr[-2] == '?')) { | 
|  | // We found an escaped newline. | 
|  | CurPtr = skipNewline(CurPtr, BufferEnd); | 
|  | } else | 
|  | return CurPtr; // Not an escaped newline. | 
|  | } | 
|  | return BufferEnd; | 
|  | } | 
|  |  | 
|  | /// Return the one past end pointer for C comments. | 
|  | /// Very dumb, does not handle escaped newlines or trigraphs. | 
|  | const char *findCCommentEnd(const char *BufferPtr, const char *BufferEnd) { | 
|  | for ( ; BufferPtr != BufferEnd; ++BufferPtr) { | 
|  | if (*BufferPtr == '*') { | 
|  | assert(BufferPtr + 1 != BufferEnd); | 
|  | if (*(BufferPtr + 1) == '/') | 
|  | return BufferPtr; | 
|  | } | 
|  | } | 
|  | llvm_unreachable("buffer end hit before '*/' was seen"); | 
|  | } | 
|  |  | 
|  | } // end anonymous namespace | 
|  |  | 
|  | void Lexer::formTokenWithChars(Token &Result, const char *TokEnd, | 
|  | tok::TokenKind Kind) { | 
|  | const unsigned TokLen = TokEnd - BufferPtr; | 
|  | Result.setLocation(getSourceLocation(BufferPtr)); | 
|  | Result.setKind(Kind); | 
|  | Result.setLength(TokLen); | 
|  | #ifndef NDEBUG | 
|  | Result.TextPtr = "<UNSET>"; | 
|  | Result.IntVal = 7; | 
|  | #endif | 
|  | BufferPtr = TokEnd; | 
|  | } | 
|  |  | 
|  | void Lexer::lexCommentText(Token &T) { | 
|  | assert(CommentState == LCS_InsideBCPLComment || | 
|  | CommentState == LCS_InsideCComment); | 
|  |  | 
|  | // Handles lexing non-command text, i.e. text and newline. | 
|  | auto HandleNonCommandToken = [&]() -> void { | 
|  | assert(State == LS_Normal); | 
|  |  | 
|  | const char *TokenPtr = BufferPtr; | 
|  | assert(TokenPtr < CommentEnd); | 
|  | switch (*TokenPtr) { | 
|  | case '\n': | 
|  | case '\r': | 
|  | TokenPtr = skipNewline(TokenPtr, CommentEnd); | 
|  | formTokenWithChars(T, TokenPtr, tok::newline); | 
|  |  | 
|  | if (CommentState == LCS_InsideCComment) | 
|  | skipLineStartingDecorations(); | 
|  | return; | 
|  |  | 
|  | default: { | 
|  | StringRef TokStartSymbols = ParseCommands ? "\n\r\\@&<" : "\n\r"; | 
|  | size_t End = StringRef(TokenPtr, CommentEnd - TokenPtr) | 
|  | .find_first_of(TokStartSymbols); | 
|  | if (End != StringRef::npos) | 
|  | TokenPtr += End; | 
|  | else | 
|  | TokenPtr = CommentEnd; | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | } | 
|  | }; | 
|  |  | 
|  | if (!ParseCommands) | 
|  | return HandleNonCommandToken(); | 
|  |  | 
|  | switch (State) { | 
|  | case LS_Normal: | 
|  | break; | 
|  | case LS_VerbatimBlockFirstLine: | 
|  | lexVerbatimBlockFirstLine(T); | 
|  | return; | 
|  | case LS_VerbatimBlockBody: | 
|  | lexVerbatimBlockBody(T); | 
|  | return; | 
|  | case LS_VerbatimLineText: | 
|  | lexVerbatimLineText(T); | 
|  | return; | 
|  | case LS_HTMLStartTag: | 
|  | lexHTMLStartTag(T); | 
|  | return; | 
|  | case LS_HTMLEndTag: | 
|  | lexHTMLEndTag(T); | 
|  | return; | 
|  | } | 
|  |  | 
|  | assert(State == LS_Normal); | 
|  | const char *TokenPtr = BufferPtr; | 
|  | assert(TokenPtr < CommentEnd); | 
|  | switch(*TokenPtr) { | 
|  | case '\\': | 
|  | case '@': { | 
|  | // Commands that start with a backslash and commands that start with | 
|  | // 'at' have equivalent semantics.  But we keep information about the | 
|  | // exact syntax in AST for comments. | 
|  | tok::TokenKind CommandKind = | 
|  | (*TokenPtr == '@') ? tok::at_command : tok::backslash_command; | 
|  | TokenPtr++; | 
|  | if (TokenPtr == CommentEnd) { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | char C = *TokenPtr; | 
|  | switch (C) { | 
|  | default: | 
|  | break; | 
|  |  | 
|  | case '\\': case '@': case '&': case '$': | 
|  | case '#':  case '<': case '>': case '%': | 
|  | case '\"': case '.': case ':': | 
|  | // This is one of \\ \@ \& \$ etc escape sequences. | 
|  | TokenPtr++; | 
|  | if (C == ':' && TokenPtr != CommentEnd && *TokenPtr == ':') { | 
|  | // This is the \:: escape sequence. | 
|  | TokenPtr++; | 
|  | } | 
|  | StringRef UnescapedText(BufferPtr + 1, TokenPtr - (BufferPtr + 1)); | 
|  | formTokenWithChars(T, TokenPtr, tok::text); | 
|  | T.setText(UnescapedText); | 
|  | return; | 
|  | } | 
|  |  | 
|  | // Don't make zero-length commands. | 
|  | if (!isCommandNameStartCharacter(*TokenPtr)) { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  |  | 
|  | TokenPtr = skipCommandName(TokenPtr, CommentEnd); | 
|  | unsigned Length = TokenPtr - (BufferPtr + 1); | 
|  |  | 
|  | // Hardcoded support for lexing LaTeX formula commands | 
|  | // \f$ \f[ \f] \f{ \f} as a single command. | 
|  | if (Length == 1 && TokenPtr[-1] == 'f' && TokenPtr != CommentEnd) { | 
|  | C = *TokenPtr; | 
|  | if (C == '$' || C == '[' || C == ']' || C == '{' || C == '}') { | 
|  | TokenPtr++; | 
|  | Length++; | 
|  | } | 
|  | } | 
|  |  | 
|  | StringRef CommandName(BufferPtr + 1, Length); | 
|  |  | 
|  | const CommandInfo *Info = Traits.getCommandInfoOrNULL(CommandName); | 
|  | if (!Info) { | 
|  | if ((Info = Traits.getTypoCorrectCommandInfo(CommandName))) { | 
|  | StringRef CorrectedName = Info->Name; | 
|  | SourceLocation Loc = getSourceLocation(BufferPtr); | 
|  | SourceLocation EndLoc = getSourceLocation(TokenPtr); | 
|  | SourceRange FullRange = SourceRange(Loc, EndLoc); | 
|  | SourceRange CommandRange(Loc.getLocWithOffset(1), EndLoc); | 
|  | Diag(Loc, diag::warn_correct_comment_command_name) | 
|  | << FullRange << CommandName << CorrectedName | 
|  | << FixItHint::CreateReplacement(CommandRange, CorrectedName); | 
|  | } else { | 
|  | formTokenWithChars(T, TokenPtr, tok::unknown_command); | 
|  | T.setUnknownCommandName(CommandName); | 
|  | Diag(T.getLocation(), diag::warn_unknown_comment_command_name) | 
|  | << SourceRange(T.getLocation(), T.getEndLocation()); | 
|  | return; | 
|  | } | 
|  | } | 
|  | if (Info->IsVerbatimBlockCommand) { | 
|  | setupAndLexVerbatimBlock(T, TokenPtr, *BufferPtr, Info); | 
|  | return; | 
|  | } | 
|  | if (Info->IsVerbatimLineCommand) { | 
|  | setupAndLexVerbatimLine(T, TokenPtr, Info); | 
|  | return; | 
|  | } | 
|  | formTokenWithChars(T, TokenPtr, CommandKind); | 
|  | T.setCommandID(Info->getID()); | 
|  | return; | 
|  | } | 
|  |  | 
|  | case '&': | 
|  | lexHTMLCharacterReference(T); | 
|  | return; | 
|  |  | 
|  | case '<': { | 
|  | TokenPtr++; | 
|  | if (TokenPtr == CommentEnd) { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | const char C = *TokenPtr; | 
|  | if (isHTMLIdentifierStartingCharacter(C)) | 
|  | setupAndLexHTMLStartTag(T); | 
|  | else if (C == '/') | 
|  | setupAndLexHTMLEndTag(T); | 
|  | else | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  |  | 
|  | default: | 
|  | return HandleNonCommandToken(); | 
|  | } | 
|  | } | 
|  |  | 
|  | void Lexer::setupAndLexVerbatimBlock(Token &T, | 
|  | const char *TextBegin, | 
|  | char Marker, const CommandInfo *Info) { | 
|  | assert(Info->IsVerbatimBlockCommand); | 
|  |  | 
|  | VerbatimBlockEndCommandName.clear(); | 
|  | VerbatimBlockEndCommandName.append(Marker == '\\' ? "\\" : "@"); | 
|  | VerbatimBlockEndCommandName.append(Info->EndCommandName); | 
|  |  | 
|  | formTokenWithChars(T, TextBegin, tok::verbatim_block_begin); | 
|  | T.setVerbatimBlockID(Info->getID()); | 
|  |  | 
|  | // If there is a newline following the verbatim opening command, skip the | 
|  | // newline so that we don't create an tok::verbatim_block_line with empty | 
|  | // text content. | 
|  | if (BufferPtr != CommentEnd && | 
|  | isVerticalWhitespace(*BufferPtr)) { | 
|  | BufferPtr = skipNewline(BufferPtr, CommentEnd); | 
|  | State = LS_VerbatimBlockBody; | 
|  | return; | 
|  | } | 
|  |  | 
|  | State = LS_VerbatimBlockFirstLine; | 
|  | } | 
|  |  | 
|  | void Lexer::lexVerbatimBlockFirstLine(Token &T) { | 
|  | again: | 
|  | assert(BufferPtr < CommentEnd); | 
|  |  | 
|  | // FIXME: It would be better to scan the text once, finding either the block | 
|  | // end command or newline. | 
|  | // | 
|  | // Extract current line. | 
|  | const char *Newline = findNewline(BufferPtr, CommentEnd); | 
|  | StringRef Line(BufferPtr, Newline - BufferPtr); | 
|  |  | 
|  | // Look for end command in current line. | 
|  | size_t Pos = Line.find(VerbatimBlockEndCommandName); | 
|  | const char *TextEnd; | 
|  | const char *NextLine; | 
|  | if (Pos == StringRef::npos) { | 
|  | // Current line is completely verbatim. | 
|  | TextEnd = Newline; | 
|  | NextLine = skipNewline(Newline, CommentEnd); | 
|  | } else if (Pos == 0) { | 
|  | // Current line contains just an end command. | 
|  | const char *End = BufferPtr + VerbatimBlockEndCommandName.size(); | 
|  | StringRef Name(BufferPtr + 1, End - (BufferPtr + 1)); | 
|  | formTokenWithChars(T, End, tok::verbatim_block_end); | 
|  | T.setVerbatimBlockID(Traits.getCommandInfo(Name)->getID()); | 
|  | State = LS_Normal; | 
|  | return; | 
|  | } else { | 
|  | // There is some text, followed by end command.  Extract text first. | 
|  | TextEnd = BufferPtr + Pos; | 
|  | NextLine = TextEnd; | 
|  | // If there is only whitespace before end command, skip whitespace. | 
|  | if (isWhitespace(BufferPtr, TextEnd)) { | 
|  | BufferPtr = TextEnd; | 
|  | goto again; | 
|  | } | 
|  | } | 
|  |  | 
|  | StringRef Text(BufferPtr, TextEnd - BufferPtr); | 
|  | formTokenWithChars(T, NextLine, tok::verbatim_block_line); | 
|  | T.setVerbatimBlockText(Text); | 
|  |  | 
|  | State = LS_VerbatimBlockBody; | 
|  | } | 
|  |  | 
|  | void Lexer::lexVerbatimBlockBody(Token &T) { | 
|  | assert(State == LS_VerbatimBlockBody); | 
|  |  | 
|  | if (CommentState == LCS_InsideCComment) | 
|  | skipLineStartingDecorations(); | 
|  |  | 
|  | if (BufferPtr == CommentEnd) { | 
|  | formTokenWithChars(T, BufferPtr, tok::verbatim_block_line); | 
|  | T.setVerbatimBlockText(""); | 
|  | return; | 
|  | } | 
|  |  | 
|  | lexVerbatimBlockFirstLine(T); | 
|  | } | 
|  |  | 
|  | void Lexer::setupAndLexVerbatimLine(Token &T, const char *TextBegin, | 
|  | const CommandInfo *Info) { | 
|  | assert(Info->IsVerbatimLineCommand); | 
|  | formTokenWithChars(T, TextBegin, tok::verbatim_line_name); | 
|  | T.setVerbatimLineID(Info->getID()); | 
|  |  | 
|  | State = LS_VerbatimLineText; | 
|  | } | 
|  |  | 
|  | void Lexer::lexVerbatimLineText(Token &T) { | 
|  | assert(State == LS_VerbatimLineText); | 
|  |  | 
|  | // Extract current line. | 
|  | const char *Newline = findNewline(BufferPtr, CommentEnd); | 
|  | StringRef Text(BufferPtr, Newline - BufferPtr); | 
|  | formTokenWithChars(T, Newline, tok::verbatim_line_text); | 
|  | T.setVerbatimLineText(Text); | 
|  |  | 
|  | State = LS_Normal; | 
|  | } | 
|  |  | 
|  | void Lexer::lexHTMLCharacterReference(Token &T) { | 
|  | const char *TokenPtr = BufferPtr; | 
|  | assert(*TokenPtr == '&'); | 
|  | TokenPtr++; | 
|  | if (TokenPtr == CommentEnd) { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | const char *NamePtr; | 
|  | bool isNamed = false; | 
|  | bool isDecimal = false; | 
|  | char C = *TokenPtr; | 
|  | if (isHTMLNamedCharacterReferenceCharacter(C)) { | 
|  | NamePtr = TokenPtr; | 
|  | TokenPtr = skipNamedCharacterReference(TokenPtr, CommentEnd); | 
|  | isNamed = true; | 
|  | } else if (C == '#') { | 
|  | TokenPtr++; | 
|  | if (TokenPtr == CommentEnd) { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | C = *TokenPtr; | 
|  | if (isHTMLDecimalCharacterReferenceCharacter(C)) { | 
|  | NamePtr = TokenPtr; | 
|  | TokenPtr = skipDecimalCharacterReference(TokenPtr, CommentEnd); | 
|  | isDecimal = true; | 
|  | } else if (C == 'x' || C == 'X') { | 
|  | TokenPtr++; | 
|  | NamePtr = TokenPtr; | 
|  | TokenPtr = skipHexCharacterReference(TokenPtr, CommentEnd); | 
|  | } else { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | } else { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | if (NamePtr == TokenPtr || TokenPtr == CommentEnd || | 
|  | *TokenPtr != ';') { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | StringRef Name(NamePtr, TokenPtr - NamePtr); | 
|  | TokenPtr++; // Skip semicolon. | 
|  | StringRef Resolved; | 
|  | if (isNamed) | 
|  | Resolved = resolveHTMLNamedCharacterReference(Name); | 
|  | else if (isDecimal) | 
|  | Resolved = resolveHTMLDecimalCharacterReference(Name); | 
|  | else | 
|  | Resolved = resolveHTMLHexCharacterReference(Name); | 
|  |  | 
|  | if (Resolved.empty()) { | 
|  | formTextToken(T, TokenPtr); | 
|  | return; | 
|  | } | 
|  | formTokenWithChars(T, TokenPtr, tok::text); | 
|  | T.setText(Resolved); | 
|  | } | 
|  |  | 
|  | void Lexer::setupAndLexHTMLStartTag(Token &T) { | 
|  | assert(BufferPtr[0] == '<' && | 
|  | isHTMLIdentifierStartingCharacter(BufferPtr[1])); | 
|  | const char *TagNameEnd = skipHTMLIdentifier(BufferPtr + 2, CommentEnd); | 
|  | StringRef Name(BufferPtr + 1, TagNameEnd - (BufferPtr + 1)); | 
|  | if (!isHTMLTagName(Name)) { | 
|  | formTextToken(T, TagNameEnd); | 
|  | return; | 
|  | } | 
|  |  | 
|  | formTokenWithChars(T, TagNameEnd, tok::html_start_tag); | 
|  | T.setHTMLTagStartName(Name); | 
|  |  | 
|  | BufferPtr = skipWhitespace(BufferPtr, CommentEnd); | 
|  |  | 
|  | const char C = *BufferPtr; | 
|  | if (BufferPtr != CommentEnd && | 
|  | (C == '>' || C == '/' || isHTMLIdentifierStartingCharacter(C))) | 
|  | State = LS_HTMLStartTag; | 
|  | } | 
|  |  | 
|  | void Lexer::lexHTMLStartTag(Token &T) { | 
|  | assert(State == LS_HTMLStartTag); | 
|  |  | 
|  | const char *TokenPtr = BufferPtr; | 
|  | char C = *TokenPtr; | 
|  | if (isHTMLIdentifierCharacter(C)) { | 
|  | TokenPtr = skipHTMLIdentifier(TokenPtr, CommentEnd); | 
|  | StringRef Ident(BufferPtr, TokenPtr - BufferPtr); | 
|  | formTokenWithChars(T, TokenPtr, tok::html_ident); | 
|  | T.setHTMLIdent(Ident); | 
|  | } else { | 
|  | switch (C) { | 
|  | case '=': | 
|  | TokenPtr++; | 
|  | formTokenWithChars(T, TokenPtr, tok::html_equals); | 
|  | break; | 
|  | case '\"': | 
|  | case '\'': { | 
|  | const char *OpenQuote = TokenPtr; | 
|  | TokenPtr = skipHTMLQuotedString(TokenPtr, CommentEnd); | 
|  | const char *ClosingQuote = TokenPtr; | 
|  | if (TokenPtr != CommentEnd) // Skip closing quote. | 
|  | TokenPtr++; | 
|  | formTokenWithChars(T, TokenPtr, tok::html_quoted_string); | 
|  | T.setHTMLQuotedString(StringRef(OpenQuote + 1, | 
|  | ClosingQuote - (OpenQuote + 1))); | 
|  | break; | 
|  | } | 
|  | case '>': | 
|  | TokenPtr++; | 
|  | formTokenWithChars(T, TokenPtr, tok::html_greater); | 
|  | State = LS_Normal; | 
|  | return; | 
|  | case '/': | 
|  | TokenPtr++; | 
|  | if (TokenPtr != CommentEnd && *TokenPtr == '>') { | 
|  | TokenPtr++; | 
|  | formTokenWithChars(T, TokenPtr, tok::html_slash_greater); | 
|  | } else | 
|  | formTextToken(T, TokenPtr); | 
|  |  | 
|  | State = LS_Normal; | 
|  | return; | 
|  | } | 
|  | } | 
|  |  | 
|  | // Now look ahead and return to normal state if we don't see any HTML tokens | 
|  | // ahead. | 
|  | BufferPtr = skipWhitespace(BufferPtr, CommentEnd); | 
|  | if (BufferPtr == CommentEnd) { | 
|  | State = LS_Normal; | 
|  | return; | 
|  | } | 
|  |  | 
|  | C = *BufferPtr; | 
|  | if (!isHTMLIdentifierStartingCharacter(C) && | 
|  | C != '=' && C != '\"' && C != '\'' && C != '>') { | 
|  | State = LS_Normal; | 
|  | return; | 
|  | } | 
|  | } | 
|  |  | 
|  | void Lexer::setupAndLexHTMLEndTag(Token &T) { | 
|  | assert(BufferPtr[0] == '<' && BufferPtr[1] == '/'); | 
|  |  | 
|  | const char *TagNameBegin = skipWhitespace(BufferPtr + 2, CommentEnd); | 
|  | const char *TagNameEnd = skipHTMLIdentifier(TagNameBegin, CommentEnd); | 
|  | StringRef Name(TagNameBegin, TagNameEnd - TagNameBegin); | 
|  | if (!isHTMLTagName(Name)) { | 
|  | formTextToken(T, TagNameEnd); | 
|  | return; | 
|  | } | 
|  |  | 
|  | const char *End = skipWhitespace(TagNameEnd, CommentEnd); | 
|  |  | 
|  | formTokenWithChars(T, End, tok::html_end_tag); | 
|  | T.setHTMLTagEndName(Name); | 
|  |  | 
|  | if (BufferPtr != CommentEnd && *BufferPtr == '>') | 
|  | State = LS_HTMLEndTag; | 
|  | } | 
|  |  | 
|  | void Lexer::lexHTMLEndTag(Token &T) { | 
|  | assert(BufferPtr != CommentEnd && *BufferPtr == '>'); | 
|  |  | 
|  | formTokenWithChars(T, BufferPtr + 1, tok::html_greater); | 
|  | State = LS_Normal; | 
|  | } | 
|  |  | 
|  | Lexer::Lexer(llvm::BumpPtrAllocator &Allocator, DiagnosticsEngine &Diags, | 
|  | const CommandTraits &Traits, SourceLocation FileLoc, | 
|  | const char *BufferStart, const char *BufferEnd, | 
|  | bool ParseCommands) | 
|  | : Allocator(Allocator), Diags(Diags), Traits(Traits), | 
|  | BufferStart(BufferStart), BufferEnd(BufferEnd), FileLoc(FileLoc), | 
|  | BufferPtr(BufferStart), CommentState(LCS_BeforeComment), State(LS_Normal), | 
|  | ParseCommands(ParseCommands) {} | 
|  |  | 
|  | void Lexer::lex(Token &T) { | 
|  | again: | 
|  | switch (CommentState) { | 
|  | case LCS_BeforeComment: | 
|  | if (BufferPtr == BufferEnd) { | 
|  | formTokenWithChars(T, BufferPtr, tok::eof); | 
|  | return; | 
|  | } | 
|  |  | 
|  | assert(*BufferPtr == '/'); | 
|  | BufferPtr++; // Skip first slash. | 
|  | switch(*BufferPtr) { | 
|  | case '/': { // BCPL comment. | 
|  | BufferPtr++; // Skip second slash. | 
|  |  | 
|  | if (BufferPtr != BufferEnd) { | 
|  | // Skip Doxygen magic marker, if it is present. | 
|  | // It might be missing because of a typo //< or /*<, or because we | 
|  | // merged this non-Doxygen comment into a bunch of Doxygen comments | 
|  | // around it: /** ... */ /* ... */ /** ... */ | 
|  | const char C = *BufferPtr; | 
|  | if (C == '/' || C == '!') | 
|  | BufferPtr++; | 
|  | } | 
|  |  | 
|  | // Skip less-than symbol that marks trailing comments. | 
|  | // Skip it even if the comment is not a Doxygen one, because //< and /*< | 
|  | // are frequent typos. | 
|  | if (BufferPtr != BufferEnd && *BufferPtr == '<') | 
|  | BufferPtr++; | 
|  |  | 
|  | CommentState = LCS_InsideBCPLComment; | 
|  | if (State != LS_VerbatimBlockBody && State != LS_VerbatimBlockFirstLine) | 
|  | State = LS_Normal; | 
|  | CommentEnd = findBCPLCommentEnd(BufferPtr, BufferEnd); | 
|  | goto again; | 
|  | } | 
|  | case '*': { // C comment. | 
|  | BufferPtr++; // Skip star. | 
|  |  | 
|  | // Skip Doxygen magic marker. | 
|  | const char C = *BufferPtr; | 
|  | if ((C == '*' && *(BufferPtr + 1) != '/') || C == '!') | 
|  | BufferPtr++; | 
|  |  | 
|  | // Skip less-than symbol that marks trailing comments. | 
|  | if (BufferPtr != BufferEnd && *BufferPtr == '<') | 
|  | BufferPtr++; | 
|  |  | 
|  | CommentState = LCS_InsideCComment; | 
|  | State = LS_Normal; | 
|  | CommentEnd = findCCommentEnd(BufferPtr, BufferEnd); | 
|  | goto again; | 
|  | } | 
|  | default: | 
|  | llvm_unreachable("second character of comment should be '/' or '*'"); | 
|  | } | 
|  |  | 
|  | case LCS_BetweenComments: { | 
|  | // Consecutive comments are extracted only if there is only whitespace | 
|  | // between them.  So we can search for the start of the next comment. | 
|  | const char *EndWhitespace = BufferPtr; | 
|  | while(EndWhitespace != BufferEnd && *EndWhitespace != '/') | 
|  | EndWhitespace++; | 
|  |  | 
|  | // Turn any whitespace between comments (and there is only whitespace | 
|  | // between them -- guaranteed by comment extraction) into a newline.  We | 
|  | // have two newlines between C comments in total (first one was synthesized | 
|  | // after a comment). | 
|  | formTokenWithChars(T, EndWhitespace, tok::newline); | 
|  |  | 
|  | CommentState = LCS_BeforeComment; | 
|  | break; | 
|  | } | 
|  |  | 
|  | case LCS_InsideBCPLComment: | 
|  | case LCS_InsideCComment: | 
|  | if (BufferPtr != CommentEnd) { | 
|  | lexCommentText(T); | 
|  | break; | 
|  | } else { | 
|  | // Skip C comment closing sequence. | 
|  | if (CommentState == LCS_InsideCComment) { | 
|  | assert(BufferPtr[0] == '*' && BufferPtr[1] == '/'); | 
|  | BufferPtr += 2; | 
|  | assert(BufferPtr <= BufferEnd); | 
|  |  | 
|  | // Synthenize newline just after the C comment, regardless if there is | 
|  | // actually a newline. | 
|  | formTokenWithChars(T, BufferPtr, tok::newline); | 
|  |  | 
|  | CommentState = LCS_BetweenComments; | 
|  | break; | 
|  | } else { | 
|  | // Don't synthesized a newline after BCPL comment. | 
|  | CommentState = LCS_BetweenComments; | 
|  | goto again; | 
|  | } | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | StringRef Lexer::getSpelling(const Token &Tok, | 
|  | const SourceManager &SourceMgr, | 
|  | bool *Invalid) const { | 
|  | SourceLocation Loc = Tok.getLocation(); | 
|  | std::pair<FileID, unsigned> LocInfo = SourceMgr.getDecomposedLoc(Loc); | 
|  |  | 
|  | bool InvalidTemp = false; | 
|  | StringRef File = SourceMgr.getBufferData(LocInfo.first, &InvalidTemp); | 
|  | if (InvalidTemp) { | 
|  | *Invalid = true; | 
|  | return StringRef(); | 
|  | } | 
|  |  | 
|  | const char *Begin = File.data() + LocInfo.second; | 
|  | return StringRef(Begin, Tok.getLength()); | 
|  | } | 
|  |  | 
|  | } // end namespace comments | 
|  | } // end namespace clang |