/external/antlr/antlr-3.4/runtime/Ruby/test/functional/lexer/ |
D | basic.rb | 20 token = lexer.next_token 21 token.name.should == 'ZERO' 23 token = lexer.next_token 24 token.name.should == '<EOF>' 30 token_types = lexer.map { |token| token.name } 38 token = lexer.next_token 62 token = lexer.next_token 63 token.name.should == 'ZERO' 65 token = lexer.next_token 66 token.name.should == 'ONE' [all …]
|
/external/antlr/antlr-3.4/runtime/C/src/ |
D | antlr3commontoken.c | 40 static pANTLR3_STRING getText (pANTLR3_COMMON_TOKEN token); 41 static void setText (pANTLR3_COMMON_TOKEN token, pANTLR3_STRING text); 42 static void setText8 (pANTLR3_COMMON_TOKEN token, pANTLR3_UINT8 text); 43 static ANTLR3_UINT32 getType (pANTLR3_COMMON_TOKEN token); 44 static void setType (pANTLR3_COMMON_TOKEN token, ANTLR3_UINT32 type); 45 static ANTLR3_UINT32 getLine (pANTLR3_COMMON_TOKEN token); 46 static void setLine (pANTLR3_COMMON_TOKEN token, ANTLR3_UINT32 line); 47 static ANTLR3_INT32 getCharPositionInLine (pANTLR3_COMMON_TOKEN token); 48 static void setCharPositionInLine (pANTLR3_COMMON_TOKEN token, ANTLR3_INT32 pos); 49 static ANTLR3_UINT32 getChannel (pANTLR3_COMMON_TOKEN token); [all …]
|
/external/libxml2/doc/ |
D | apibuild.py | 405 def push(self, token): argument 406 self.tokens.insert(0, token); 413 def token(self): member in CLexer 632 def error(self, msg, token=-1): argument 637 if token != -1: 638 print("Got token ", token) 642 def debug(self, msg, token=-1): argument 644 if token != -1: 645 print("Got token ", token) 676 def parseComment(self, token): argument [all …]
|
/external/chromium-trace/trace-viewer/tracing/third_party/closure_linter/closure_linter/ |
D | tokenutil.py | 32 def GetFirstTokenInSameLine(token): argument 41 while not token.IsFirstInLine(): 42 token = token.previous 43 return token 46 def GetFirstTokenInPreviousLine(token): argument 56 first_in_line = GetFirstTokenInSameLine(token) 63 def GetLastTokenInSameLine(token): argument 72 while not token.IsLastInLine(): 73 token = token.next 74 return token [all …]
|
D | indentation.py | 81 def __init__(self, token, is_block=False): argument 88 self.token = token 92 self.is_transient = not is_block and not token.type in ( 94 self.line_number = token.line_number 97 result = '\n %s' % self.token 100 result, self.overridden_by.token.string) 127 def CheckToken(self, token, state): argument 139 token_type = token.type 142 is_first = self._IsFirstNonWhitespaceTokenInLine(token) 158 goog_scope = self._GoogScopeOrNone(start_token.token) [all …]
|
D | ecmalintrules.py | 101 def HandleMissingParameterDoc(self, token, param_name): argument 114 token = last_token 119 while token and token.line_number == line_number: 120 if state.IsTypeToken(token): 121 line.insert(0, 'x' * len(token.string)) 122 elif token.type in (Type.IDENTIFIER, Type.NORMAL): 124 line.insert(0, token.string.replace('.', ' ')) 126 line.insert(0, token.string) 127 token = token.previous 164 def _CheckJsDocType(self, token): argument [all …]
|
D | closurizednamespacesinfo_test.py | 76 token = self._GetRequireTokens('package.Something') 81 self.assertFalse(namespaces_info.IsExtraRequire(token), 87 self.assertTrue(namespaces_info.IsExtraRequire(token), 96 token = self._tokenizer.TokenizeFile(input_lines) 97 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], []) 99 self.assertFalse(namespaces_info.IsExtraProvide(token), 108 token = self._tokenizer.TokenizeFile(input_lines) 109 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], []) 111 self.assertFalse(namespaces_info.IsExtraProvide(token), 117 token = self._tokenizer.TokenizeFile(input_lines) [all …]
|
D | error_fixer.py | 90 for token in tokens: 91 self._file_changed_lines.add(token.line_number) 100 token = error.token 103 iterator = token.attached_object.type_start_token 112 while iterator and iterator != token.attached_object.type_end_token.next: 118 token.attached_object = javascriptstatetracker.JsDocFlag(token) 119 self._AddFix(token) 122 iterator = token.attached_object.type_end_token 131 token.attached_object = javascriptstatetracker.JsDocFlag(token) 132 self._AddFix(token) [all …]
|
D | javascriptlintrules.py | 57 def HandleMissingParameterDoc(self, token, param_name): argument 60 'Missing docs for parameter: "%s"' % param_name, token) 62 def __ContainsRecordType(self, token): argument 74 token and token.type == Type.DOC_FLAG and 75 token.attached_object.type is not None and 76 token.attached_object.type.find('{') != token.string.rfind('{')) 78 def CheckToken(self, token, state): argument 85 if self.__ContainsRecordType(token): 92 super(JavaScriptLintRules, self).CheckToken(token, state) 99 if token.type == Type.SIMPLE_LVALUE: [all …]
|
D | requireprovidesorter.py | 46 def CheckProvides(self, token): argument 64 provide_tokens = self._GetRequireOrProvideTokens(token, 'goog.provide') 71 def CheckRequires(self, token): argument 89 require_tokens = self._GetRequireOrProvideTokens(token, 'goog.require') 96 def FixProvides(self, token): argument 103 self._GetRequireOrProvideTokens(token, 'goog.provide')) 105 def FixRequires(self, token): argument 112 self._GetRequireOrProvideTokens(token, 'goog.require')) 153 def _GetRequireOrProvideTokens(self, token, token_string): argument 166 while token: [all …]
|
/external/libxml2/ |
D | rngparser.c | 59 const xmlChar *token; member 92 tokenPtr token; member 252 tokenPtr token; in xmlCRNGNextToken() local 256 token = &(ctxt->tokens[(ctxt->firstToken + ctxt->nbTokens) % MAX_TOKEN]); in xmlCRNGNextToken() 257 token->toktype = CRNG_NONE; in xmlCRNGNextToken() 291 token->toklen = cur - ctxt->cur; in xmlCRNGNextToken() 292 token->token = xmlDictLookup(ctxt->dict, ctxt->cur, token->toklen); in xmlCRNGNextToken() 293 token->toktype = CRNG_LITERAL_SEGMENT; in xmlCRNGNextToken() 294 token->prefix = NULL; in xmlCRNGNextToken() 308 token->toklen = 2; in xmlCRNGNextToken() [all …]
|
/external/doclava/src/com/google/doclava/apicheck/ |
D | ApiFile.java | 71 String token = tokenizer.getToken(); in parseApi() local 72 if (token == null) { in parseApi() 75 if ("package".equals(token)) { in parseApi() 78 throw new ApiParseException("expected package got " + token, tokenizer.getLine()); in parseApi() 90 String token; in parsePackage() local 94 token = tokenizer.requireToken(); in parsePackage() 95 assertIdent(tokenizer, token); in parsePackage() 96 name = token; in parsePackage() 98 token = tokenizer.requireToken(); in parsePackage() 99 if (!"{".equals(token)) { in parsePackage() [all …]
|
/external/antlr/antlr-3.4/runtime/Python/tests/ |
D | t011lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.IDENTIFIER 29 assert token.start == 0, token.start 30 assert token.stop == 5, token.stop 31 assert token.text == 'foobar', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.WS 35 assert token.start == 6, token.start 36 assert token.stop == 6, token.stop 37 assert token.text == ' ', token.text [all …]
|
D | t010lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.IDENTIFIER 29 assert token.start == 0, token.start 30 assert token.stop == 5, token.stop 31 assert token.text == 'foobar', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.WS 35 assert token.start == 6, token.start 36 assert token.stop == 6, token.stop 37 assert token.text == ' ', token.text [all …]
|
D | t004lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.FOO 29 assert token.start == 0, token.start 30 assert token.stop == 0, token.stop 31 assert token.text == 'f', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.FOO 35 assert token.start == 1, token.start 36 assert token.stop == 2, token.stop 37 assert token.text == 'fo', token.text [all …]
|
D | t009lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.DIGIT 29 assert token.start == 0, token.start 30 assert token.stop == 0, token.stop 31 assert token.text == '0', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.DIGIT 35 assert token.start == 1, token.start 36 assert token.stop == 1, token.stop 37 assert token.text == '8', token.text [all …]
|
D | t008lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.FOO 29 assert token.start == 0, token.start 30 assert token.stop == 0, token.stop 31 assert token.text == 'f', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.FOO 35 assert token.start == 1, token.start 36 assert token.stop == 2, token.stop 37 assert token.text == 'fa', token.text [all …]
|
D | t005lexer.py | 27 token = lexer.nextToken() 28 assert token.type == self.lexerModule.FOO 29 assert token.start == 0, token.start 30 assert token.stop == 1, token.stop 31 assert token.text == 'fo', token.text 33 token = lexer.nextToken() 34 assert token.type == self.lexerModule.FOO 35 assert token.start == 2, token.start 36 assert token.stop == 4, token.stop 37 assert token.text == 'foo', token.text [all …]
|
/external/chromium-trace/trace-viewer/tracing/third_party/vinn/third_party/parse5/lib/tree_construction/ |
D | parser.js | 431 var token = this.tokenizer.getNextToken(); 436 if (token.type === Tokenizer.WHITESPACE_CHARACTER_TOKEN && token.chars[0] === '\n') { 437 if (token.chars.length === 1) 440 token.chars = token.chars.substr(1); 444 if (this._shouldProcessTokenInForeignContent(token)) 445 this._processTokenInForeignContent(token); 448 this._processToken(token); 512 Parser.prototype._setDocumentType = function (token) { argument 513 this.treeAdapter.setDocumentType(this.document, token.name, token.publicId, token.systemId); 527 Parser.prototype._appendElement = function (token, namespaceURI) { argument [all …]
|
/external/apache-xml/src/main/java/org/apache/xml/utils/ |
D | StylesheetPIHandler.java | 156 String token = ""; in processingInstruction() local 160 token = tokenizer.nextToken(); in processingInstruction() 164 (token.equals(" ") || token.equals("\t") || token.equals("="))) in processingInstruction() 167 String name = token; in processingInstruction() 170 token = tokenizer.nextToken(); in processingInstruction() 172 (token.equals(" " ) || token.equals("\t") || token.equals("="))) in processingInstruction() 173 token = tokenizer.nextToken(); in processingInstruction() 174 type = token.substring(1, token.length() - 1); in processingInstruction() 179 token = tokenizer.nextToken(); in processingInstruction() 181 (token.equals(" " ) || token.equals("\t") || token.equals("="))) in processingInstruction() [all …]
|
/external/antlr/antlr-3.4/runtime/ObjC/Framework/test/runtime/token/ |
D | ANTLRCommonTokenTest.m | 17 ANTLRCommonToken *token = [[ANTLRCommonToken newToken] retain]; 18 STAssertNotNil(token, @"Token was nil"); 19 [token release]; 24 ANTLRCommonToken *token = [[ANTLRCommonToken eofToken] retain]; 25 STAssertNotNil(token, @"Token was nil"); 26 …STAssertEquals(token.type, (NSInteger)ANTLRTokenTypeEOF, @"Token was not of type ANTLRTokenTypeEOF… 27 [token release]; 32 ANTLRCommonToken *token = [[ANTLRCommonToken newToken:ANTLRTokenTypeUP] retain]; 33 token.text = @"<UP>"; 34 STAssertNotNil(token, @"Token was nil"); [all …]
|
/external/jsoncpp/src/lib_json/ |
D | json_reader.cpp | 163 Token token; in parse() local 164 skipCommentTokens( token ); in parse() 172 token.type_ = tokenError; in parse() 173 token.start_ = beginDoc; in parse() 174 token.end_ = endDoc; in parse() 176 token ); in parse() 187 Token token; in readValue() local 188 skipCommentTokens( token ); in readValue() 198 switch ( token.type_ ) in readValue() 201 successful = readObject( token ); in readValue() [all …]
|
/external/owasp/sanitizer/src/main/org/owasp/html/ |
D | CssGrammar.java | 79 if (!(it.hasTokenAfterSpace() && ":".equals(it.token()))) { in parsePropertyGroup() 96 String token = it.token(); in parsePropertyValue() local 103 handler.startFunction(token); in parsePropertyValue() 105 handler.endFunction(token); in parsePropertyValue() 108 handler.identifier(token); in parsePropertyValue() 111 if (token.length() == 4 || token.length() == 7) { in parsePropertyValue() 112 handler.hash(token); in parsePropertyValue() 116 handler.quotedString(token); in parsePropertyValue() 119 handler.url(token); in parsePropertyValue() 124 handler.quantity(token); in parsePropertyValue() [all …]
|
/external/libcxx/test/std/numerics/rand/rand.device/ |
D | ctor.pass.cpp | 25 bool is_valid_random_device(const std::string &token) { in is_valid_random_device() argument 29 return token == "/dev/urandom"; in is_valid_random_device() 32 return token == "/dev/urandom" || token == "/dev/random"; in is_valid_random_device() 36 void check_random_device_valid(const std::string &token) { in check_random_device_valid() argument 37 std::random_device r(token); in check_random_device_valid() 40 void check_random_device_invalid(const std::string &token) { in check_random_device_invalid() argument 42 std::random_device r(token); in check_random_device_invalid() 63 std::string token = "wrong file"; in main() local 64 if (is_valid_random_device(token)) in main() 65 check_random_device_valid(token); in main() [all …]
|
/external/google-breakpad/src/testing/scripts/generator/cpp/ |
D | ast.py | 286 for token in self.alias: 287 if token is not None and name == token.name: 341 for token in token_list: 342 if token.name == node.name: 466 token = tokens[end] 468 if token.name == '<': 470 elif token.name == '>': 506 token = tokens[i] 507 if token.name == '<': 514 elif token.name == ',': [all …]
|