Home
last modified time | relevance | path

Searched refs:token (Results 1 – 25 of 1095) sorted by relevance

12345678910>>...44

/external/antlr/antlr-3.4/runtime/Ruby/test/functional/lexer/
Dbasic.rb20 token = lexer.next_token
21 token.name.should == 'ZERO'
23 token = lexer.next_token
24 token.name.should == '<EOF>'
30 token_types = lexer.map { |token| token.name }
38 token = lexer.next_token
62 token = lexer.next_token
63 token.name.should == 'ZERO'
65 token = lexer.next_token
66 token.name.should == 'ONE'
[all …]
/external/antlr/antlr-3.4/runtime/C/src/
Dantlr3commontoken.c40 static pANTLR3_STRING getText (pANTLR3_COMMON_TOKEN token);
41 static void setText (pANTLR3_COMMON_TOKEN token, pANTLR3_STRING text);
42 static void setText8 (pANTLR3_COMMON_TOKEN token, pANTLR3_UINT8 text);
43 static ANTLR3_UINT32 getType (pANTLR3_COMMON_TOKEN token);
44 static void setType (pANTLR3_COMMON_TOKEN token, ANTLR3_UINT32 type);
45 static ANTLR3_UINT32 getLine (pANTLR3_COMMON_TOKEN token);
46 static void setLine (pANTLR3_COMMON_TOKEN token, ANTLR3_UINT32 line);
47 static ANTLR3_INT32 getCharPositionInLine (pANTLR3_COMMON_TOKEN token);
48 static void setCharPositionInLine (pANTLR3_COMMON_TOKEN token, ANTLR3_INT32 pos);
49 static ANTLR3_UINT32 getChannel (pANTLR3_COMMON_TOKEN token);
[all …]
/external/libxml2/doc/
Dapibuild.py405 def push(self, token): argument
406 self.tokens.insert(0, token);
413 def token(self): member in CLexer
632 def error(self, msg, token=-1): argument
637 if token != -1:
638 print("Got token ", token)
642 def debug(self, msg, token=-1): argument
644 if token != -1:
645 print("Got token ", token)
676 def parseComment(self, token): argument
[all …]
/external/chromium-trace/trace-viewer/tracing/third_party/closure_linter/closure_linter/
Dtokenutil.py32 def GetFirstTokenInSameLine(token): argument
41 while not token.IsFirstInLine():
42 token = token.previous
43 return token
46 def GetFirstTokenInPreviousLine(token): argument
56 first_in_line = GetFirstTokenInSameLine(token)
63 def GetLastTokenInSameLine(token): argument
72 while not token.IsLastInLine():
73 token = token.next
74 return token
[all …]
Dindentation.py81 def __init__(self, token, is_block=False): argument
88 self.token = token
92 self.is_transient = not is_block and not token.type in (
94 self.line_number = token.line_number
97 result = '\n %s' % self.token
100 result, self.overridden_by.token.string)
127 def CheckToken(self, token, state): argument
139 token_type = token.type
142 is_first = self._IsFirstNonWhitespaceTokenInLine(token)
158 goog_scope = self._GoogScopeOrNone(start_token.token)
[all …]
Decmalintrules.py101 def HandleMissingParameterDoc(self, token, param_name): argument
114 token = last_token
119 while token and token.line_number == line_number:
120 if state.IsTypeToken(token):
121 line.insert(0, 'x' * len(token.string))
122 elif token.type in (Type.IDENTIFIER, Type.NORMAL):
124 line.insert(0, token.string.replace('.', ' '))
126 line.insert(0, token.string)
127 token = token.previous
164 def _CheckJsDocType(self, token): argument
[all …]
Dclosurizednamespacesinfo_test.py76 token = self._GetRequireTokens('package.Something')
81 self.assertFalse(namespaces_info.IsExtraRequire(token),
87 self.assertTrue(namespaces_info.IsExtraRequire(token),
96 token = self._tokenizer.TokenizeFile(input_lines)
97 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], [])
99 self.assertFalse(namespaces_info.IsExtraProvide(token),
108 token = self._tokenizer.TokenizeFile(input_lines)
109 namespaces_info = self._GetInitializedNamespacesInfo(token, ['package'], [])
111 self.assertFalse(namespaces_info.IsExtraProvide(token),
117 token = self._tokenizer.TokenizeFile(input_lines)
[all …]
Derror_fixer.py90 for token in tokens:
91 self._file_changed_lines.add(token.line_number)
100 token = error.token
103 iterator = token.attached_object.type_start_token
112 while iterator and iterator != token.attached_object.type_end_token.next:
118 token.attached_object = javascriptstatetracker.JsDocFlag(token)
119 self._AddFix(token)
122 iterator = token.attached_object.type_end_token
131 token.attached_object = javascriptstatetracker.JsDocFlag(token)
132 self._AddFix(token)
[all …]
Djavascriptlintrules.py57 def HandleMissingParameterDoc(self, token, param_name): argument
60 'Missing docs for parameter: "%s"' % param_name, token)
62 def __ContainsRecordType(self, token): argument
74 token and token.type == Type.DOC_FLAG and
75 token.attached_object.type is not None and
76 token.attached_object.type.find('{') != token.string.rfind('{'))
78 def CheckToken(self, token, state): argument
85 if self.__ContainsRecordType(token):
92 super(JavaScriptLintRules, self).CheckToken(token, state)
99 if token.type == Type.SIMPLE_LVALUE:
[all …]
Drequireprovidesorter.py46 def CheckProvides(self, token): argument
64 provide_tokens = self._GetRequireOrProvideTokens(token, 'goog.provide')
71 def CheckRequires(self, token): argument
89 require_tokens = self._GetRequireOrProvideTokens(token, 'goog.require')
96 def FixProvides(self, token): argument
103 self._GetRequireOrProvideTokens(token, 'goog.provide'))
105 def FixRequires(self, token): argument
112 self._GetRequireOrProvideTokens(token, 'goog.require'))
153 def _GetRequireOrProvideTokens(self, token, token_string): argument
166 while token:
[all …]
/external/libxml2/
Drngparser.c59 const xmlChar *token; member
92 tokenPtr token; member
252 tokenPtr token; in xmlCRNGNextToken() local
256 token = &(ctxt->tokens[(ctxt->firstToken + ctxt->nbTokens) % MAX_TOKEN]); in xmlCRNGNextToken()
257 token->toktype = CRNG_NONE; in xmlCRNGNextToken()
291 token->toklen = cur - ctxt->cur; in xmlCRNGNextToken()
292 token->token = xmlDictLookup(ctxt->dict, ctxt->cur, token->toklen); in xmlCRNGNextToken()
293 token->toktype = CRNG_LITERAL_SEGMENT; in xmlCRNGNextToken()
294 token->prefix = NULL; in xmlCRNGNextToken()
308 token->toklen = 2; in xmlCRNGNextToken()
[all …]
/external/doclava/src/com/google/doclava/apicheck/
DApiFile.java71 String token = tokenizer.getToken(); in parseApi() local
72 if (token == null) { in parseApi()
75 if ("package".equals(token)) { in parseApi()
78 throw new ApiParseException("expected package got " + token, tokenizer.getLine()); in parseApi()
90 String token; in parsePackage() local
94 token = tokenizer.requireToken(); in parsePackage()
95 assertIdent(tokenizer, token); in parsePackage()
96 name = token; in parsePackage()
98 token = tokenizer.requireToken(); in parsePackage()
99 if (!"{".equals(token)) { in parsePackage()
[all …]
/external/antlr/antlr-3.4/runtime/Python/tests/
Dt011lexer.py27 token = lexer.nextToken()
28 assert token.type == self.lexerModule.IDENTIFIER
29 assert token.start == 0, token.start
30 assert token.stop == 5, token.stop
31 assert token.text == 'foobar', token.text
33 token = lexer.nextToken()
34 assert token.type == self.lexerModule.WS
35 assert token.start == 6, token.start
36 assert token.stop == 6, token.stop
37 assert token.text == ' ', token.text
[all …]
Dt010lexer.py27 token = lexer.nextToken()
28 assert token.type == self.lexerModule.IDENTIFIER
29 assert token.start == 0, token.start
30 assert token.stop == 5, token.stop
31 assert token.text == 'foobar', token.text
33 token = lexer.nextToken()
34 assert token.type == self.lexerModule.WS
35 assert token.start == 6, token.start
36 assert token.stop == 6, token.stop
37 assert token.text == ' ', token.text
[all …]
Dt004lexer.py27 token = lexer.nextToken()
28 assert token.type == self.lexerModule.FOO
29 assert token.start == 0, token.start
30 assert token.stop == 0, token.stop
31 assert token.text == 'f', token.text
33 token = lexer.nextToken()
34 assert token.type == self.lexerModule.FOO
35 assert token.start == 1, token.start
36 assert token.stop == 2, token.stop
37 assert token.text == 'fo', token.text
[all …]
Dt009lexer.py27 token = lexer.nextToken()
28 assert token.type == self.lexerModule.DIGIT
29 assert token.start == 0, token.start
30 assert token.stop == 0, token.stop
31 assert token.text == '0', token.text
33 token = lexer.nextToken()
34 assert token.type == self.lexerModule.DIGIT
35 assert token.start == 1, token.start
36 assert token.stop == 1, token.stop
37 assert token.text == '8', token.text
[all …]
Dt008lexer.py27 token = lexer.nextToken()
28 assert token.type == self.lexerModule.FOO
29 assert token.start == 0, token.start
30 assert token.stop == 0, token.stop
31 assert token.text == 'f', token.text
33 token = lexer.nextToken()
34 assert token.type == self.lexerModule.FOO
35 assert token.start == 1, token.start
36 assert token.stop == 2, token.stop
37 assert token.text == 'fa', token.text
[all …]
Dt005lexer.py27 token = lexer.nextToken()
28 assert token.type == self.lexerModule.FOO
29 assert token.start == 0, token.start
30 assert token.stop == 1, token.stop
31 assert token.text == 'fo', token.text
33 token = lexer.nextToken()
34 assert token.type == self.lexerModule.FOO
35 assert token.start == 2, token.start
36 assert token.stop == 4, token.stop
37 assert token.text == 'foo', token.text
[all …]
/external/chromium-trace/trace-viewer/tracing/third_party/vinn/third_party/parse5/lib/tree_construction/
Dparser.js431 var token = this.tokenizer.getNextToken();
436 if (token.type === Tokenizer.WHITESPACE_CHARACTER_TOKEN && token.chars[0] === '\n') {
437 if (token.chars.length === 1)
440 token.chars = token.chars.substr(1);
444 if (this._shouldProcessTokenInForeignContent(token))
445 this._processTokenInForeignContent(token);
448 this._processToken(token);
512 Parser.prototype._setDocumentType = function (token) { argument
513 this.treeAdapter.setDocumentType(this.document, token.name, token.publicId, token.systemId);
527 Parser.prototype._appendElement = function (token, namespaceURI) { argument
[all …]
/external/apache-xml/src/main/java/org/apache/xml/utils/
DStylesheetPIHandler.java156 String token = ""; in processingInstruction() local
160 token = tokenizer.nextToken(); in processingInstruction()
164 (token.equals(" ") || token.equals("\t") || token.equals("="))) in processingInstruction()
167 String name = token; in processingInstruction()
170 token = tokenizer.nextToken(); in processingInstruction()
172 (token.equals(" " ) || token.equals("\t") || token.equals("="))) in processingInstruction()
173 token = tokenizer.nextToken(); in processingInstruction()
174 type = token.substring(1, token.length() - 1); in processingInstruction()
179 token = tokenizer.nextToken(); in processingInstruction()
181 (token.equals(" " ) || token.equals("\t") || token.equals("="))) in processingInstruction()
[all …]
/external/antlr/antlr-3.4/runtime/ObjC/Framework/test/runtime/token/
DANTLRCommonTokenTest.m17 ANTLRCommonToken *token = [[ANTLRCommonToken newToken] retain];
18 STAssertNotNil(token, @"Token was nil");
19 [token release];
24 ANTLRCommonToken *token = [[ANTLRCommonToken eofToken] retain];
25 STAssertNotNil(token, @"Token was nil");
26 …STAssertEquals(token.type, (NSInteger)ANTLRTokenTypeEOF, @"Token was not of type ANTLRTokenTypeEOF…
27 [token release];
32 ANTLRCommonToken *token = [[ANTLRCommonToken newToken:ANTLRTokenTypeUP] retain];
33 token.text = @"<UP>";
34 STAssertNotNil(token, @"Token was nil");
[all …]
/external/jsoncpp/src/lib_json/
Djson_reader.cpp163 Token token; in parse() local
164 skipCommentTokens( token ); in parse()
172 token.type_ = tokenError; in parse()
173 token.start_ = beginDoc; in parse()
174 token.end_ = endDoc; in parse()
176 token ); in parse()
187 Token token; in readValue() local
188 skipCommentTokens( token ); in readValue()
198 switch ( token.type_ ) in readValue()
201 successful = readObject( token ); in readValue()
[all …]
/external/owasp/sanitizer/src/main/org/owasp/html/
DCssGrammar.java79 if (!(it.hasTokenAfterSpace() && ":".equals(it.token()))) { in parsePropertyGroup()
96 String token = it.token(); in parsePropertyValue() local
103 handler.startFunction(token); in parsePropertyValue()
105 handler.endFunction(token); in parsePropertyValue()
108 handler.identifier(token); in parsePropertyValue()
111 if (token.length() == 4 || token.length() == 7) { in parsePropertyValue()
112 handler.hash(token); in parsePropertyValue()
116 handler.quotedString(token); in parsePropertyValue()
119 handler.url(token); in parsePropertyValue()
124 handler.quantity(token); in parsePropertyValue()
[all …]
/external/libcxx/test/std/numerics/rand/rand.device/
Dctor.pass.cpp25 bool is_valid_random_device(const std::string &token) { in is_valid_random_device() argument
29 return token == "/dev/urandom"; in is_valid_random_device()
32 return token == "/dev/urandom" || token == "/dev/random"; in is_valid_random_device()
36 void check_random_device_valid(const std::string &token) { in check_random_device_valid() argument
37 std::random_device r(token); in check_random_device_valid()
40 void check_random_device_invalid(const std::string &token) { in check_random_device_invalid() argument
42 std::random_device r(token); in check_random_device_invalid()
63 std::string token = "wrong file"; in main() local
64 if (is_valid_random_device(token)) in main()
65 check_random_device_valid(token); in main()
[all …]
/external/google-breakpad/src/testing/scripts/generator/cpp/
Dast.py286 for token in self.alias:
287 if token is not None and name == token.name:
341 for token in token_list:
342 if token.name == node.name:
466 token = tokens[end]
468 if token.name == '<':
470 elif token.name == '>':
506 token = tokens[i]
507 if token.name == '<':
514 elif token.name == ',':
[all …]

12345678910>>...44