/external/minijail/tools/ |
D | parser_unittest.py | 41 def _tokenize(line): member in TokenizerTests 49 for token in TokenizerTests._tokenize('@include /minijail.policy') 56 for token in TokenizerTests._tokenize('@include ./minijail.policy') 62 [(token.type, token.value) for token in TokenizerTests._tokenize( 94 TokenizerTests._tokenize('%invalid-token%') 105 def _tokenize(self, line): member in ParseConstantTests 112 self.parser.parse_value(self._tokenize('0x80000000')), 0x80000000) 115 self.parser.parse_value(self._tokenize('0x8000000000000000')), 123 self.parser.parse_value(self._tokenize('0x100000000')) 126 self.parser.parse_value(self._tokenize('0x10000000000000000')) [all …]
|
/external/tensorflow/tensorflow/examples/saved_model/integration_tests/ |
D | export_simple_text_embedding.py | 69 def _tokenize(self, sentences): member in TextEmbeddingModel 87 token_ids, token_values, token_dense_shape = self._tokenize(sentences)
|
D | export_text_rnn_model.py | 48 def _tokenize(self, sentences): member in TextRnnModel 82 token_ids, token_values, token_dense_shape = self._tokenize(sentences)
|
/external/python/cpython3/Lib/ |
D | tokenize.py | 484 return _tokenize(chain(consumed, rl_gen, empty).__next__, encoding) 487 def _tokenize(readline, encoding): function 673 return _tokenize(readline, None) 709 tokens = _tokenize(sys.stdin.readline, None)
|
D | gettext.py | 86 def _tokenize(plural): function 183 result, nexttok = _parse(_tokenize(plural))
|
/external/python/cpython2/Lib/idlelib/ |
D | EditorWindow.py | 1588 _tokenize = tokenize variable 1614 INDENT=_tokenize.INDENT, 1615 NAME=_tokenize.NAME, 1626 save_tabsize = _tokenize.tabsize 1627 _tokenize.tabsize = self.tabwidth 1630 _tokenize.tokenize(self.readline, self.tokeneater) 1631 except (_tokenize.TokenError, SyntaxError): 1636 _tokenize.tabsize = save_tabsize
|
/external/python/cpython2/Lib/ |
D | gettext.py | 84 def _tokenize(plural): function 177 result, nexttok = _parse(_tokenize(plural))
|
/external/python/cpython3/Lib/test/ |
D | test_tokenize.py | 2 from tokenize import (tokenize, _tokenize, untokenize, NUMBER, NAME, OP,
|
/external/u-boot/tools/buildman/ |
D | kconfiglib.py | 560 return self._eval_expr(self._parse_expr(self._tokenize(s, True), # Feed 665 tokens = self._tokenize(line, False, line_feeder.filename, 852 tokens = self._tokenize(line, False, filename, linenr) 1216 def _tokenize(self, s, for_eval, filename=None, linenr=None): member in Config
|
/external/python/cpython3/Doc/library/ |
D | tokenize.rst | 136 .. _tokenize-cli:
|
/external/libchrome/third_party/jinja2/ |
D | environment.py | 524 def _tokenize(self, source, name, filename=None, state=None): member in Environment
|
D | parser.py | 40 self.stream = environment._tokenize(source, name, filename, state)
|