Home
last modified time | relevance | path

Searched refs:_tokenize (Results 1 – 12 of 12) sorted by relevance

/external/minijail/tools/
Dparser_unittest.py41 def _tokenize(line): member in TokenizerTests
49 for token in TokenizerTests._tokenize('@include /minijail.policy')
56 for token in TokenizerTests._tokenize('@include ./minijail.policy')
62 [(token.type, token.value) for token in TokenizerTests._tokenize(
94 TokenizerTests._tokenize('%invalid-token%')
105 def _tokenize(self, line): member in ParseConstantTests
112 self.parser.parse_value(self._tokenize('0x80000000')), 0x80000000)
115 self.parser.parse_value(self._tokenize('0x8000000000000000')),
123 self.parser.parse_value(self._tokenize('0x100000000'))
126 self.parser.parse_value(self._tokenize('0x10000000000000000'))
[all …]
/external/tensorflow/tensorflow/examples/saved_model/integration_tests/
Dexport_simple_text_embedding.py69 def _tokenize(self, sentences): member in TextEmbeddingModel
87 token_ids, token_values, token_dense_shape = self._tokenize(sentences)
Dexport_text_rnn_model.py48 def _tokenize(self, sentences): member in TextRnnModel
82 token_ids, token_values, token_dense_shape = self._tokenize(sentences)
/external/python/cpython3/Lib/
Dtokenize.py484 return _tokenize(chain(consumed, rl_gen, empty).__next__, encoding)
487 def _tokenize(readline, encoding): function
673 return _tokenize(readline, None)
709 tokens = _tokenize(sys.stdin.readline, None)
Dgettext.py86 def _tokenize(plural): function
183 result, nexttok = _parse(_tokenize(plural))
/external/python/cpython2/Lib/idlelib/
DEditorWindow.py1588 _tokenize = tokenize variable
1614 INDENT=_tokenize.INDENT,
1615 NAME=_tokenize.NAME,
1626 save_tabsize = _tokenize.tabsize
1627 _tokenize.tabsize = self.tabwidth
1630 _tokenize.tokenize(self.readline, self.tokeneater)
1631 except (_tokenize.TokenError, SyntaxError):
1636 _tokenize.tabsize = save_tabsize
/external/python/cpython2/Lib/
Dgettext.py84 def _tokenize(plural): function
177 result, nexttok = _parse(_tokenize(plural))
/external/python/cpython3/Lib/test/
Dtest_tokenize.py2 from tokenize import (tokenize, _tokenize, untokenize, NUMBER, NAME, OP,
/external/u-boot/tools/buildman/
Dkconfiglib.py560 return self._eval_expr(self._parse_expr(self._tokenize(s, True), # Feed
665 tokens = self._tokenize(line, False, line_feeder.filename,
852 tokens = self._tokenize(line, False, filename, linenr)
1216 def _tokenize(self, s, for_eval, filename=None, linenr=None): member in Config
/external/python/cpython3/Doc/library/
Dtokenize.rst136 .. _tokenize-cli:
/external/libchrome/third_party/jinja2/
Denvironment.py524 def _tokenize(self, source, name, filename=None, state=None): member in Environment
Dparser.py40 self.stream = environment._tokenize(source, name, filename, state)