Searched refs:untokenize (Results 1 - 4 of 4) sorted by relevance
/external/python/cpython2/Lib/test/ |
H A D | test_tokenize.py | 2 from tokenize import (untokenize, generate_tokens, NUMBER, NAME, OP, namespace 531 return untokenize(result) 587 self.assertEqual(u.untokenize(iter([token])), 'Hello ') 594 Test roundtrip for `untokenize`. `f` is an open file or a string. 596 via tokenize.untokenize(), and tokenized again from the latter. 603 new_text = untokenize(tokens1) 622 # untokenize a little trickier. Note that this test involves trailing 704 return untokenize(tokens).decode('utf-8')
|
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/ |
H A D | offset_token.py | 115 return tokenize.untokenize(tokenize_tokens).replace('\\\n', ' \\\n')
|
/external/python/cpython2/Lib/lib2to3/pgen2/ |
H A D | tokenize.py | 38 "generate_tokens", "untokenize"] 197 def untokenize(self, iterable): member in class:Untokenizer 326 def untokenize(iterable): function 339 newcode = untokenize(t1) 345 return ut.untokenize(iterable)
|
/external/python/cpython2/Lib/ |
H A D | tokenize.py | 35 __all__ += ["COMMENT", "tokenize", "generate_tokens", "NL", "untokenize"] 199 def untokenize(self, iterable): member in class:Untokenizer 266 def untokenize(iterable): function 279 newcode = untokenize(t1) 285 return ut.untokenize(iterable)
|
Completed in 162 milliseconds