/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/thirdparty/coverage/ |
H A D | phystokens.py | 3 import keyword, re, token, tokenize namespace 9 tokenize.generate_tokens() doesn't return a token for the backslash that 38 if last_ttype == tokenize.COMMENT: 77 ws_tokens = [token.INDENT, token.DEDENT, token.NEWLINE, tokenize.NL] 81 tokgen = tokenize.generate_tokens(StringIO(source).readline) 98 tok_class = tokenize.tok_name.get(ttype, 'xx').lower()[:3]
|
H A D | backward.py | 75 # Python 3.2 provides `tokenize.open`, the best way to open source files. 76 import tokenize namespace 78 open_source = tokenize.open # pylint: disable=E1101 81 detect_encoding = tokenize.detect_encoding # pylint: disable=E1101
|
H A D | parser.py | 3 import opcode, re, sys, token, tokenize namespace 105 tokgen = tokenize.generate_tokens(StringIO(self.text).readline) 109 tokenize.tok_name.get(toktype, toktype), 145 if ttext.strip() and toktype != tokenize.COMMENT:
|
/external/chromium_org/third_party/WebKit/Source/core/html/parser/ |
H A D | CSSPreloadScanner.cpp | 59 tokenize(*it, source); 79 inline void CSSPreloadScanner::tokenize(UChar c, const SegmentedString& source) function in class:blink::CSSPreloadScanner
|
/external/deqp/framework/randomshaders/ |
H A D | rsgShader.cpp | 93 void Shader::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::Shader 101 m_globalStatements[ndx]->tokenize(state, str); 107 m_functions[ndx]->tokenize(state, str); 112 m_mainFunction.tokenize(state, str); 125 void Function::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::Function 147 m_functionBlock.tokenize(state, str);
|
H A D | rsgBuiltinFunctions.hpp | 43 void tokenize (GeneratorState& state, TokenStream& str) const; 95 void UnaryBuiltinVecFunc<GetValueRangeWeight, ComputeValueRange, Evaluate>::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::UnaryBuiltinVecFunc 98 m_child->tokenize(state, str);
|
H A D | rsgExpression.hpp | 57 virtual void tokenize (GeneratorState& state, TokenStream& str) const = DE_NULL; 74 void tokenize (GeneratorState& state, TokenStream& str) const { DE_UNREF(state); str << Token(m_variable->getName()); } function in class:rsg::VariableAccess 113 void tokenize (GeneratorState& state, TokenStream& str) const; 131 void tokenize (GeneratorState& state, TokenStream& str) const; 149 void tokenize (GeneratorState& state, TokenStream& str) const; 167 void tokenize (GeneratorState& state, TokenStream& str) const; 189 void tokenize (GeneratorState& state, TokenStream& str) const; 214 void tokenize (GeneratorState& state, TokenStream& str) const; 233 void tokenize (GeneratorState& state, TokenStream& str) const; 255 void tokenize (GeneratorStat [all...] |
H A D | rsgBinaryOps.cpp | 93 void BinaryOp<Precedence, Assoc>::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::BinaryOp 95 m_leftValueExpr->tokenize(state, str); 97 m_rightValueExpr->tokenize(state, str);
|
H A D | rsgStatement.cpp | 203 void BlockStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::BlockStatement 208 (*i)->tokenize(state, str); 219 void ExpressionStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::ExpressionStatement 222 m_expression->tokenize(state, str); 333 void DeclarationStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::DeclarationStatement 340 m_expression->tokenize(state, str); 456 void ConditionalStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::ConditionalStatement 462 m_condition->tokenize(state, str); 469 m_trueStatement->tokenize(state, str); 473 m_trueStatement->tokenize(stat 556 void AssignStatement::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::AssignStatement [all...] |
H A D | rsgExpression.cpp | 411 void FloatLiteral::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::FloatLiteral 465 void IntLiteral::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::IntLiteral 502 void BoolLiteral::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::BoolLiteral 774 void ConstructorOp::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::ConstructorOp 786 (*i)->tokenize(state, str); 928 void AssignOp::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::AssignOp 930 m_lvalueExpr->tokenize(state, str); 932 m_rvalueExpr->tokenize(state, str); 1245 void ParenOp::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::ParenOp 1248 m_child->tokenize(stat 1334 void SwizzleOp::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::SwizzleOp 1544 void TexLookup::tokenize (GeneratorState& state, TokenStream& str) const function in class:rsg::TexLookup [all...] |
/external/apache-xml/src/main/java/org/apache/xpath/compiler/ |
H A D | Lexer.java | 96 void tokenize(String pat) throws javax.xml.transform.TransformerException method in class:Lexer 98 tokenize(pat, null); 109 void tokenize(String pat, Vector targetStrings) method in class:Lexer
|
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/source_frame/ |
H A D | CodeMirrorUtils.js | 144 function tokenize(line, callback) 154 return tokenize;
|
/external/chromium_org/third_party/webrtc/base/ |
H A D | stringencode.cc | 553 size_t tokenize(const std::string& source, char delimiter, function in namespace:rtc 577 tokenize(source, delimiter, &new_fields); 582 size_t tokenize(const std::string& source, char delimiter, char start_mark, function in namespace:rtc 600 // We have found the matching marks. First tokenize the pre-mask. Then add
|
/external/sqlite/android/ |
H A D | sqlite3_android.cpp | 258 static void tokenize(sqlite3_context * context, int argc, sqlite3_value ** argv) function 260 //ALOGD("enter tokenize"); 333 // Get the raw bytes for the string to tokenize 449 err = sqlite3_create_function(handle, "_TOKENIZE", 4, SQLITE_UTF16, collator, tokenize, NULL, NULL); 453 err = sqlite3_create_function(handle, "_TOKENIZE", 5, SQLITE_UTF16, collator, tokenize, NULL, NULL); 457 err = sqlite3_create_function(handle, "_TOKENIZE", 6, SQLITE_UTF16, collator, tokenize, NULL, NULL);
|
/external/chromium_org/third_party/WebKit/Source/core/css/parser/ |
H A D | MediaQueryTokenizer.cpp | 211 void MediaQueryTokenizer::tokenize(String string, Vector<MediaQueryToken>& outTokens) function in class:blink::MediaQueryTokenizer
|
/external/chromium_org/third_party/cython/src/Cython/Tempita/ |
H A D | _tempita.py | 40 import tokenize namespace 1030 tokens = tokenize.generate_tokens(StringIO(sig_text).readline) 1040 return tokenize.ENDMARKER, '' 1048 if tok_type == tokenize.ENDMARKER: 1050 if tok_type == tokenize.OP and (tok_string == '*' or tok_string == '**'): 1053 if tok_type != tokenize.NAME: 1058 if tok_type == tokenize.ENDMARKER or (tok_type == tokenize.OP and tok_string == ','): 1065 if tok_type == tokenize.ENDMARKER: 1071 if tok_type == tokenize [all...] |
/external/chromium_org/third_party/jinja2/ |
H A D | lexer.py | 542 def tokenize(self, source, name=None, filename=None, state=None): member in class:Lexer 543 """Calls tokeniter + tokenize and wraps it in a token stream. 549 """This is called with the stream as returned by `tokenize` and wraps 595 generator. Use this method if you just want to tokenize a template.
|
/external/chromium_org/third_party/WebKit/Source/devtools/front_end/script_formatter_worker/ |
H A D | ScriptFormatterWorker.js | 39 function tokenize(line, callback) 49 return tokenize;
|
/external/chromium_org/third_party/WebKit/Tools/Scripts/webkitpy/thirdparty/ |
H A D | pep8.py | 57 import tokenize namespace 91 NEWLINE = frozenset([tokenize.NL, tokenize.NEWLINE]) 92 SKIP_TOKENS = NEWLINE.union([tokenize.INDENT, tokenize.DEDENT]) 94 SKIP_COMMENTS = SKIP_TOKENS.union([tokenize.COMMENT, tokenize.ERRORTOKEN]) 115 COMMENT_WITH_NL = tokenize.generate_tokens(['#\n'].pop).send(None)[1] == '#\n' 447 close_bracket = (token_type == tokenize.OP and text in ']})') 504 if (parens[row] and token_type not in (tokenize [all...] |
H A D | autopep8.py | 57 import tokenize namespace 129 from lib2to3.pgen2 import tokenize as lib2to3_tokenize 221 token_type not in (tokenize.NL, tokenize.NEWLINE)) 232 close_bracket = (token_type == tokenize.OP and text in ']})') 286 if (parens[row] and token_type not in (tokenize.NL, tokenize.COMMENT) 291 elif (token_type in (tokenize.STRING, tokenize.COMMENT) or 302 if token_type == tokenize [all...] |
/external/e2fsprogs/e2fsck/ |
H A D | dict.c | 1227 static int tokenize(char *string, ...) function 1322 if (tokenize(in+1, &tok1, &tok2, (char **) 0) != 2) { 1393 if (tokenize(in+1, &tok1, &tok2, (char **) 0) != 2) { 1414 if (tokenize(in+1, &tok1, (char **) 0) != 1) { 1436 if (tokenize(in+1, &tok1, (char **) 0) != 1) { 1486 if (tokenize(in+1, &tok1, (char **) 0) != 1) { 1499 if (tokenize(in+1, &tok1, &tok2, (char **) 0) != 2) {
|
/external/chromium_org/remoting/webapp/ |
H A D | client_plugin_impl.js | 320 var tokenize = function(str) { 334 tokenize(getStringAttr(message.data, 'apiFeatures')); 342 tokenize(getStringAttr(message.data, 'requestedCapabilities')); 349 tokenize(getStringAttr(message.data, 'supportedCapabilities')); 448 var capabilities = tokenize(getStringAttr(message.data, 'capabilities'));
|
/external/owasp/sanitizer/lib/htmlparser-1.3/ |
H A D | htmlparser-1.3-with-transitions.jar | META-INF/MANIFEST.MF nu/validator/htmlparser/tools/XSLT4HTML5XOM.class XSLT4HTML5XOM.java package nu. ... |
H A D | htmlparser-1.3.jar | META-INF/MANIFEST.MF nu/validator/htmlparser/tools/XSLT4HTML5XOM.class XSLT4HTML5XOM.java package nu. ... |
/external/owasp/sanitizer/tools/findbugs/lib/ |
H A D | jdepend-2.9.jar | META-INF/ META-INF/MANIFEST.MF jdepend/ jdepend/framework/ jdepend/framework/AbstractParser.class AbstractParser. ... |