/external/chromium-trace/catapult/third_party/closure_linter/closure_linter/common/ |
H A D | matcher.py | 46 def __init__(self, regex, token_type, result_mode=None, line_start=False): 51 token_type: The type of token a successful match indicates. 58 self.type = token_type
|
H A D | tokens.py | 50 def __init__(self, string, token_type, line, line_number, values=None, 56 token_type: The type of token. 66 self.type = token_type 103 def IsType(self, token_type): 107 token_type: The type to test for. 112 return self.type == token_type
|
H A D | tokenizer.py | 78 def _CreateToken(self, string, token_type, line, line_number, values=None): 83 token_type: The type of token. 93 return tokens.Token(string, token_type, line, line_number, values,
|
/external/autotest/client/common_lib/cros/ |
H A D | dbus_send.py | 47 token_type = token_stream.pop(0) 48 if token_type == 'variant': 49 token_type = token_stream.pop(0) 50 if token_type == 'object': 51 token_type = token_stream.pop(0) # Should be 'path' 55 if token_type in INT_TYPES: 57 if token_type == 'string' or token_type == 'path': 59 if token_type == 'boolean': 61 if token_type [all...] |
/external/chromium-trace/catapult/third_party/html5lib-python/html5lib/ |
H A D | sanitizer.py | 182 # accommodate filters which use token_type differently 183 token_type = token["type"] 184 if token_type in list(tokenTypes.keys()): 185 token_type = tokenTypes[token_type] 187 if token_type in (tokenTypes["StartTag"], tokenTypes["EndTag"], 190 return self.allowed_token(token, token_type) 192 return self.disallowed_token(token, token_type) 193 elif token_type == tokenTypes["Comment"]: 198 def allowed_token(self, token, token_type) [all...] |
/external/google-breakpad/src/testing/scripts/generator/cpp/ |
H A D | tokenize.py | 74 def __init__(self, token_type, name, start, end): 75 self.token_type = token_type 147 token_type = UNKNOWN 151 token_type = NAME 159 token_type = CONSTANT 162 token_type = CONSTANT 173 token_type = SYNTAX 183 token_type = SYNTAX 186 token_type [all...] |
H A D | ast.py | 549 if parts[-1].token_type == tokenize.NAME: 579 if (type_name and type_name[-1].token_type == tokenize.NAME and 580 p.token_type == tokenize.NAME): 713 if token.token_type == _INTERNAL_TOKEN: 738 if token.token_type == tokenize.NAME: 749 if next.token_type == tokenize.SYNTAX and next.name == '(': 805 elif token.token_type == tokenize.SYNTAX: 811 if (token.token_type == tokenize.NAME and 815 elif token.token_type == tokenize.PREPROCESSOR: 854 while (last_token.token_type ! [all...] |
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/ |
H A D | offset_token.py | 26 def __init__(self, token_type, string, offset): 27 self._type = token_type 70 token_type, string, (srow, scol), _, _ = next_token 72 offset_tokens.append(OffsetToken(token_type, string, (0, 0))) 76 offset_tokens.append(OffsetToken(token_type, string, (0, scol - ecol))) 79 token_type, string, (srow - erow, scol)))
|
H A D | snippet.py | 97 def __init__(self, token_type, tokens): 102 self._type = token_type 107 def Create(cls, token_type, string, offset=(0, 0)): 108 return cls(token_type, 109 [offset_token.OffsetToken(token_type, string, offset)])
|
/external/openfst/src/include/fst/extensions/far/ |
H A D | print-strings.h | 46 typename StringPrinter<Arc>::TokenType token_type; local 48 token_type = StringPrinter<Arc>::SYMBOL; 50 token_type = StringPrinter<Arc>::BYTE; 52 token_type = StringPrinter<Arc>::UTF8; 95 token_type, syms ? syms : fst.InputSymbols()); 127 if (token_type == StringPrinter<Arc>::SYMBOL)
|
H A D | compile-strings.h | 54 TokenType token_type, 59 token_type_(token_type), symbols_(syms), done_(false), 60 compiler_(token_type, syms, unknown_label, allow_negative_labels) { 166 typename StringCompiler<Arc>::TokenType token_type; local 168 token_type = StringCompiler<Arc>::SYMBOL; 170 token_type = StringCompiler<Arc>::BYTE; 172 token_type = StringCompiler<Arc>::UTF8; 246 entry_type, token_type, allow_negative_labels, 260 << " in file " << inputs[i] << " failed with token_type = " 51 StringReader(istream &istrm, const string &source, EntryType entry_type, TokenType token_type, bool allow_negative_labels, const SymbolTable *syms = 0, Label unknown_label = kNoStateId) argument
|
H A D | farscript.h | 211 const FarTokenType token_type; member in struct:fst::script::FarPrintStringsArgs 224 const FarTokenType token_type, const string &begin_key, 229 ifilenames(ifilenames), entry_type(entry_type), token_type(token_type), 240 args->ifilenames, args->entry_type, args->token_type, 250 const FarTokenType token_type, 222 FarPrintStringsArgs( const vector<string> &ifilenames, const FarEntryType entry_type, const FarTokenType token_type, const string &begin_key, const string &end_key, const bool print_key, const bool print_weight, const string &symbols_fname, const bool initial_symbols, const int32 generate_filenames, const string &filename_prefix, const string &filename_suffix) argument
|
/external/protobuf/gtest/scripts/ |
H A D | pump.py | 141 def __init__(self, start=None, end=None, value=None, token_type=None): 151 self.token_type = token_type 155 self.start, self.value, self.token_type) 161 self.token_type) 172 for (regex, token_type) in token_table: 179 best_match_token_type = token_type 198 (start_column, length, token_type) = m 203 return MakeToken(lines, found_start, found_end, token_type) 227 def MakeToken(lines, start, end, token_type) [all...] |
/external/google-breakpad/src/testing/gtest/scripts/ |
H A D | pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type) [all...] |
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/ |
H A D | pump.py | 140 def __init__(self, start=None, end=None, value=None, token_type=None): 150 self.token_type = token_type 154 self.start, self.value, self.token_type) 160 self.token_type) 171 for (regex, token_type) in token_table: 178 best_match_token_type = token_type 197 (start_column, length, token_type) = m 202 return MakeToken(lines, found_start, found_end, token_type) 237 def MakeToken(lines, start, end, token_type) [all...] |
/external/chromium-trace/catapult/third_party/closure_linter/closure_linter/ |
H A D | ecmametadatapass.py | 311 token_type = token.type 326 if token_type == TokenType.START_PAREN: 334 elif token_type == TokenType.END_PAREN: 362 elif (token_type == TokenType.KEYWORD and 370 elif token_type == TokenType.START_PARAMETERS: 373 elif token_type == TokenType.END_PARAMETERS: 376 elif token_type == TokenType.START_BRACKET: 383 elif token_type == TokenType.END_BRACKET: 386 elif token_type == TokenType.START_BLOCK: 401 elif token_type [all...] |
H A D | indentation.py | 139 token_type = token.type 145 if token_type == Type.END_PAREN: 148 elif token_type == Type.END_PARAMETERS: 151 elif token_type == Type.END_BRACKET: 154 elif token_type == Type.END_BLOCK: 181 elif token_type == Type.KEYWORD and token.string in ('case', 'default'): 184 elif token_type == Type.SEMICOLON: 188 token_type not in (Type.COMMENT, Type.DOC_PREFIX, Type.STRING_TEXT)): 202 if token_type in Type.COMMENT_TYPES: 221 if token_type [all...] |
H A D | ecmalintrules.py | 314 token_type = token.type 326 if token_type == Type.PARAMETERS: 341 elif (token_type == Type.START_BLOCK and 345 elif token_type == Type.END_BLOCK: 385 elif token_type == Type.SEMICOLON: 422 elif token_type == Type.START_PAREN: 446 elif token_type == Type.START_BRACKET: 448 elif token_type in (Type.END_PAREN, Type.END_BRACKET): 462 elif token_type == Type.WHITESPACE: 494 elif token_type [all...] |
/external/bison/examples/calc++/ |
H A D | calc++-driver.hh | 10 yy::calcxx_parser::token_type \
|
/external/v8/src/parsing/ |
H A D | token.cc | 35 const char Token::token_type[] = { member in class:v8::internal::Token
|
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/tree/ |
H A D | wizard.rb | 197 @token_type = tokenizer.next_token 201 case @token_type 205 @token_type == EOF and return node 214 @token_type != :open and return nil 215 @token_type = @tokenizer.next_token 219 case @token_type 229 @token_type == :close or return nil 230 @token_type = @tokenizer.next_token 236 if @token_type == :percent 237 ( @token_type [all...] |
H A D | debug.rb | 27 def create_from_token( token_type, from_token, text = nil ) 33 def create_from_type( token_type, text )
|
/external/openfst/src/extensions/far/ |
H A D | farcompilestrings.cc | 43 DEFINE_string(token_type, "symbol", "Token type: one of : "
|
H A D | farprintstrings.cc | 37 DEFINE_string(token_type, "symbol", "Token type: one of : "
|
H A D | farscript.cc | 105 const FarTokenType token_type, 115 FarPrintStringsArgs args(ifilenames, entry_type, token_type, 102 FarPrintStrings(const vector<string> &ifilenames, const string &arc_type, const FarEntryType entry_type, const FarTokenType token_type, const string &begin_key, const string &end_key, const bool print_key, const bool print_weight, const string &symbols_fname, const bool initial_symbols, const int32 generate_filenames, const string &filename_prefix, const string &filename_suffix) argument
|