/external/chromium_org/chrome/browser/ui/app_list/search/ |
H A D | tokenized_string_char_iterator_unittest.cc | 41 TokenizedString tokens(UTF8ToUTF16(text)); 42 TokenizedStringCharIterator iter(tokens); 55 TokenizedString tokens(UTF8ToUTF16(text)); 56 TokenizedStringCharIterator iter(tokens); 70 TokenizedString tokens(UTF8ToUTF16(text)); 71 TokenizedStringCharIterator iter(tokens);
|
/external/chromium/testing/gmock/scripts/generator/cpp/ |
H A D | ast.py | 340 # TODO(nnorwitz): bases are tokens, do name comparision. 383 # TODO(nnorwitz): parameters are tokens, do name comparision. 462 def _GetTemplateEnd(self, tokens, start): 466 token = tokens[end] 474 return tokens[start:end-1], end 476 def ToType(self, tokens): 489 # Partition tokens into name and modifier tokens. 504 end = len(tokens) 506 token = tokens[ [all...] |
/external/antlr/antlr-3.4/runtime/CSharp3/Sources/Antlr3.Runtime.Test/Composition/ |
H A D | Program.cs | 38 var tokens = new CommonTokenStream(lex); 39 var g = new VecMathParser(tokens);
|
/external/antlr/antlr-3.4/runtime/Python/unittests/ |
H A D | testdottreegen.py | 20 self.tokens = [ 23 self.wiz = TreeWizard(self.adaptor, self.tokens)
|
H A D | testtreewizard.py | 138 self.tokens = [ 141 self.wizard = TreeWizard(self.adaptor, tokenNames=self.tokens) 240 self.tokens = [ 265 tokenNames=self.tokens 284 wiz = TreeWizard(self.adaptor, self.tokens) 292 wiz = TreeWizard(self.adaptor, self.tokens) 300 wiz = TreeWizard(self.adaptor, self.tokens) 308 wiz = TreeWizard(self.adaptor, self.tokens) 316 wiz = TreeWizard(self.adaptor, self.tokens) 324 wiz = TreeWizard(self.adaptor, self.tokens) [all...] |
/external/antlr/antlr-3.4/tool/src/test/java/org/antlr/test/ |
H A D | TestTreeIterator.java | 36 static final String[] tokens = new String[] { field in class:TestTreeIterator 42 TreeWizard wiz = new TreeWizard(adaptor, tokens); 53 TreeWizard wiz = new TreeWizard(adaptor, tokens); 64 TreeWizard wiz = new TreeWizard(adaptor, tokens); 75 TreeWizard wiz = new TreeWizard(adaptor, tokens); 86 TreeWizard wiz = new TreeWizard(adaptor, tokens); 97 TreeWizard wiz = new TreeWizard(adaptor, tokens); 108 TreeWizard wiz = new TreeWizard(adaptor, tokens);
|
/external/chromium_org/chrome/browser/extensions/api/log_private/ |
H A D | syslog_parser.cc | 45 << "Error when parsing data. Expect: At least 3 tokens. Actual: 0"; 55 << "Error when parsing data. Expect: At least 3 tokens. Actual: 1"; 60 << "Error when parsing data. Expect: At least 3 tokens. Actual: 2"; 87 std::string tokens[kExpectedTimeTokenNum]; local 89 if (ParseTimeHelper(&tokenizer, &(tokens[i])) != SyslogParser::SUCCESS) 93 std::string buffer = tokens[1] + '-' + tokens[2] + '-' + tokens[0] + ' ' + 94 tokens[3] + ':' + tokens[ [all...] |
/external/chromium_org/chrome/renderer/safe_browsing/ |
H A D | phishing_url_feature_extractor.h | 6 // the client-side phishing detection model. These include tokens in the 38 // tokens. 41 std::vector<std::string>* tokens);
|
/external/chromium_org/net/ftp/ |
H A D | ftp_util.cc | 38 std::vector<std::string> tokens; local 40 tokens.push_back(tokenizer.token()); 45 if (tokens.empty()) { 50 if (tokens.size() == 1) 53 std::string result(tokens[0] + ":["); 54 if (tokens.size() == 2) { 58 result.append(tokens[1]); 59 for (size_t i = 2; i < tokens.size() - 1; i++) 60 result.append("." + tokens[i]); 62 result.append("]" + tokens[token [all...] |
/external/chromium_org/third_party/mesa/src/src/gallium/auxiliary/draw/ |
H A D | draw_fs.c | 50 tgsi_scan_shader(shader->tokens, &dfs->info);
|
/external/mesa3d/src/gallium/auxiliary/draw/ |
H A D | draw_fs.c | 50 tgsi_scan_shader(shader->tokens, &dfs->info);
|
/external/robolectric/src/main/java/com/xtremelabs/robolectric/shadows/ |
H A D | ShadowTextUtils.java | 47 public static String join(CharSequence delimiter, Iterable tokens) { argument 48 return Join.join((String) delimiter, (Collection) tokens); 52 public static String join(CharSequence delimiter, Object[] tokens) { argument 53 return Join.join((String) delimiter, tokens);
|
/external/antlr/antlr-3.4/gunit/src/main/java/org/antlr/gunit/swingui/model/ |
H A D | TestSuite.java | 38 protected CommonTokenStream tokens; field in class:TestSuite 84 public CommonTokenStream getTokens() { return tokens; } 86 public void setTokens(CommonTokenStream ts) { tokens = ts; }
|
/external/antlr/antlr-3.4/runtime/Ruby/test/functional/token-rewrite/ |
H A D | via-parser.rb | 92 tokens = ANTLR3::TokenRewriteStream.new( lexer ) 93 parser = TokenRewrites::Parser.new( tokens ) 96 tokens.render.strip.should == expected_output
|
/external/chromium-trace/trace-viewer/build/ |
H A D | parse_deps_test.py | 14 tokens = list(parse_deps._tokenize_js('')) 15 self.assertEquals([], tokens) 18 tokens = list(parse_deps._tokenize_js('\n')) 19 self.assertEquals(['\n'], tokens) 22 tokens = list(parse_deps._tokenize_js('A // foo')) 23 self.assertEquals(['A ', '//', ' foo'], tokens) 26 tokens = list(parse_deps._tokenize_js("""A // foo 29 self.assertEquals(['A ', '//', ' foo', '\n', 'bar'], tokens) 32 tokens = list(parse_deps._tokenize_js("""A /* foo */""")) 33 self.assertEquals(['A ', '/*', ' foo ', '*/'], tokens) [all...] |
/external/chromium_org/third_party/mesa/src/src/gallium/state_trackers/d3d1x/gd3d1x/tools/ |
H A D | dxbc2tgsi.cpp | 71 const struct tgsi_token* tokens = (const struct tgsi_token*)sm4_to_tgsi(*sm4); local 72 if(tokens) 76 tgsi_dump(tokens, 0);
|
/external/mesa3d/src/gallium/state_trackers/d3d1x/gd3d1x/tools/ |
H A D | dxbc2tgsi.cpp | 71 const struct tgsi_token* tokens = (const struct tgsi_token*)sm4_to_tgsi(*sm4); local 72 if(tokens) 76 tgsi_dump(tokens, 0);
|
/external/nist-sip/java/gov/nist/javax/sip/header/ |
H A D | UserAgent.java | 50 /** Product tokens. 60 StringBuffer tokens = new StringBuffer(); 64 tokens.append((String) it.next()); 67 return tokens.toString();
|
/external/smack/src/org/xbill/DNS/ |
H A D | Address.java | 77 String [] tokens = s.split(":", -1); 80 int last = tokens.length - 1; 82 if (tokens[0].length() == 0) { 83 // If the first two tokens are empty, it means the string 86 if (last - first > 0 && tokens[1].length() == 0) 92 if (tokens[last].length() == 0) { 93 // If the last two tokens are empty, it means the string 96 if (last - first > 0 && tokens[last - 1].length() == 0) 107 if (tokens[i].length() == 0) { 114 if (tokens[ [all...] |
/external/antlr/antlr-3.4/runtime/Ruby/lib/antlr3/ |
H A D | streams.rb | 72 parsers with the means to sequential walk through series of tokens. 79 In a similar fashion to CommonTokenStream, CommonTreeNodeStream feeds tokens 82 the two-dimensional shape of the tree using special UP and DOWN tokens. The 99 is the <i>integer token type of the token</i> <tt>k</tt> tokens ahead of the 108 <b>TokenStreams</b>, this is the <i>full token structure</i> <tt>k</tt> tokens 282 sequence of tokens. Unlike simple character-based streams, such as StringStream, 287 <i>channel</i> feature, which allows you to hold on to all tokens of interest 288 while only presenting a specific set of interesting tokens to a parser. For 291 whitespace to channel value HIDDEN as it creates the tokens. 295 yield tokens tha 1044 def tokens( start = nil, stop = nil ) method in class:ANTLR3.that.CommonTokenStream [all...] |
/external/antlr/antlr-3.4/runtime/Java/src/main/java/org/antlr/runtime/tree/ |
H A D | CommonTreeNodeStream.java | 45 protected TokenStream tokens; field in class:CommonTreeNodeStream 108 public TokenStream getTokenStream() { return tokens; } 110 public void setTokenStream(TokenStream tokens) { this.tokens = tokens; } argument
|
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/common/ |
H A D | tokens_test.py | 20 from closure_linter.common import tokens namespace 24 return tokens.Token('foo', None, 1, 1) 82 a = tokens.Token('foo', 'fakeType1', 1, 1) 87 a = tokens.Token('foo', 'fakeType1', 1, 1) 92 a = tokens.Token('foo', 'fakeType1', 1, 1)
|
/external/chromium_org/third_party/closure_linter/closure_linter/common/ |
H A D | tokens_test.py | 20 from closure_linter.common import tokens namespace 24 return tokens.Token('foo', None, 1, 1) 82 a = tokens.Token('foo', 'fakeType1', 1, 1) 87 a = tokens.Token('foo', 'fakeType1', 1, 1) 92 a = tokens.Token('foo', 'fakeType1', 1, 1)
|
/external/chromium_org/third_party/mesa/src/src/gallium/auxiliary/tgsi/ |
H A D | tgsi_parse.c | 36 const struct tgsi_token *tokens ) 38 ctx->FullHeader.Header = *(struct tgsi_header *) &tokens[0]; 40 ctx->FullHeader.Processor = *(struct tgsi_processor *) &tokens[1]; 46 ctx->Tokens = tokens; 306 tgsi_dup_tokens(const struct tgsi_token *tokens) argument 308 unsigned n = tgsi_num_tokens(tokens); 312 memcpy(new_tokens, tokens, bytes); 318 * Allocate memory for num_tokens tokens. 329 tgsi_dump_tokens(const struct tgsi_token *tokens) argument 331 const unsigned *dwords = (const unsigned *)tokens; [all...] |
/external/libvpx/libvpx/vp9/encoder/ |
H A D | vp9_encodemb.c | 103 // to scan previous tokens to work out the correct context. 122 vp9_token_state tokens[1025][2]; local 154 tokens[eob][0].rate = 0; 155 tokens[eob][0].error = 0; 156 tokens[eob][0].next = default_eob; 157 tokens[eob][0].token = DCT_EOB_TOKEN; 158 tokens[eob][0].qc = 0; 159 *(tokens[eob] + 1) = *(tokens[eob] + 0); 173 error0 = tokens[nex [all...] |