/external/mesa3d/src/glsl/glcpp/tests/ |
H A D | 000-content-with-spaces.c | 1 this is four tokens
|
/external/antlr/antlr-3.4/runtime/Python/tests/ |
H A D | t048rewrite.py | 25 tokens = self._parse("abc") 26 tokens.insertBefore(0, "0") 28 result = tokens.toString() 34 tokens = self._parse("abc") 35 tokens.insertAfter(2, "x") 37 result = tokens.toString() 43 tokens = self._parse("abc") 44 tokens.insertBefore(1, "x") 45 tokens.insertAfter(1, "x") 47 result = tokens [all...] |
/external/apache-harmony/regex/src/test/java/org/apache/harmony/tests/java/util/regex/ |
H A D | SplitTest.java | 45 String tokens[]; 47 tokens = p.split(input, 1); 48 assertEquals(1, tokens.length); 49 assertTrue(tokens[0].equals(input)); 50 tokens = p.split(input, 2); 51 assertEquals(2, tokens.length); 52 assertEquals("poodle", tokens[0]); 53 assertEquals("zoo", tokens[1]); 54 tokens = p.split(input, 5); 55 assertEquals(2, tokens [all...] |
/external/antlr/antlr-3.4/runtime/Python/antlr3/ |
H A D | __init__.py | 152 from tokens import *
|
/external/qemu/ |
H A D | json-parser.h | 20 QObject *json_parser_parse(QList *tokens, va_list *ap);
|
/external/clang/bindings/python/tests/cindex/ |
H A D | test_tokens.py | 15 tokens = list(tu.get_tokens(extent=r)) 17 assert len(tokens) == 5 18 assert tokens[1].spelling == 'i' 19 assert tokens[1].kind == TokenKind.IDENTIFIER 21 cursor = tokens[1].cursor 23 assert tokens[1].cursor == tokens[2].cursor 31 tokens = list(tu.get_tokens(extent=r)) 32 eq_(len(tokens), 4) 34 loc = tokens[ [all...] |
/external/antlr/antlr-3.4/tool/src/test/java/org/antlr/test/ |
H A D | TestCommonTokenStream.java | 52 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); 54 String result = tokens.LT(1).getText(); 73 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); 75 String result = tokens.LT(2).getText(); 94 BufferedTokenStream tokens = new BufferedTokenStream(lexEngine); 97 Token t = tokens.LT(i); 100 t = tokens.LT(i); 102 tokens.LT(i++); // push it past end 103 tokens.LT(i++); 105 String result = tokens [all...] |
H A D | TestTokenRewriteStream.java | 51 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); 52 tokens.insertBefore(0, "0"); 53 String result = tokens.toString(); 66 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); 67 tokens.insertAfter(2, "x"); 68 String result = tokens.toString(); 81 TokenRewriteStream tokens = new TokenRewriteStream(lexEngine); 82 tokens.fill(); 83 tokens.insertBefore(1, "x"); 84 tokens [all...] |
/external/chromium/net/tools/flip_server/ |
H A D | balsa_headers_token_utils.cc | 13 BalsaHeaders::HeaderTokenList* tokens) { 14 CHECK(tokens); 24 ParseTokenList(value_begin, line_end, tokens); 46 BalsaHeaders::HeaderTokenList tokens; local 47 TokenizeHeaderLine(*headers, *header_line, &tokens); 49 if (tokens.empty()) { 53 } else if (tokens.size() == 1) { 57 const base::StringPiece& new_last_token = tokens[tokens.size() - 2]; 84 BalsaHeaders::HeaderTokenList tokens; local 10 TokenizeHeaderLine( const BalsaHeaders& headers, const BalsaHeaders::HeaderLineDescription& header_line, BalsaHeaders::HeaderTokenList* tokens) argument 92 TokenizeHeaderValue( const BalsaHeaders& headers, const base::StringPiece& key, BalsaHeaders::HeaderTokenList* tokens) argument 109 ParseTokenList( const char* start, const char* end, BalsaHeaders::HeaderTokenList* tokens) argument [all...] |
H A D | balsa_headers_token_utils.h | 5 // Utility class that performs basic operations on header value tokens: parsing 6 // them out, checking for presense of certain tokens, and removing them. 30 // with that key, all of them will be tokenized and tokens will be added to 34 BalsaHeaders::HeaderTokenList* tokens); 43 // in some buffer, populates tokens list with beginning and end indices 44 // of all tokens present in the value string. 47 BalsaHeaders::HeaderTokenList* tokens); 54 BalsaHeaders::HeaderTokenList* tokens);
|
/external/webkit/Source/WebCore/loader/ |
H A D | FTPDirectoryParser.cpp | 119 const char *tokens[16]; /* 16 is more than enough */ local 120 unsigned int toklen[WTF_ARRAY_LENGTH(tokens)]; 130 tokens[0] = state.carryBuffer; 136 while (pos < linelen && numtoks < WTF_ARRAY_LENGTH(tokens)) 143 tokens[numtoks] = &line[pos]; 147 if (tokens[numtoks] != &line[pos]) 149 toklen[numtoks] = (&line[pos] - tokens[numtoks]); 158 linelen_sans_wsp = &(tokens[numtoks-1][toklen[numtoks-1]]) - tokens[0]; 159 if (numtoks == WTF_ARRAY_LENGTH(tokens)) [all...] |
/external/jmonkeyengine/engine/src/core-plugins/com/jme3/font/plugins/ |
H A D | BitmapFontLoader.java | 64 String[] tokens = line.split(regex); 65 if (tokens[0].equals("info")){ 67 for (int i = 1; i < tokens.length; i++){ 68 if (tokens[i].equals("size")){ 69 charSet.setRenderedSize(Integer.parseInt(tokens[i + 1])); 72 }else if (tokens[0].equals("common")){ 74 for (int i = 1; i < tokens.length; i++){ 75 String token = tokens[i]; 77 charSet.setLineHeight(Integer.parseInt(tokens[i + 1])); 79 charSet.setBase(Integer.parseInt(tokens[ [all...] |
/external/clang/test/Preprocessor/ |
H A D | macro_paste_msextensions.c | 12 #define comment /##/ dead tokens live here 19 #define nested(x) int x comment cute little dead tokens... 21 nested(baz) rise of the dead tokens
|
/external/antlr/antlr-3.4/runtime/Python/unittests/ |
H A D | testrecognizers.py | 33 self.tokens = [ 42 return self.tokens.pop(0) 46 tokens = [] 48 tokens.append(token.type) 50 self.failUnlessEqual(tokens, [1, 2, 3, 4])
|
/external/antlr/antlr-3.4/runtime/JavaScript/tests/functional/ |
H A D | rhino-python.extensions | 33 tokens should be sent to the parser sometimes without a corresponding 48 This TokenStream normally just passes tokens through to the parser. 63 A queue of tokens is built up to hold multiple DEDENT tokens that 77 /** The queue of tokens */ 78 this.tokens = []; 117 if (this.tokens.length>0 ) { 118 var t = this.tokens[0]; 119 this.tokens.splice(0,1); 137 this.tokens [all...] |
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
H A D | requireprovidesorter.py | 49 Iterates over tokens in given token stream, identifies goog.provide tokens, 54 token: A token in the token stream before any goog.provide tokens. 74 Iterates over tokens in given token stream, identifies goog.require tokens, 79 token: A token in the token stream before any goog.require tokens. 114 def _FixProvidesOrRequires(self, tokens): 118 tokens: A list of goog.provide or goog.require tokens in the order they 122 strings = self._GetRequireOrProvideTokenStrings(tokens) [all...] |
/external/webp/src/utils/ |
H A D | huffman_encode.c | 282 HuffmanTreeToken* tokens, 286 tokens->code = value; 287 tokens->extra_bits = 0; 288 ++tokens; 295 tokens->code = value; 296 tokens->extra_bits = 0; 297 ++tokens; 301 tokens->code = 16; 302 tokens->extra_bits = repetitions - 3; 303 ++tokens; 281 CodeRepeatedValues(int repetitions, HuffmanTreeToken* tokens, int value, int prev_value) argument 315 CodeRepeatedZeros(int repetitions, HuffmanTreeToken* tokens) argument 346 VP8LCreateCompressedHuffmanTree(const HuffmanTreeCode* const tree, HuffmanTreeToken* tokens, int max_tokens) argument [all...] |
/external/antlr/antlr-3.4/runtime/CSharp2/Sources/Antlr3.Runtime/Antlr.Runtime/ |
H A D | LegacyCommonTokenStream.cs | 43 * The most common stream of tokens is one where every token is buffered up 44 * and tokens are prefiltered for a certain channel (the parser will only 45 * see these tokens and cannot change the filter channel number during the 49 * <remarks>TODO: how to access the full token stream? How to track all tokens matched per rule?</remarks> 61 protected List<IToken> tokens; field in class:Antlr.Runtime.LegacyCommonTokenStream 66 /** <summary>Set<tokentype>; discard any tokens with this type</summary> */ 69 /** <summary>Skip tokens on any channel but this one; this is how we skip whitespace...</summary> */ 72 /** <summary>By default, track all incoming tokens</summary> */ 79 * The index into the tokens list of the current token (next token 80 * to consume). p==-1 indicates that the tokens lis [all...] |
/external/antlr/antlr-3.4/runtime/CSharp3/Sources/Antlr3.Runtime/ |
H A D | LegacyCommonTokenStream.cs | 41 * The most common stream of tokens is one where every token is buffered up 42 * and tokens are prefiltered for a certain channel (the parser will only 43 * see these tokens and cannot change the filter channel number during the 47 * <remarks>TODO: how to access the full token stream? How to track all tokens matched per rule?</remarks> 60 protected List<IToken> tokens; field in class:Antlr.Runtime.LegacyCommonTokenStream 65 /** <summary>Set<tokentype>; discard any tokens with this type</summary> */ 68 /** <summary>Skip tokens on any channel but this one; this is how we skip whitespace...</summary> */ 71 /** <summary>By default, track all incoming tokens</summary> */ 78 * The index into the tokens list of the current token (next token 79 * to consume). p==-1 indicates that the tokens lis [all...] |
/external/antlr/antlr-3.4/runtime/Java/src/main/java/org/antlr/runtime/ |
H A D | BufferedTokenStream.java | 35 /** Buffer all input tokens but do on-demand fetching of new tokens from 37 * proper lexing of future tokens. The ST template parser needs this, 43 * tokens to the parser. The stream can't ignore off-channel tokens. 47 * to confuse small moving window of tokens it uses for the full buffer. 57 protected List<Token> tokens = new ArrayList<Token>(100); field in class:BufferedTokenStream 62 /** The index into the tokens list of the current token (next token 63 * to consume). tokens[p] should be LT(1). p=-1 indicates need 108 public int size() { return tokens [all...] |
H A D | LegacyCommonTokenStream.java | 32 /** The most common stream of tokens is one where every token is buffered up 33 * and tokens are prefiltered for a certain channel (the parser will only 34 * see these tokens and cannot change the filter channel number during the 37 * TODO: how to access the full token stream? How to track all tokens matched per rule? 45 protected List tokens; field in class:LegacyCommonTokenStream 50 /** Set<tokentype>; discard any tokens with this type */ 53 /** Skip tokens on any channel but this one; this is how we skip whitespace... */ 56 /** By default, track all incoming tokens */ 64 /** The index into the tokens list of the current token (next token 65 * to consume). p==-1 indicates that the tokens lis [all...] |
/external/antlr/antlr-3.4/runtime/Perl5/examples/expr/ |
H A D | expr.pl | 22 my $tokens = ANTLR::Runtime::CommonTokenStream->new({ token_source => $lexer }); 23 my $parser = ExprParser->new({ input => $tokens });
|
/external/antlr/antlr-3.4/runtime/Perl5/examples/mexpr/ |
H A D | mexpr.pl | 17 my $tokens = ANTLR::Runtime::CommonTokenStream->new({ token_source => $lexer }); 18 my $parser = MExprParser->new($tokens);
|
/external/antlr/antlr-3.4/runtime/Perl5/examples/tweak/ |
H A D | tweak.pl | 13 my $tokens = ANTLR::Runtime::TokenRewriteStream({ token_source => $lexer }); 14 my $parser = TParser->new({ input => $tokens }); 16 print "$tokens\n";
|
/external/nist-sip/java/gov/nist/javax/sip/header/ |
H A D | Server.java | 53 /** Product tokens. 62 StringBuffer tokens = new StringBuffer(); 66 tokens.append((String) it.next()); 68 tokens.append('/'); 72 return tokens.toString();
|