/external/protobuf/src/google/protobuf/io/ |
H A D | tokenizer_unittest.cc | 182 EXPECT_TRUE(Tokenizer::ParseInteger(text, kuint64max, &result)); 197 Tokenizer::TokenType type; 207 { "hello", Tokenizer::TYPE_IDENTIFIER }, 210 { "123", Tokenizer::TYPE_INTEGER }, 211 { "0xab6", Tokenizer::TYPE_INTEGER }, 212 { "0XAB6", Tokenizer::TYPE_INTEGER }, 213 { "0X1234567", Tokenizer::TYPE_INTEGER }, 214 { "0x89abcdef", Tokenizer::TYPE_INTEGER }, 215 { "0x89ABCDEF", Tokenizer::TYPE_INTEGER }, 216 { "01234567", Tokenizer [all...] |
H A D | tokenizer.h | 51 class Tokenizer; 81 class LIBPROTOBUF_EXPORT Tokenizer { class in namespace:google::protobuf::io 83 // Construct a Tokenizer that reads and tokenizes text from the given 86 Tokenizer(ZeroCopyInputStream* input, ErrorCollector* error_collector); 87 ~Tokenizer(); 138 // comes from a TYPE_FLOAT token parsed by Tokenizer. If it doesn't, the 143 // comes from a TYPE_STRING token parsed by Tokenizer. If it doesn't, the 153 // parsed by a Tokenizer, the result is undefined (possibly an assert 180 GOOGLE_DISALLOW_EVIL_CONSTRUCTORS(Tokenizer); 290 inline const Tokenizer [all...] |
H A D | tokenizer.cc | 105 // For instance, Tokenizer::ConsumeZeroOrMore<Whitespace>() will eat 180 Tokenizer::Tokenizer(ZeroCopyInputStream* input, function in class:google::protobuf::io::Tokenizer 201 Tokenizer::~Tokenizer() { 212 void Tokenizer::NextChar() { 233 void Tokenizer::Refresh() { 263 inline void Tokenizer::StartToken() { 271 inline void Tokenizer::EndToken() { 286 inline bool Tokenizer [all...] |
/external/ganymed-ssh2/src/main/java/ch/ethz/ssh2/util/ |
H A D | Tokenizer.java | 8 * Tokenizer. Why? Because StringTokenizer is not available in J2ME. 13 public class Tokenizer class
|
/external/protobuf/src/google/protobuf/compiler/ |
H A D | parser.cc | 114 inline bool Parser::LookingAtType(io::Tokenizer::TokenType token_type) { 119 return LookingAtType(io::Tokenizer::TYPE_END); 150 if (LookingAtType(io::Tokenizer::TYPE_IDENTIFIER)) { 161 if (LookingAtType(io::Tokenizer::TYPE_INTEGER)) { 163 if (!io::Tokenizer::ParseInteger(input_->current().text, 179 if (LookingAtType(io::Tokenizer::TYPE_INTEGER)) { 180 if (!io::Tokenizer::ParseInteger(input_->current().text, max_value, 195 if (LookingAtType(io::Tokenizer::TYPE_FLOAT)) { 196 *output = io::Tokenizer::ParseFloat(input_->current().text); 199 } else if (LookingAtType(io::Tokenizer [all...] |
H A D | parser.h | 73 bool Parse(io::Tokenizer* input, FileDescriptorProto* file); 145 inline bool LookingAtType(io::Tokenizer::TokenType token_type); 291 io::Tokenizer* input_;
|
H A D | parser_unittest.cc | 104 input_.reset(new io::Tokenizer(raw_input_.get(), &error_collector_)); 118 EXPECT_EQ(io::Tokenizer::TYPE_END, input_->current().type); 135 EXPECT_EQ(io::Tokenizer::TYPE_END, input_->current().type); 158 EXPECT_EQ(io::Tokenizer::TYPE_END, input_->current().type); 172 scoped_ptr<io::Tokenizer> input_; 1220 EXPECT_EQ(io::Tokenizer::TYPE_END, input_->current().type);
|
H A D | importer.cc | 82 // - It implements the ErrorCollector interface (used by Tokenizer and Parser) 134 io::Tokenizer tokenizer(input.get(), &file_error_collector);
|
/external/doclava/src/com/google/doclava/apicheck/ |
H A D | ApiFile.java | 67 final Tokenizer tokenizer = new Tokenizer(filename, (new String(buf, 0, size)).toCharArray()); 88 private static void parsePackage(ApiInfo api, Tokenizer tokenizer) 113 private static void parseClass(ApiInfo api, PackageInfo pkg, Tokenizer tokenizer, String token) 222 private static void parseConstructor(Tokenizer tokenizer, ClassInfo cl, String token) 273 private static void parseMethod(Tokenizer tokenizer, ClassInfo cl, String token) 346 private static void parseField(Tokenizer tokenizer, ClassInfo cl, String token, boolean isEnum) 471 private static void parseParameterList(Tokenizer tokenizer, AbstractMethodInfo method, 501 private static String parseThrows(Tokenizer tokenizer, AbstractMethodInfo method) 533 public static void assertIdent(Tokenizer tokenize 539 static class Tokenizer { class in class:ApiFile 544 Tokenizer(String filename, char[] buf) { method in class:ApiFile.Tokenizer [all...] |
/external/protobuf/src/google/protobuf/ |
H A D | text_format.cc | 133 tokenizer_.set_comment_style(io::Tokenizer::SH_COMMENT_STYLE); 147 if (LookingAtType(io::Tokenizer::TYPE_END)) { 162 return suc && LookingAtType(io::Tokenizer::TYPE_END); 451 bool LookingAtType(io::Tokenizer::TokenType token_type) { 458 if (!LookingAtType(io::Tokenizer::TYPE_IDENTIFIER)) { 472 if (!LookingAtType(io::Tokenizer::TYPE_STRING)) { 478 while (LookingAtType(io::Tokenizer::TYPE_STRING)) { 479 io::Tokenizer::ParseStringAppend(tokenizer_.current().text, text); 490 if (!LookingAtType(io::Tokenizer::TYPE_INTEGER)) { 495 if (!io::Tokenizer [all...] |
/external/ganymed-ssh2/src/main/java/ch/ethz/ssh2/packets/ |
H A D | TypesReader.java | 11 import ch.ethz.ssh2.util.Tokenizer; 171 return Tokenizer.parseTokens(readString(), ',');
|
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/common/ |
H A D | tokenizer.py | 28 class Tokenizer(object): class in inherits:object
|
/external/chromium-trace/trace-viewer/third_party/closure_linter/closure_linter/ |
H A D | javascripttokenizer.py | 47 class JavaScriptTokenizer(tokenizer.Tokenizer): 347 tokenizer.Tokenizer.__init__(self, JavaScriptModes.TEXT_MODE, matchers,
|
/external/webkit/Source/WebCore/dom/ |
H A D | XMLDocumentParser.h | 33 #include "Tokenizer.h" 48 class XMLTokenizer : public Tokenizer, public CachedResourceClient { 56 // from Tokenizer
|
H A D | xml_expat_tokenizer.cpp | 896 Tokenizer *newXMLTokenizer(Document *d, FrameView *v) 913 Tokenizer::stopParsing();
|
/external/ganymed-ssh2/src/main/java/ch/ethz/ssh2/transport/ |
H A D | TransportManager.java | 36 import ch.ethz.ssh2.util.Tokenizer; 183 String[] quad = Tokenizer.parseTokens(host, '.');
|
/external/protobuf/java/src/main/java/com/google/protobuf/ |
H A D | TextFormat.java | 407 private static final class Tokenizer { class in class:TextFormat 446 private Tokenizer(final CharSequence text) { method in class:TextFormat.Tokenizer 852 final Tokenizer tokenizer = new Tokenizer(input); 863 private static void mergeField(final Tokenizer tokenizer,
|
/external/eclipse-basebuilder/basebuilder-3.6.2/org.eclipse.releng.basebuilder/plugins/ |
H A D | org.apache.lucene.analysis_1.9.1.v20100518-1140.jar | META-INF/MANIFEST.MF META-INF/ECLIPSEF.SF META-INF/ECLIPSEF.RSA META ... |
H A D | org.apache.lucene_1.9.1.v20100518-1140.jar | META-INF/MANIFEST.MF META-INF/ECLIPSEF.SF META-INF/ECLIPSEF.RSA META ... |
H A D | org.eclipse.osgi_3.6.1.R36x_v20100806.jar | META-INF/MANIFEST.MF META-INF/ECLIPSEF.SF META-INF/ECLIPSEF.RSA META ... |
H A D | org.eclipse.osgi_3.6.2.R36x_v20101103.jar | META-INF/MANIFEST.MF META-INF/ECLIPSEF.SF META-INF/ECLIPSEF.RSA META ... |
/external/antlr/antlr-3.4/runtime/Delphi/Sources/Antlr3.Runtime/ |
H A D | Antlr.Runtime.Tree.pas | 4012 Tokenizer: ITreePatternLexer; 4015 Tokenizer := TTreePatternLexer.Create(Pattern); 4016 Parser := TTreePatternParser.Create(Tokenizer, Self, FAdaptor); 4034 Tokenizer: ITreePatternLexer; 4043 Tokenizer := TTreePatternLexer.Create(Pattern); 4044 Parser := TTreePatternParser.Create(Tokenizer, Self, TTreePatternTreeAdaptor.Create); 4101 Tokenizer: ITreePatternLexer; 4105 Tokenizer := TTreePatternLexer.Create(Pattern); 4106 Parser := TTreePatternParser.Create(Tokenizer, Self, TTreePatternTreeAdaptor.Create); 4114 Tokenizer [all...] |
/external/v8/tools/ |
H A D | test.py | 890 class Tokenizer(object): class in inherits:object 1047 tokens = Tokenizer(expr).Tokenize()
|
/external/eclipse-basebuilder/basebuilder-3.6.2/org.eclipse.releng.basebuilder/plugins/org.apache.ant_1.7.1.v20090120-1145/lib/ |
H A D | ant.jar | META-INF/ META-INF/MANIFEST.MF org/ org/apache/ org/apache/tools/ org/apache/tools/ant/ ... |