Searched refs:Tokenize (Results 1 - 10 of 10) sorted by relevance

/external/google-breakpad/src/processor/
H A Dtokenize.h30 // Implements a Tokenize function for splitting up strings.
50 // Tokenize, but may be treated as a failure if the caller expects an
53 bool Tokenize(char *line,
57 // For convenience, since you need a char* to pass to Tokenize.
H A Dtokenize.cc45 bool Tokenize(char *line, function in namespace:google_breakpad
H A Dbasic_source_line_resolver.cc471 if (!Tokenize(file_line, kWhitespace, 2, &tokens)) {
499 if (!Tokenize(function_line, kWhitespace, 4, &tokens)) {
531 if (!Tokenize(line_line, kWhitespace, 4, &tokens)) {
580 if (!Tokenize(public_line, kWhitespace, 3, &tokens)) {
H A Dwindows_frame_info.h126 if (!Tokenize(&buffer[0], " \r\n", 11, &tokens))
/external/vixl/src/vixl/a64/
H A Ddebugger-a64.cc61 static Token* Tokenize(const char* arg);
94 static Token* Tokenize(const char* arg);
116 static Token* Tokenize(const char* arg);
140 static Token* Tokenize(const char* arg);
158 static Token* Tokenize(const char* arg);
175 static Token* Tokenize(const char* arg);
207 static Token* Tokenize(const char* arg);
805 Token* Token::Tokenize(const char* arg) { function in class:vixl::Token
810 // The order is important. For example Identifier::Tokenize would consider
813 Token* token = RegisterToken::Tokenize(ar
866 Token* RegisterToken::Tokenize(const char* arg) { function in class:vixl::RegisterToken
894 Token* FPRegisterToken::Tokenize(const char* arg) { function in class:vixl::FPRegisterToken
943 Token* IdentifierToken::Tokenize(const char* arg) { function in class:vixl::IdentifierToken
972 Token* AddressToken::Tokenize(const char* arg) { function in class:vixl::AddressToken
992 Token* IntegerToken::Tokenize(const char* arg) { function in class:vixl::IntegerToken
1002 Token* FormatToken::Tokenize(const char* arg) { function in class:vixl::FormatToken
[all...]
/external/chromium-trace/catapult/common/py_utils/py_utils/refactor/
H A Doffset_token.py55 def Tokenize(f): function
H A Dsnippet.py206 tokens = offset_token.Tokenize(f)
/external/google-breakpad/src/testing/gtest/scripts/
H A Dpump.py382 def Tokenize(s): function
579 tokens = list(Tokenize(pump_src_text))
/external/protobuf/gtest/scripts/
H A Dpump.py376 def Tokenize(s): function
571 for token in Tokenize(s):
/external/vulkan-validation-layers/tests/gtest-1.7.0/scripts/
H A Dpump.py382 def Tokenize(s): function
579 tokens = list(Tokenize(pump_src_text))

Completed in 200 milliseconds