Searched defs:UChar32 (Results 1 - 9 of 9) sorted by relevance
/external/clang/test/ARCMT/ |
H A D | Common.h | 17 typedef int32_t UChar32; typedef
|
/external/chromium/base/third_party/icu/ |
H A D | icu_utf.h | 24 typedef uint32 UChar32; typedef in namespace:base_icu 32 * (take or) return single code points (UChar32). 45 * @see UChar32 172 UChar32 utf8_nextCharSafeBody(const uint8 *s, int32 *pi, int32 length, UChar32 c, UBool strict); 188 * @param c output UChar32 variable, set to <0 in case of an error 297 (((base_icu::UChar32)(lead)<<10UL)+(base_icu::UChar32)(trail)-CBU16_SURROGATE_OFFSET) 350 * @param c output UChar32 variable
|
/external/icu4c/common/unicode/ |
H A D | umachine.h | 299 /* UChar and UChar32 definitions -------------------------------------------- */ 329 * Define UChar32 as a type for single Unicode code points. 330 * UChar32 is a signed 32-bit integer (same as int32_t). 337 * Before ICU 2.4 (Jitterbug 2146), UChar32 was defined 340 * That is, the definition of UChar32 was platform-dependent. 345 typedef int32_t UChar32; typedef
|
/external/webkit/Source/JavaScriptCore/icu/unicode/ |
H A D | umachine.h | 306 /* UChar and UChar32 definitions -------------------------------------------- */ 330 * Define UChar32 as a type for single Unicode code points. 331 * UChar32 is a signed 32-bit integer (same as int32_t). 338 * Before ICU 2.4 (Jitterbug 2146), UChar32 was defined 341 * That is, the definition of UChar32 was platform-dependent. 346 typedef int32_t UChar32; typedef
|
/external/webkit/Source/JavaScriptGlue/icu/unicode/ |
H A D | umachine.h | 306 /* UChar and UChar32 definitions -------------------------------------------- */ 330 * Define UChar32 as a type for single Unicode code points. 331 * UChar32 is a signed 32-bit integer (same as int32_t). 338 * Before ICU 2.4 (Jitterbug 2146), UChar32 was defined 341 * That is, the definition of UChar32 was platform-dependent. 346 typedef int32_t UChar32; typedef
|
/external/webkit/Source/WebCore/icu/unicode/ |
H A D | umachine.h | 306 /* UChar and UChar32 definitions -------------------------------------------- */ 330 * Define UChar32 as a type for single Unicode code points. 331 * UChar32 is a signed 32-bit integer (same as int32_t). 338 * Before ICU 2.4 (Jitterbug 2146), UChar32 was defined 341 * That is, the definition of UChar32 was platform-dependent. 346 typedef int32_t UChar32; typedef
|
/external/webkit/Source/WebKit/mac/icu/unicode/ |
H A D | umachine.h | 306 /* UChar and UChar32 definitions -------------------------------------------- */ 330 * Define UChar32 as a type for single Unicode code points. 331 * UChar32 is a signed 32-bit integer (same as int32_t). 338 * Before ICU 2.4 (Jitterbug 2146), UChar32 was defined 341 * That is, the definition of UChar32 was platform-dependent. 346 typedef int32_t UChar32; typedef
|
/external/webkit/Source/JavaScriptCore/wtf/unicode/glib/ |
H A D | UnicodeGLib.h | 38 typedef int32_t UChar32; typedef 127 UChar32 foldCase(UChar32); 133 inline UChar32 toLower(UChar32 c) 138 inline UChar32 toUpper(UChar32 c) 145 inline UChar32 toTitleCase(UChar32 c) 150 inline bool isArabicChar(UChar32 [all...] |
/external/webkit/Source/JavaScriptCore/wtf/unicode/qt4/ |
H A D | UnicodeQt4.h | 71 typedef uint32_t UChar32; typedef 157 inline UChar32 toLower(UChar32 ch) 215 inline UChar32 toUpper(UChar32 c) 273 inline int toTitleCase(UChar32 c) 278 inline UChar32 foldCase(UChar32 c) 296 inline bool isArabicChar(UChar32 c) 301 inline bool isPrintableChar(UChar32 [all...] |
Completed in 194 milliseconds