/external/valgrind/main/memcheck/tests/ |
H A D | sh-mem.c | 15 typedef unsigned long long U8; typedef 24 U8 a [SZB_OF_a / 8]; // Type is U8 to ensure it's 8-aligned 25 U8 b [SZB_OF_a / 8]; // same size as a[] 38 U8 build(int size, U1 byte) 41 U8 mask = 0; 42 U8 shres; 43 U8 res = 0xffffffffffffffffULL, res2; 49 mask |= (U8)byte; 109 assert(8 == sizeof(U8)); [all...] |
H A D | sh-mem-random.c | 16 typedef unsigned long long U8; typedef 37 U8 build(int size, U1 byte) 40 U8 mask = 0; 41 U8 shres; 42 U8 res = 0xffffffffffffffffULL, res2; 48 mask |= (U8)byte; 75 U8 sum = 0; 85 sum += (U8)arr_i; 153 case 3: { // U8 156 *(U8*)(ar [all...] |
H A D | sh-mem.stderr.exp | 16 -- NNN: 8 U8 U8 ------------------------ 25 -- NNN: 8 F8 U8 ------------------------
|
/external/libhevc/decoder/arm/ |
H A D | ihevcd_fmt_conv_420sp_to_rgba8888.s | 181 VSUBL.U8 Q2,D2,D1 @//(U-128) 182 VSUBL.U8 Q3,D3,D1 @//(V-128) 218 VADDW.U8 Q7,Q4,D30 @//Q7 - HAS Y + B 219 VADDW.U8 Q8,Q5,D30 @//Q8 - HAS Y + R 220 VADDW.U8 Q9,Q6,D30 @//Q9 - HAS Y + G 222 VADDW.U8 Q10,Q4,D31 @//Q10 - HAS Y + B 223 VADDW.U8 Q11,Q5,D31 @//Q11 - HAS Y + R 224 VADDW.U8 Q12,Q6,D31 @//Q12 - HAS Y + G 259 VADDW.U8 Q7,Q4,D28 @//Q7 - HAS Y + B 260 VADDW.U8 Q [all...] |
/external/libhevc/common/arm/ |
H A D | ihevc_sao_edge_offset_class0_chroma.s | 163 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 167 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 170 VSUB.U8 Q10,Q9,Q8 @sign_left = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 177 VCGT.U8 Q13,Q15,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 180 VCLT.U8 Q12,Q15,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 191 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 194 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 197 VSUB.U8 Q11,Q9,Q8 @sign_right = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 200 VADD.U8 Q7,Q1,Q10 @edge_idx = vaddq_s8(const_2, sign_left) 202 VADD.U8 Q [all...] |
H A D | ihevc_sao_edge_offset_class1_chroma.s | 137 VCGT.U8 Q6,Q5,Q14 @vcgtq_u8(pu1_cur_row, pu1_top_row) 140 VCLT.U8 Q7,Q5,Q14 @vcltq_u8(pu1_cur_row, pu1_top_row) 142 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 152 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 157 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 160 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 161 VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 164 VMOVL.U8 Q14,D19 @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 167 VCGT.U8 Q11,Q9,Q15 @II vcgtq_u8(pu1_cur_row, pu1_top_row) 171 VCLT.U8 Q1 [all...] |
H A D | ihevc_sao_edge_offset_class1.s | 132 VCGT.U8 Q6,Q5,Q4 @vcgtq_u8(pu1_cur_row, pu1_top_row) 135 VCLT.U8 Q7,Q5,Q4 @vcltq_u8(pu1_cur_row, pu1_top_row) 137 VSUB.U8 Q8,Q7,Q6 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 147 VCGT.U8 Q6,Q5,Q9 @vcgtq_u8(pu1_cur_row, pu1_top_row) 152 VCLT.U8 Q7,Q5,Q9 @vcltq_u8(pu1_cur_row, pu1_top_row) 155 VSUB.U8 Q10,Q7,Q6 @sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 156 VMOVL.U8 Q13,D18 @II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 159 VMOVL.U8 Q14,D19 @II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 162 VCGT.U8 Q11,Q9,Q15 @II vcgtq_u8(pu1_cur_row, pu1_top_row) 166 VCLT.U8 Q1 [all...] |
H A D | ihevc_sao_band_offset_luma.s | 143 VCLE.U8 D12,D4,D29 @vcle_u8(band_table.val[3], vdup_n_u8(16)) 146 VORR.U8 D4,D4,D12 @band_table.val[3] = vorr_u8(band_table.val[3], au1_cmp) 151 VCLE.U8 D11,D3,D29 @vcle_u8(band_table.val[2], vdup_n_u8(16)) 154 VORR.U8 D3,D3,D11 @band_table.val[2] = vorr_u8(band_table.val[2], au1_cmp) 156 VAND.U8 D4,D4,D12 @band_table.val[3] = vand_u8(band_table.val[3], au1_cmp) 161 VCLE.U8 D10,D2,D29 @vcle_u8(band_table.val[1], vdup_n_u8(16)) 164 VORR.U8 D2,D2,D10 @band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 166 VAND.U8 D3,D3,D11 @band_table.val[2] = vand_u8(band_table.val[2], au1_cmp) 173 VCLE.U8 D9,D1,D29 @vcle_u8(band_table.val[0], vdup_n_u8(16)) 174 VORR.U8 D [all...] |
H A D | ihevc_sao_band_offset_chroma.s | 157 VCLE.U8 D13,D4,D30 @vcle_u8(band_table.val[3], vdup_n_u8(16)) 160 VORR.U8 D4,D4,D13 @band_table.val[3] = vorr_u8(band_table.val[3], au1_cmp) 166 VCLE.U8 D14,D3,D30 @vcle_u8(band_table.val[2], vdup_n_u8(16)) 168 VORR.U8 D3,D3,D14 @band_table.val[2] = vorr_u8(band_table.val[2], au1_cmp) 170 VAND.U8 D4,D4,D13 @band_table.val[3] = vand_u8(band_table.val[3], au1_cmp) 176 VCLE.U8 D15,D2,D30 @vcle_u8(band_table.val[1], vdup_n_u8(16)) 178 VORR.U8 D2,D2,D15 @band_table.val[1] = vorr_u8(band_table.val[1], au1_cmp) 180 VAND.U8 D3,D3,D14 @band_table.val[2] = vand_u8(band_table.val[2], au1_cmp) 186 VCLE.U8 D16,D1,D30 @vcle_u8(band_table.val[0], vdup_n_u8(16)) 187 VORR.U8 D [all...] |
H A D | ihevc_sao_edge_offset_class0.s | 157 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 162 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 174 VCGT.U8 Q15,Q13,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 180 VCLT.U8 Q0,Q13,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 185 VCGT.U8 Q8,Q6,Q7 @vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 188 VCLT.U8 Q9,Q6,Q7 @vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 199 VMOVL.U8 Q9,D12 @pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 203 VCGT.U8 Q15,Q13,Q14 @II vcgtq_u8(pu1_cur_row, pu1_cur_row_tmp) 205 VCLT.U8 Q0,Q13,Q14 @II vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 211 VMOVL.U8 Q [all...] |
H A D | ihevc_sao_edge_offset_class2.s | 247 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 251 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 255 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 292 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 295 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 296 VSUB.U8 Q5,Q9,Q5 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 308 VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 313 VMOVL.U8 Q11,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 359 VCGT.U8 Q12,Q6,Q11 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 363 VCLT.U8 Q1 [all...] |
H A D | ihevc_sao_edge_offset_class2_chroma.s | 339 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 343 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 355 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 401 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 403 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 404 VSUB.U8 Q11,Q11,Q10 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 415 VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 418 VMOVL.U8 Q9,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 475 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 483 VCLT.U8 Q1 [all...] |
H A D | ihevc_sao_edge_offset_class3.s | 262 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 272 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 275 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 307 VCGT.U8 Q5,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 308 VCLT.U8 Q9,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 309 VSUB.U8 Q5,Q9,Q5 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 319 VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 324 VMOVL.U8 Q11,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 384 VCGT.U8 Q12,Q6,Q9 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 388 VCLT.U8 Q1 [all...] |
H A D | ihevc_sao_edge_offset_class3_chroma.s | 329 VCGT.U8 Q7,Q6,Q5 @vcgtq_u8(pu1_cur_row, pu1_top_row) 333 VCLT.U8 Q8,Q6,Q5 @vcltq_u8(pu1_cur_row, pu1_top_row) 337 VSUB.U8 Q7,Q8,Q7 @sign_up = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 393 VCGT.U8 Q10,Q6,Q9 @I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 395 VCLT.U8 Q11,Q6,Q9 @I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 396 VSUB.U8 Q11,Q11,Q10 @I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 406 VMOVL.U8 Q10,D12 @I pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 414 VMOVL.U8 Q9,D13 @I pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 476 VCGT.U8 Q11,Q6,Q14 @II vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 486 VCLT.U8 Q1 [all...] |
/external/chromium_org/third_party/skia/include/xml/ |
H A D | SkBML_WXMLParser.h | 42 // important that these are U8, so we get automatic wrap-around 43 U8 fNextElem, fNextAttrName, fNextAttrValue;
|
/external/skia/include/xml/ |
H A D | SkBML_WXMLParser.h | 42 // important that these are U8, so we get automatic wrap-around 43 U8 fNextElem, fNextAttrName, fNextAttrValue;
|
/external/compiler-rt/lib/asan/tests/ |
H A D | asan_asm_test.cc | 56 DECLARE_ASM_WRITE(U8, "8", "movq", "r"); 57 DECLARE_ASM_READ(U8, "8", "movq", "=r"); 174 TestAsmWrite<U8>("WRITE of size 8"); 182 TestAsmRead<U8>("READ of size 8");
|
H A D | asan_test_utils.h | 61 typedef uint64_t U8; typedef
|
/external/compiler-rt/lib/msan/tests/ |
H A D | msan_test.cc | 77 typedef unsigned long long U8; // NOLINT typedef 133 static U8 poisoned_array[100]; 418 EXPECT_UMR(break_optimization((void*)(U8)a[GetPoisonedZero()])); 1468 TestOverlapMemmove<U8, 4>(); 1469 TestOverlapMemmove<U8, 1000>(); 2519 U8 b; 2595 U8 a8, b8; 3422 U8 y8 = 0; 3491 U8 y = 0; 3523 U8 [all...] |
/external/clang/test/Misc/ |
H A D | diag-template-diffing.cpp | 234 template<typename T> using U8 = S8<int, char, T>; 236 int k8 = f8(U8<char>());
|
/external/valgrind/main/VEX/priv/ |
H A D | ir_opt.c | 1091 case Ico_U8: return toBool( c1->Ico.U8 == c2->Ico.U8 ); 1211 case Ico_U8: return toBool( e->Iex.Const.con->Ico.U8 == 0); 1224 case Ico_U8: return toBool( e->Iex.Const.con->Ico.U8 == 0xFF); 1396 /* signed */ Int s32 = e->Iex.Unop.arg->Iex.Const.con->Ico.U8; 1411 0xFFULL & e->Iex.Unop.arg->Iex.Const.con->Ico.U8)); 1419 0xFF & e->Iex.Unop.arg->Iex.Const.con->Ico.U8)); 1422 /* signed */ Short s16 = e->Iex.Unop.arg->Iex.Const.con->Ico.U8; 1430 0xFF & e->Iex.Unop.arg->Iex.Const.con->Ico.U8)); 1473 ~ (e->Iex.Unop.arg->Iex.Const.con->Ico.U8)))); [all...] |
H A D | host_arm_isel.c | 980 case Ico_U8: u = 0xFF & (e->Iex.Const.con->Ico.U8); break; 1034 case Ico_U8: u = 0xFF & (e->Iex.Const.con->Ico.U8); break; 1384 index = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; 1409 index = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; 1861 case Ico_U8: u = 0xFF & (e->Iex.Const.con->Ico.U8); break; 2687 imm = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; 2711 imm = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; 2735 imm = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; 2782 Int nshift = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; 3116 imm6 = e->Iex.Binop.arg2->Iex.Const.con->Ico.U8; [all...] |
/external/valgrind/main/memcheck/tests/common/ |
H A D | sh-mem-vec128.tmpl.c | 25 typedef unsigned long long U8; typedef
|
/external/llvm/unittests/ADT/ |
H A D | StringRefTest.cpp | 414 uint8_t U8; local 420 bool U8Success = StringRef(Unsigned[i].Str).getAsInteger(0, U8); 423 EXPECT_EQ(U8, Unsigned[i].Expected);
|
/external/chromium_org/media/mojo/services/ |
H A D | media_type_converters.cc | 90 ASSERT_ENUM_EQ(SampleFormat, kSampleFormat, SAMPLE_FORMAT_, U8);
|