/external/boringssl/ios-aarch64/crypto/fipsmodule/ |
H A D | ghashv8-armx64.S | 14 ushr v18.2d,v19.2d,#63 16 ext v16.16b,v18.16b,v19.16b,#8 //t0=0xc2....01 17 ushr v18.2d,v3.2d,#63 19 and v18.16b,v18.16b,v16.16b 21 ext v18.16b,v18.16b,v18.16b,#8 23 orr v3.16b,v3.16b,v18.16b //H<<<=1 35 eor v18 [all...] |
H A D | aesv8-armx64.S | 302 ld1 {v18.4s,v19.4s},[x7],#32 358 aese v0.16b,v18.16b 397 aese v0.16b,v18.16b 417 ld1 {v18.16b},[x0],#16 422 orr v19.16b,v18.16b,v18.16b 425 orr v1.16b,v18.16b,v18.16b 426 ld1 {v18.16b},[x0],#16 429 orr v19.16b,v18 [all...] |
/external/boringssl/linux-aarch64/crypto/fipsmodule/ |
H A D | ghashv8-armx64.S | 15 ushr v18.2d,v19.2d,#63 17 ext v16.16b,v18.16b,v19.16b,#8 //t0=0xc2....01 18 ushr v18.2d,v3.2d,#63 20 and v18.16b,v18.16b,v16.16b 22 ext v18.16b,v18.16b,v18.16b,#8 24 orr v3.16b,v3.16b,v18.16b //H<<<=1 36 eor v18 [all...] |
H A D | aesv8-armx64.S | 303 ld1 {v18.4s,v19.4s},[x7],#32 359 aese v0.16b,v18.16b 398 aese v0.16b,v18.16b 418 ld1 {v18.16b},[x0],#16 423 orr v19.16b,v18.16b,v18.16b 426 orr v1.16b,v18.16b,v18.16b 427 ld1 {v18.16b},[x0],#16 430 orr v19.16b,v18 [all...] |
/external/libavc/common/armv8/ |
H A D | ih264_inter_pred_luma_horz_qpel_vert_qpel_av8.s | 156 ld1 {v18.2s, v19.2s}, [x6], x2 // horz row0, col 0 162 ext v23.8b, v18.8b , v19.8b , #5 163 ext v20.8b, v18.8b , v19.8b , #2 164 ext v21.8b, v18.8b , v19.8b , #3 165 ext v22.8b, v18.8b , v19.8b , #4 166 ext v19.8b, v18.8b , v19.8b , #1 168 uaddl v28.8h, v18.8b, v23.8b 173 ld1 {v18.2s, v19.2s}, [x11], x2 // horz row 0, col 1 180 ext v23.8b, v18.8b , v19.8b , #5 181 ext v20.8b, v18 [all...] |
H A D | ih264_inter_pred_luma_horz_hpel_vert_hpel_av8.s | 115 uaddl v18.8h, v0.8b, v10.8b 117 mla v18.8h, v20.8h , v28.8h 123 mls v18.8h, v22.8h , v30.8h 130 ext v24.16b, v18.16b , v20.16b , #4 131 ext v26.16b, v18.16b , v20.16b , #6 133 ext v23.16b, v18.16b , v20.16b , #10 135 ext v24.16b, v18.16b , v20.16b , #2 136 ext v26.16b, v18.16b , v20.16b , #8 139 saddl v26.4s, v18.4h, v23.4h 143 saddl2 v23.4s, v18 [all...] |
H A D | ih264_inter_pred_luma_horz_qpel_vert_hpel_av8.s | 177 uaddl v18.8h, v0.8b, v10.8b 179 mla v18.8h, v20.8h , v28.8h 185 mls v18.8h, v22.8h , v30.8h 191 st1 {v18.4s }, [x9], #16 193 ext v24.16b, v18.16b , v20.16b , #4 194 ext v26.16b, v18.16b , v20.16b , #6 196 ext v22.16b, v18.16b , v20.16b , #10 198 ext v24.16b, v18.16b , v20.16b , #2 199 ext v26.16b, v18.16b , v20.16b , #8 202 saddl v26.4s, v18 [all...] |
H A D | ih264_inter_pred_luma_horz_hpel_vert_qpel_av8.s | 242 uaddl v18.8h, v2.8b, v3.8b 244 mla v16.8h, v18.8h , v22.8h 247 uaddl v18.8h, v1.8b, v4.8b 249 mls v16.8h, v18.8h , v24.8h 258 saddl v18.4s, v6.4h, v16.4h 266 smlal v18.4s, v30.4h, v22.4h 267 smlsl v18.4s, v28.4h, v24.4h 273 sqrshrun v18.4h, v18.4s, #10 281 uqxtn v18 [all...] |
H A D | ih264_inter_pred_filters_luma_vert_av8.s | 144 uaddl v18.8h, v5.8b, v7.8b // temp3 = src[2_8] + src[3_8] 145 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 151 uaddl v18.8h, v4.8b, v10.8b 160 mls v16.8h, v18.8h , v24.8h 162 uaddl v18.8h, v4.8b, v2.8b 166 mla v18.8h, v12.8h , v22.8h 174 mls v18.8h, v20.8h , v24.8h 184 sqrshrun v30.8b, v18.8h, #5 185 uaddl v18.8h, v7.8b, v5.8b 187 mla v18 [all...] |
H A D | ih264_inter_pred_luma_vert_qpel_av8.s | 151 uaddl v18.8h, v5.8b, v7.8b // temp3 = src[2_8] + src[3_8] 152 mla v20.8h, v18.8h , v22.8h // temp4 += temp3 * 20 158 uaddl v18.8h, v4.8b, v10.8b 167 mls v16.8h, v18.8h , v24.8h 172 uaddl v18.8h, v4.8b, v2.8b 175 mla v18.8h, v12.8h , v22.8h 183 mls v18.8h, v20.8h , v24.8h 195 sqrshrun v30.8b, v18.8h, #5 196 uaddl v18.8h, v7.8b, v5.8b 198 mla v18 [all...] |
H A D | ih264_deblk_chroma_av8.s | 121 cmhs v18.16b, v22.16b, v20.16b 127 orr v18.16b, v18.16b , v24.16b //Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) 131 orr v18.16b, v18.16b , v26.16b //Q9 = ( ABS(p0 - q0) >= Alpha ) | ( ABS(q1 - q0) >= Beta ) | ( ABS(p1 - p0) >= Beta ) 132 bit v10.16b, v4.16b , v18.16b // 133 bit v8.16b, v0.16b , v18.16b // 236 uaddl v18.8h, v0.8b, v4.8b 239 umlal v18.8h, v6.8b, v31.8b 246 rshrn v18 [all...] |
/external/libmpeg2/common/armv8/ |
H A D | ideint_spatial_filter_av8.s | 69 movi v18.8h, #0 112 uabal v18.8h, v1.8b, v5.8b 126 addp v18.8h, v18.8h, v18.8h 130 uaddlp v18.2s, v18.4h 137 mul v18.2s, v18.2s, v31.2s 148 smov x6, v18 [all...] |
/external/libhevc/common/arm64/ |
H A D | ihevc_sao_edge_offset_class3_chroma.s | 363 movi v18.16b, #0 394 movi v18.16b, #0 //I 404 mov v18.h[7], w5 //I vsetq_lane_u8 408 EXT v18.16b, v18.16b , v16.16b,#14 //I pu1_next_row_tmp = vextq_u8(pu1_next_row_tmp, pu1_next_row, 14) 441 cmhi v20.16b, v5.16b , v18.16b //I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 443 cmhi v22.16b, v18.16b , v5.16b //I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 446 ADD v18.16b, v0.16b , v17.16b //I edge_idx = vaddq_s8(const_2, sign_up) 447 ADD v18.16b, v18 [all...] |
H A D | ihevc_weighted_pred_bi.s | 251 smull v18.4s, v2.4h, v7.h[0] //vmull_n_s16(pi2_src1_val2, (int16_t) wgt0) iv iteration 263 add v18.4s, v18.4s , v20.4s //vaddq_s32(i4_tmp2_t1, i4_tmp2_t2) iv iteration 266 add v18.4s, v18.4s , v30.4s //vaddq_s32(i4_tmp2_t1, tmp_lvl_shift_t) iv iteration 271 sshl v18.4s,v18.4s,v28.4s 279 sqxtun v18.4h, v18.4s //vqmovun_s32(sto_res_tmp1) iv iteration 280 //mov v19, v18 //vcombine_u1 [all...] |
H A D | ihevc_sao_edge_offset_class2.s | 290 movi v18.16b, #0 298 mov v18.b[0], w5 //I pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], pu1_next_row_tmp, 0) 300 EXT v18.16b, v16.16b , v18.16b,#1 //I pu1_next_row_tmp = vextq_u8(pu1_next_row, pu1_next_row_tmp, 1) 319 cmhi v3.16b, v5.16b , v18.16b //I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 322 cmhi v18.16b, v18.16b , v5.16b //I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 323 SUB v3.16b, v18.16b , v3.16b //I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 326 TBL v18.16b, {v6.16b},v24.16b //I vtbl1_s8(edge_idx_tbl, vget_low_s8(edge_idx)) 329 AND v18 [all...] |
H A D | ihevc_sao_edge_offset_class3.s | 296 movi v18.16b, #0 311 mov v18.b[15], w8 //I vsetq_lane_u8 314 EXT v18.16b, v18.16b , v16.16b,#15 //I pu1_next_row_tmp = vextq_u8(pu1_next_row_tmp, pu1_next_row, 15) 332 cmhi v3.16b, v5.16b , v18.16b //I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 333 cmhi v18.16b, v18.16b , v5.16b //I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 334 SUB v3.16b, v18.16b , v3.16b //I sign_down = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 336 ADD v18.16b, v0.16b , v17.16b //I edge_idx = vaddq_s8(const_2, sign_up) 337 ADD v18 [all...] |
H A D | ihevc_sao_edge_offset_class0.s | 162 cmhi v18.16b, v21.16b , v17.16b //vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 165 SUB v20.16b, v18.16b , v16.16b //sign_left = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 188 cmhi v18.16b, v21.16b , v17.16b //vcltq_u8(pu1_cur_row, pu1_cur_row_tmp) 192 SUB v22.16b, v18.16b , v16.16b //sign_right = vreinterpretq_s8_u8(vsubq_u8(cmp_lt, cmp_gt)) 199 Uxtl v18.8h, v17.8b //pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 216 SADDW v18.8h, v18.8h , v16.8b 218 SMAX v18.8h, v18.8h , v4.8h //pi2_tmp_cur_row.val[0] = vmaxq_s16(pi2_tmp_cur_row.val[0], const_min_clip) 221 UMIN v18 [all...] |
H A D | ihevc_sao_edge_offset_class2_chroma.s | 411 movi v18.16b, #0 415 mov v18.h[0], w5 //I pu1_next_row_tmp = vsetq_lane_u8(pu1_src_cpy[src_strd + 16], pu1_next_row_tmp, 0) 419 EXT v18.16b, v16.16b , v18.16b,#2 //I pu1_next_row_tmp = vextq_u8(pu1_next_row, pu1_next_row_tmp, 2) 455 cmhi v20.16b, v5.16b , v18.16b //I vcgtq_u8(pu1_cur_row, pu1_next_row_tmp) 457 cmhi v22.16b, v18.16b , v5.16b //I vcltq_u8(pu1_cur_row, pu1_next_row_tmp) 460 ADD v18.16b, v0.16b , v17.16b //I edge_idx = vaddq_s8(const_2, sign_up) 461 ADD v18.16b, v18.16b , v22.16b //I edge_idx = vaddq_s8(edge_idx, sign_down) 463 TBL v18 [all...] |
H A D | ihevc_intra_pred_chroma_horz.s | 120 ld1 { v18.8h},[x12] //load 16 values. d1[7] will have the 1st value. 149 dup v2.8h, v18.h[7] 153 dup v4.8h, v18.h[6] 157 dup v6.8h, v18.h[5] 161 dup v1.8h, v18.h[4] 165 dup v2.8h, v18.h[3] 169 dup v4.8h, v18.h[2] 173 dup v6.8h, v18.h[1] 178 dup v1.8h, v18.h[0] 206 dup v18 [all...] |
H A D | ihevc_sao_edge_offset_class1.s | 145 LD1 {v18.16b},[x10] //pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 148 cmhi v5.16b, v3.16b , v18.16b //vcgtq_u8(pu1_cur_row, pu1_top_row) 151 cmhi v17.16b, v18.16b , v3.16b //vcltq_u8(pu1_cur_row, pu1_top_row) 155 Uxtl v26.8h, v18.8b //II pi2_tmp_cur_row.val[0] = vreinterpretq_s16_u16(vmovl_u8(vget_low_u8(pu1_cur_row))) 158 Uxtl2 v28.8h, v18.16b //II pi2_tmp_cur_row.val[1] = vreinterpretq_s16_u16(vmovl_u8(vget_high_u8(pu1_cur_row))) 161 cmhi v22.16b, v18.16b , v30.16b //II vcgtq_u8(pu1_cur_row, pu1_top_row) 165 cmhi v24.16b, v30.16b , v18.16b //II vcltq_u8(pu1_cur_row, pu1_top_row) 222 LD1 {v18.16b},[x10] //pu1_next_row = vld1q_u8(pu1_src_cpy + src_strd) 223 cmhi v5.16b, v3.16b , v18.16b //vcgtq_u8(pu1_cur_row, pu1_top_row) 224 cmhi v17.16b, v18 [all...] |
H A D | ihevc_itrans_recon_8x8.s | 189 smull v18.4s, v3.4h, v1.h[2] //// y2 * sin2 (q3 is freed by this time)(part of d1) 239 smlsl v18.4s, v11.4h, v0.h[2] //// d1 = y2 * sin2 - y6 * cos2(part of a0 and a1) 252 sub v22.4s, v20.4s , v18.4s //// a2 = c1 - d1(part of x2,x5) 253 add v18.4s, v20.4s , v18.4s //// a1 = c1 + d1(part of x1,x6) 261 add v28.4s, v18.4s , v26.4s //// a1 + b1(part of x1) 262 sub v18.4s, v18.4s , v26.4s //// a1 - b1(part of x6) 272 sqrshrn v11.4h, v18.4s,#shift_stage1_idct //// x6 = (a1 - b1 + rnd) >> 7(shift_stage1_idct) 314 smull v18 [all...] |
H A D | ihevc_intra_pred_chroma_dc.s | 118 mov v18.s[0], w9 119 mov v18.s[1], w9 145 uadalp v18.1d, v3.2s 154 uadalp v18.1d, v3.2s 168 uadalp v18.1d, v3.2s 177 uadalp v18.1d, v3.2s 183 smov x1, v18.s[0] 266 uadalp v18.1d, v2.2s 275 uadalp v18.1d, v2.2s 278 smov x11, v18 [all...] |
H A D | ihevc_intra_pred_luma_mode_3_to_9.s | 222 umull v18.8h, v23.8b, v7.8b //mul (row 3) 223 umlal v18.8h, v25.8b, v6.8b //mul (row 3) 230 rshrn v18.8b, v18.8h,#5 //round shft (row 3) 240 st1 {v18.8b},[x2], x3 //st (row 3) 259 umull v18.8h, v23.8b, v7.8b //mul (row 7) 260 umlal v18.8h, v25.8b, v6.8b //mul (row 7) 264 rshrn v18.8b, v18.8h,#5 //round shft (row 7) 270 st1 {v18 [all...] |
/external/valgrind/none/tests/arm64/ |
H A D | memory.stdout.exp | 129 0000000000000000 v18.d[0] (xor, xfer vecreg #2) 130 0000000000000000 v18.d[1] (xor, xfer vecreg #2) 159 0000000000000000 v18.d[0] (xor, xfer vecreg #2) 160 0000000000000000 v18.d[1] (xor, xfer vecreg #2) 189 0000000000000000 v18.d[0] (xor, xfer vecreg #2) 190 0000000000000000 v18.d[1] (xor, xfer vecreg #2) 219 0000000000000000 v18.d[0] (xor, xfer vecreg #2) 220 0000000000000000 v18.d[1] (xor, xfer vecreg #2) 249 0000000000000000 v18.d[0] (xor, xfer vecreg #2) 250 0000000000000000 v18 [all...] |
/external/llvm/test/MC/Hexagon/ |
H A D | v60-permute.s | 5 #CHECK: 1fd2d5cf { v15.b = vpack(v21.h{{ *}},{{ *}}v18.h):sat } 6 v15.b=vpack(v21.h,v18.h):sat 29 #CHECK: 1e01d256 { v23:22.h = vunpack(v18.b) } 30 v23:22.h=vunpack(v18.b)
|