/external/libhevc/common/arm64/ |
H A D | ihevc_intra_pred_chroma_ver.s | 121 ld2 {v22.8b, v23.8b}, [x6] //16 loads (col 16:31) 132 st2 {v22.8b, v23.8b}, [x2], x11 133 st2 {v22.8b, v23.8b}, [x5], x11 134 st2 {v22.8b, v23.8b}, [x8], x11 135 st2 {v22.8b, v23.8b}, [x10], x11 145 st2 {v22.8b, v23.8b}, [x2], x11 146 st2 {v22.8b, v23.8b}, [x5], x11 147 st2 {v22.8b, v23.8b}, [x8], x11 148 st2 {v22.8b, v23.8b}, [x10], x11 158 st2 {v22.8b, v23 [all...] |
H A D | ihevc_intra_pred_luma_vert.s | 122 ld1 {v22.8b, v23.8b}, [x6] //16 loads (col 16:31) 131 st1 {v22.8b, v23.8b}, [x2], x11 132 st1 {v22.8b, v23.8b}, [x5], x11 133 st1 {v22.8b, v23.8b}, [x8], x11 134 st1 {v22.8b, v23.8b}, [x10], x11 144 st1 {v22.8b, v23.8b}, [x2], x11 145 st1 {v22.8b, v23.8b}, [x5], x11 146 st1 {v22.8b, v23.8b}, [x8], x11 147 st1 {v22.8b, v23.8b}, [x10], x11 156 st1 {v22.8b, v23 [all...] |
H A D | ihevc_intra_pred_chroma_mode_3_to_9.s | 213 tbl v23.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 3) 222 umlal v18.8h, v23.8b, v6.8b //mul (row 3) 259 tbl v23.8b, { v0.16b, v1.16b}, v5.8b //load from ref_main_idx + 1 (row 7) 261 umlal v18.8h, v23.8b, v6.8b //mul (row 7) 298 sqxtn v23.8b, v25.8h 299 shl v23.8b, v23.8b,#1 311 sub v23.8b, v23.8b , v27.8b //ref_main_idx (sub row) 315 sub v2.8b, v26.8b , v23 [all...] |
H A D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 129 dup v23.4h, v0.h[1] //coeffabs_1 = vdup_lane_u8(coeffabs, 1)// 155 smull v19.4s, v1.4h, v23.4h //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 171 smull v20.4s, v2.4h, v23.4h //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 188 smull v21.4s, v3.4h, v23.4h 202 smull v30.4s, v4.4h, v23.4h 229 smull v19.4s, v1.4h, v23.4h //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 246 smull v20.4s, v2.4h, v23.4h //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 262 smull v21.4s, v3.4h, v23.4h 289 smull v30.4s, v4.4h, v23.4h 316 smull v19.4s, v1.4h, v23 [all...] |
H A D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 138 dup v23.4h,v0.h[1] //coeffabs_1 = vdup_lane_u8(coeffabs, 1)// 166 smull v19.4s,v1.4h,v23.4h //mul_res1 = smull_u8(src_tmp2, coeffabs_1)// 182 smull v20.4s,v2.4h,v23.4h //mul_res2 = smull_u8(src_tmp3, coeffabs_1)// 199 smull v21.4s,v3.4h,v23.4h 214 smull v31.4s,v4.4h,v23.4h 243 smull v19.4s,v1.4h,v23.4h //mul_res1 = smull_u8(src_tmp2, coeffabs_1)// 261 smull v20.4s,v2.4h,v23.4h //mul_res2 = smull_u8(src_tmp3, coeffabs_1)// 278 smull v21.4s,v3.4h,v23.4h 306 smull v31.4s,v4.4h,v23.4h 334 smull v19.4s,v1.4h,v23 [all...] |
H A D | ihevc_intra_pred_filters_chroma_mode_19_to_25.s | 281 umull v23.8h, v7.8b, v30.8b //(i row)vmull_u8(ref_main_idx, dup_const_32_fract) 284 umlal v23.8h, v19.8b, v31.8b //(i row)vmull_u8(ref_main_idx_1, dup_const_fract) 298 rshrn v23.8b, v23.8h,#5 //(i row)shift_res = vrshrn_n_u16(add_res, 5) 312 st1 {v23.8b},[x2],#8 //(i row) 339 umull v23.8h, v7.8b, v30.8b //(v)vmull_u8(ref_main_idx, dup_const_32_fract) 342 umlal v23.8h, v19.8b, v31.8b //(v)vmull_u8(ref_main_idx_1, dup_const_fract) 364 rshrn v23.8b, v23.8h,#5 //(v)shift_res = vrshrn_n_u16(add_res, 5) 378 st1 {v23 [all...] |
H A D | ihevc_sao_edge_offset_class1_chroma.s | 220 //TBL v23.8b, {v6.16b},v23.8b //II vtbl1_s8(edge_idx_tbl, vget_high_s8(edge_idx)) 222 mov v23.d[0], v22.d[1] 223 UZP1 v27.8b, v22.8b, v23.8b 224 UZP2 v23.8b, v22.8b, v23.8b 235 TBL v25.8b, {v1.16b},v23.8b 281 //TBL v23.8b, {v6.16b},v23.8b //vtbl1_s8(edge_idx_tbl, vget_high_s8(edge_idx)) 283 mov v23 [all...] |
H A D | ihevc_intra_pred_luma_mode2.s | 154 rev64 v23.8b, v7.8b 180 st1 {v23.8b},[x14],x5 226 rev64 v23.8b, v7.8b 239 st1 {v23.8b},[x14],x5
|
H A D | ihevc_inter_pred_filters_luma_vert.s | 137 dup v23.8b, v0.b[1] //coeffabs_1 = vdup_lane_u8(coeffabs, 1)// 164 umull v19.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 182 umull v20.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 212 umull v21.8h, v3.8b, v23.8b 228 umull v30.8h, v4.8b, v23.8b 253 umull v19.8h, v1.8b, v23.8b //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 281 umull v20.8h, v2.8b, v23.8b //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 307 umull v21.8h, v3.8b, v23.8b 337 umull v30.8h, v4.8b, v23.8b 365 umull v19.8h, v1.8b, v23 [all...] |
/external/llvm/test/MC/Hexagon/ |
H A D | v60-permute.s | 8 #CHECK: 1fd7d7a2 { v2.ub = vpack(v23.h{{ *}},{{ *}}v23.h):sat } 9 v2.ub=vpack(v23.h,v23.h):sat 29 #CHECK: 1e01d256 { v23:22.h = vunpack(v18.b) } 30 v23:22.h=vunpack(v18.b) 38 #CHECK: 1e01d778 { v25:24.w = vunpack(v23.h) } 39 v25:24.w=vunpack(v23.h)
|
H A D | v60-vmpy1.s | 20 #CHECK: 1908da97 { v23.w = vrmpy(v26.ub,{{ *}}r8.b) } 21 v23.w=vrmpy(v26.ub,r8.b) 71 #CHECK: 1946de76 { v23:22.uw = vmpy(v30.uh,{{ *}}r6.uh) } 72 v23:22.uw=vmpy(v30.uh,r6.uh) 80 #CHECK: 1c34d937 { v23.h = vmpy(v25.h,{{ *}}v20.h):<<1:rnd:sat } 81 v23.h=vmpy(v25.h,v20.h):<<1:rnd:sat 98 #CHECK: 1c16d77c { v28.w = vdmpy(v23.h,{{ *}}v22.h):sat } 99 v28.w=vdmpy(v23.h,v22.h):sat 131 #CHECK: 1ff7cbfa { v26.w = vmpyo(v11.w{{ *}},{{ *}}v23.h):<<1:sat } 132 v26.w=vmpyo(v11.w,v23 [all...] |
H A D | v60-vcmp.s | 44 #CHECK: 1c9bf795 { q1 ^= vcmp.gt(v23.h{{ *}},{{ *}}v27.h) } 45 q1^=vcmp.gt(v23.h,v27.h) 77 #CHECK: 1c91f75a { q2 |= vcmp.gt(v23.w{{ *}},{{ *}}v17.w) } 78 q2|=vcmp.gt(v23.w,v17.w)
|
H A D | v60-shift.s | 14 #CHECK: 1965d7ac { v12.w = vasr(v23.w,{{ *}}r5) } 15 v12.w=vasr(v23.w,r5)
|
H A D | v60-vmpy-acc.s | 5 #CHECK: 1936ee37 { v23.w += vdmpy(v15:14.h,r22.uh,#1):sat } 6 v23.w += vdmpy(v15:14.h,r22.uh,#1):sat 35 #CHECK: 1987f71e { v31:30.uh += vmpy(v23.ub,r7.ub) } 36 v31:30.uh += vmpy(v23.ub,r7.ub) 62 #CHECK: 194efe36 { v23:22.uw += vmpy(v30.uh,r14.uh) } 63 v23:22.uw += vmpy(v30.uh,r14.uh) 104 #CHECK: 1c17fcc0 { v1:0.h += vmpy(v28.ub,v23.b) } 105 v1:0.h += vmpy(v28.ub,v23.b)
|
/external/libavc/common/armv8/ |
H A D | ih264_resi_trans_quant_av8.s | 146 shl v23.4h, v21.4h, #1 //u_shift(x3,1,shft) 152 add v25.4h, v23.4h , v20.4h //x6 = u_shift(x3,1,shft) + x2; 155 dup v23.4s, w8 //load round factor values 176 add v20.4s, v0.4s, v23.4s 177 add v21.4s, v1.4s, v23.4s 178 add v22.4s, v2.4s, v23.4s 179 add v23.4s, v3.4s, v23.4s 186 sshl v23.4s, v23 [all...] |
H A D | ih264_inter_pred_filters_luma_horz_av8.s | 180 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row1) 189 st1 {v23.8b, v24.8b}, [x1], x3 ////Store dest row1 233 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row3) 241 st1 {v23.8b, v24.8b}, [x1], x3 ////Store dest row3 285 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row5) 295 st1 {v23.8b, v24.8b}, [x1], x3 ////Store dest row5 335 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row7) 339 st1 {v23.8b, v24.8b}, [x1], x3 ////Store dest row7 359 ext v23.8b, v5.8b , v6.8b, #1 ////extract a[1] (column1,row1) 365 umlsl v14.8h, v23 [all...] |
H A D | ih264_iquant_itrans_recon_av8.s | 137 ld4 {v20.4h - v23.4h}, [x5] // load pu2_iscal_mat[i], i =0..15 145 mul v23.4h, v23.4h, v29.4h // x[i]=(scale[i] * dequant[i]) where i = 12..14 150 smull v6.4s, v19.4h, v23.4h // q3 = p[i] = (x[i] * trns_coeff[i]) where i = 12..15 216 sub v23.4h, v14.4h, v17.4h // x0 - x3 219 mov v22.d[1], v23.d[0] 330 ld4 {v20.4h - v23.4h}, [x5] // load pu2_iscal_mat[i], i =0..15 338 mul v23.4h, v23.4h, v29.4h // x[i]=(scale[i] * dequant[i]) where i = 12..14 343 smull v6.4s, v19.4h, v23 [all...] |
H A D | ih264_inter_pred_luma_horz_hpel_vert_hpel_av8.s | 133 ext v23.16b, v18.16b , v20.16b , #10 139 saddl v26.4s, v18.4h, v23.4h 143 saddl2 v23.4s, v18.8h, v23.8h 144 smlal2 v23.4s, v0.8h, v28.8h 145 smlsl2 v23.4s, v24.8h, v30.8h 148 sqrshrun v19.4h, v23.4s, #10 214 ext v23.16b, v18.16b , v20.16b , #10 220 saddl v26.4s, v18.4h, v23.4h 224 saddl2 v23 [all...] |
H A D | ih264_intra_pred_luma_16x16_av8.s | 225 dup v23.16b, v0.b[2] 231 st1 {v23.16b}, [x1], x3 514 sqrshrun v23.8b, v28.8h, #5 518 st1 {v22.2s, v23.2s}, [x2], x3 524 sqrshrun v23.8b, v28.8h, #5 528 st1 {v22.2s, v23.2s}, [x2], x3 534 sqrshrun v23.8b, v28.8h, #5 538 st1 {v22.2s, v23.2s}, [x2], x3 544 sqrshrun v23.8b, v28.8h, #5 548 st1 {v22.2s, v23 [all...] |
/external/libavc/encoder/armv8/ |
H A D | ih264e_half_pel_av8.s | 177 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row1) 183 st1 {v23.8b, v24.8b}, [x1], #16 ////Store dest row1 321 mov v23.d[0], v22.d[1] 338 ext v31.8b, v22.8b , v23.8b , #2 354 ext v31.8b, v23.8b , v24.8b , #2 363 ext v30.8b, v22.8b , v23.8b , #4 364 ext v29.8b, v22.8b , v23.8b , #6 367 ext v28.8b, v22.8b , v23.8b , #2 371 smlsl v2.4s, v23.4h, v0.h[0] //// a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 (set3) 375 ext v30.8b, v23 [all...] |
/external/llvm/test/MC/AArch64/ |
H A D | neon-scalar-by-elem-mla.s | 29 fmls s5, s12, v23.s[2] 37 // CHECK: fmls s5, s12, v23.s[2] // encoding: [0x85,0x59,0x97,0x5f]
|
/external/libhevc/decoder/arm64/ |
H A D | ihevcd_fmt_conv_420sp_to_rgba8888.s | 242 movi v23.8b, #0 254 ZIP1 v27.8b, v22.8b, v23.8b 255 ZIP2 v23.8b, v22.8b, v23.8b 261 mov v22.d[1], v23.d[0] 308 movi v23.8b, #0 320 ZIP1 v27.8b, v22.8b, v23.8b 321 ZIP2 v23.8b, v22.8b, v23.8b 327 mov v22.d[1], v23 [all...] |
/external/linux-kselftest/tools/testing/selftests/powerpc/include/ |
H A D | vmx_asm.h | 21 stvx v23,reg,%r1; \ 48 lvx v23,reg,%r1; \ 78 lvx v23,r5,r3 variable
|
/external/capstone/suite/MC/AArch64/ |
H A D | neon-scalar-by-elem-mla.s.cs | 10 0x85,0x59,0x97,0x5f = fmls s5, s12, v23.s[2]
|
/external/compiler-rt/lib/tsan/rtl/ |
H A D | tsan_ppc_regs.h | 88 #define v23 23 macro
|