/external/libhevc/common/arm64/ |
H A D | ihevc_weighted_pred_bi.s | 222 smull v4.4s, v0.4h, v7.h[0] //vmull_n_s16(pi2_src1_val1, (int16_t) wgt0) 224 smull v5.4s, v1.4h, v7.h[1] //vmull_n_s16(pi2_src2_val1, (int16_t) wgt1) 229 smull v6.4s, v2.4h, v7.h[0] //vmull_n_s16(pi2_src1_val2, (int16_t) wgt0) ii iteration 233 smull v19.4s, v0.4h, v7.h[0] //vmull_n_s16(pi2_src1_val1, (int16_t) wgt0) iii iteration 236 smull v17.4s, v3.4h, v7.h[1] //vmull_n_s16(pi2_src2_val2, (int16_t) wgt1) ii iteration 243 smull v16.4s, v1.4h, v7.h[1] //vmull_n_s16(pi2_src2_val1, (int16_t) wgt1) iii iteration 251 smull v18.4s, v2.4h, v7.h[0] //vmull_n_s16(pi2_src1_val2, (int16_t) wgt0) iv iteration 257 smull v20.4s, v3.4h, v7.h[1] //vmull_n_s16(pi2_src2_val2, (int16_t) wgt1) iv iteration
|
H A D | ihevc_itrans_recon_8x8.s | 187 smull v20.4s, v2.4h, v0.h[0] //// y0 * cos4(part of c0 and c1) 189 smull v18.4s, v3.4h, v1.h[2] //// y2 * sin2 (q3 is freed by this time)(part of d1) 192 smull v24.4s, v6.4h, v0.h[1] //// y1 * cos1(part of b0) 194 smull v26.4s, v6.4h, v0.h[3] //// y1 * cos3(part of b1) 196 smull v28.4s, v6.4h, v1.h[1] //// y1 * sin3(part of b2) 198 smull v30.4s, v6.4h, v1.h[3] //// y1 * sin1(part of b3) 208 smull v22.4s, v10.4h, v0.h[0] //// y4 * cos4(part of c0 and c1) 210 smull v6.4s, v3.4h, v0.h[2] //// y2 * cos2(part of d0) 304 smull v24.4s, v6.4h, v0.h[1] //// y1 * cos1(part of b0) 305 smull v2 [all...] |
H A D | ihevc_itrans_recon_4x4_ttype1.s | 140 smull v6.4s, v1.4h, v4.h[2] //74 * pi2_src[1] 145 smull v5.4s, v1.4h, v4.h[2] //74 * pi2_src[1] 150 smull v7.4s, v0.4h, v4.h[2] // 74 * pi2_src[0] 154 smull v20.4s, v2.4h, v4.h[1] // 55 * pi2_src[2] 183 smull v6.4s, v22.4h, v4.h[2] //74 * pi2_src[1] 188 smull v5.4s, v22.4h, v4.h[2] //74 * pi2_src[1] 193 smull v7.4s, v21.4h, v4.h[2] // 74 * pi2_src[0] 198 smull v20.4s, v16.4h, v4.h[1] // 55 * pi2_src[2]
|
H A D | ihevc_inter_pred_chroma_vert_w16inp.s | 144 smull v0.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 147 smull v7.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 192 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 200 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 213 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 224 smull v24.4s, v3.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 241 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 252 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 266 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 279 smull v2 [all...] |
H A D | ihevc_inter_pred_chroma_vert_w16inp_w16out.s | 144 smull v0.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 147 smull v7.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 191 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 199 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 212 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 222 smull v24.4s, v3.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 238 smull v30.4s, v0.4h, v16.4h //vmull_s16(src_tmp1, coeff_0) 248 smull v28.4s, v1.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 261 smull v26.4s, v2.4h, v16.4h //vmull_s16(src_tmp2, coeff_0) 273 smull v2 [all...] |
H A D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 155 smull v19.4s, v1.4h, v23.4h //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 171 smull v20.4s, v2.4h, v23.4h //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 188 smull v21.4s, v3.4h, v23.4h 202 smull v30.4s, v4.4h, v23.4h 229 smull v19.4s, v1.4h, v23.4h //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 246 smull v20.4s, v2.4h, v23.4h //mul_res2 = vmull_u8(src_tmp3, coeffabs_1)// 262 smull v21.4s, v3.4h, v23.4h 289 smull v30.4s, v4.4h, v23.4h 316 smull v19.4s, v1.4h, v23.4h //mul_res1 = vmull_u8(src_tmp2, coeffabs_1)// 330 smull v2 [all...] |
H A D | ihevc_inter_pred_luma_vert_w16inp_w16out.s | 166 smull v19.4s,v1.4h,v23.4h //mul_res1 = smull_u8(src_tmp2, coeffabs_1)// 182 smull v20.4s,v2.4h,v23.4h //mul_res2 = smull_u8(src_tmp3, coeffabs_1)// 199 smull v21.4s,v3.4h,v23.4h 214 smull v31.4s,v4.4h,v23.4h 243 smull v19.4s,v1.4h,v23.4h //mul_res1 = smull_u8(src_tmp2, coeffabs_1)// 261 smull v20.4s,v2.4h,v23.4h //mul_res2 = smull_u8(src_tmp3, coeffabs_1)// 278 smull v21.4s,v3.4h,v23.4h 306 smull v31.4s,v4.4h,v23.4h 334 smull v19.4s,v1.4h,v23.4h //mul_res1 = smull_u8(src_tmp2, coeffabs_1)// 349 smull v2 [all...] |
H A D | ihevc_itrans_recon_4x4.s | 143 smull v6.4s, v1.4h, v4.h[1] //83 * pi2_src[1] 145 smull v5.4s, v1.4h, v4.h[3] //36 * pi2_src[1] 176 smull v6.4s, v1.4h, v4.h[1] //83 * pi2_src[1] 179 smull v5.4s, v1.4h, v4.h[3] //36 * pi2_src[1]
|
H A D | ihevc_weighted_pred_uni.s | 175 smull v4.4s, v1.4h, v0.h[0] //vmull_n_s16(pi2_src_val1, (int16_t) wgt0) 180 smull v6.4s, v2.4h, v0.h[0] //vmull_n_s16(pi2_src_val2, (int16_t) wgt0) ii iteration 187 smull v7.4s, v3.4h, v0.h[0] //vmull_n_s16(pi2_src_val1, (int16_t) wgt0) iii iteration 196 smull v16.4s, v5.4h, v0.h[0] //vmull_n_s16(pi2_src_val2, (int16_t) wgt0) iv iteration
|
H A D | ihevc_itrans_recon_16x16.s | 255 smull v24.4s, v6.4h, v0.h[1] //// y1 * cos1(part of b0) 256 smull v26.4s, v6.4h, v0.h[3] //// y1 * cos3(part of b1) 257 smull v28.4s, v6.4h, v1.h[1] //// y1 * sin3(part of b2) 258 smull v30.4s, v6.4h, v1.h[3] //// y1 * sin1(part of b3) 270 smull v12.4s, v10.4h, v0.h[0] 272 smull v14.4s, v10.4h, v0.h[0] 274 smull v16.4s, v10.4h, v0.h[0] 276 smull v18.4s, v10.4h, v0.h[0] 433 smull v24.4s, v6.4h, v2.h[1] //// y1 * cos1(part of b0) 434 smull v2 [all...] |
H A D | ihevc_itrans_recon_32x32.s | 216 smull v24.4s, v8.4h, v0.h[1] //// y1 * cos1(part of b0) 217 smull v26.4s, v8.4h, v0.h[3] //// y1 * cos3(part of b1) 218 smull v28.4s, v8.4h, v1.h[1] //// y1 * sin3(part of b2) 219 smull v30.4s, v8.4h, v1.h[3] //// y1 * sin1(part of b3) 230 smull v20.4s, v10.4h, v0.h[0] 234 smull v22.4s, v10.4h, v0.h[0] 237 smull v16.4s, v10.4h, v0.h[0] 240 smull v18.4s, v10.4h, v0.h[0] 577 smull v24.4s, v8.4h, v2.h[1] //// y1 * cos1(part of b0) 578 smull v2 [all...] |
H A D | ihevc_intra_pred_luma_mode_3_to_9.s | 139 smull v22.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 294 smull v12.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 389 smull v14.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 481 smull v22.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col)
|
H A D | ihevc_intra_pred_filters_luma_mode_11_to_17.s | 255 smull v22.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 414 smull v12.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 509 smull v14.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 608 smull v22.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col)
|
H A D | ihevc_intra_pred_chroma_mode_3_to_9.s | 133 smull v22.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 295 smull v25.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col) 401 smull v14.8h, v30.8b, v31.8b //(col+1)*intra_pred_angle [0:7](col)
|
/external/libmpeg2/common/armv8/ |
H A D | impeg2_idct.s | 387 smull v20.4s, v2.4h, v0.h[0] //// y0 * cos4(part of c0 and c1) 389 smull v18.4s, v3.4h, v1.h[2] //// y2 * sin2 (q3 is freed by this time)(part of d1) 392 smull v24.4s, v6.4h, v0.h[1] //// y1 * cos1(part of b0) 394 smull v26.4s, v6.4h, v0.h[3] //// y1 * cos3(part of b1) 396 smull v28.4s, v6.4h, v1.h[1] //// y1 * sin3(part of b2) 398 smull v30.4s, v6.4h, v1.h[3] //// y1 * sin1(part of b3) 408 smull v22.4s, v10.4h, v0.h[0] //// y4 * cos4(part of c0 and c1) 410 smull v6.4s, v3.4h, v0.h[2] //// y2 * cos2(part of d0) 505 smull v24.4s, v6.4h, v0.h[1] //// y1 * cos1(part of b0) 506 smull v2 [all...] |
/external/llvm/test/MC/ARM/ |
H A D | directive-arch-armv4.s | 34 smull r4, r5, r6, r3
|
/external/libavc/common/armv8/ |
H A D | ih264_iquant_itrans_recon_av8.s | 147 smull v0.4s, v16.4h, v20.4h // q0 = p[i] = (x[i] * trns_coeff[i]) where i = 0..3 148 smull v2.4s, v17.4h, v21.4h // q1 = p[i] = (x[i] * trns_coeff[i]) where i = 4..7 149 smull v4.4s, v18.4h, v22.4h // q2 = p[i] = (x[i] * trns_coeff[i]) where i = 8..11 150 smull v6.4s, v19.4h, v23.4h // q3 = p[i] = (x[i] * trns_coeff[i]) where i = 12..15 340 smull v0.4s, v16.4h, v20.4h // q0 = p[i] = (x[i] * trns_coeff[i]) where i = 0..3 341 smull v2.4s, v17.4h, v21.4h // q1 = p[i] = (x[i] * trns_coeff[i]) where i = 4..7 342 smull v4.4s, v18.4h, v22.4h // q2 = p[i] = (x[i] * trns_coeff[i]) where i = 8..11 343 smull v6.4s, v19.4h, v23.4h // q3 = p[i] = (x[i] * trns_coeff[i]) where i = 12..15 560 smull v16.4s, v0.4h, v8.4h 562 smull v1 [all...] |
H A D | ih264_resi_trans_quant_av8.s | 171 smull v0.4s, v0.4h, v28.4h //multiply and add row 1 172 smull v1.4s, v1.4h, v29.4h //multiply and add row 2 173 smull v2.4s, v2.4h, v30.4h //multiply and add row 3 174 smull v3.4s, v3.4h, v31.4h //multiply and add row 4 373 smull v0.4s, v0.4h, v28.4h //multiply and add row 1 374 smull v1.4s, v1.4h, v29.4h //multiply and add row 2 375 smull v2.4s, v2.4h, v30.4h //multiply and add row 3 376 smull v3.4s, v3.4h, v31.4h //multiply and add row 4
|
H A D | ih264_iquant_itrans_recon_dc_av8.s | 141 smull v0.4s, v0.4h, v2.4h 357 smull v0.4s, v0.4h, v1.4h
|
H A D | ih264_intra_pred_chroma_av8.s | 455 smull v22.4s, v14.4h, v18.4h 456 smull v24.4s, v15.4h, v18.4h 457 smull v26.4s, v16.4h, v18.4h 458 smull v28.4s, v17.4h, v18.4h
|
/external/libjpeg-turbo/simd/ |
H A D | jsimd_arm64_neon.S | 291 smull v18.4s, v18.4h, XFIX_P_0_541 /* z1l z1 = MULTIPLY(z2 + z3, FIX_0_541196100); */ 331 smull v10.4s, v9.4h, XFIX_P_0_298 /* tmp0 = MULTIPLY(tmp0, FIX_0_298631336) */ 332 smull v12.4s, v7.4h, XFIX_P_2_053 /* tmp1 = MULTIPLY(tmp1, FIX_2_053119869) */ 333 smull v14.4s, v5.4h, XFIX_P_3_072 /* tmp2 = MULTIPLY(tmp2, FIX_3_072711026) */ 334 smull v16.4s, v3.4h, XFIX_P_1_501 /* tmp3 = MULTIPLY(tmp3, FIX_1_501321110) */ 335 smull v26.4s, v26.4h, XFIX_P_1_175 /* z5l z5 = MULTIPLY(z3 + z4, FIX_1_175875602) */ 336 smull v22.4s, v22.4h, XFIX_N_1_961 /* z3 = MULTIPLY(z3, - FIX_1_961570560) */ 337 smull v24.4s, v24.4h, XFIX_N_0_390 /* z4 = MULTIPLY(z4, - FIX_0_390180644) */ 338 smull v18.4s, v18.4h, XFIX_N_0_899 /* z1 = MULTIPLY(z1, - FIX_0_899976223) */ 339 smull v2 [all...] |
/external/freetype/include/freetype/internal/ |
H A D | ftcalc.h | 55 smull t2, t, b, a /* (lo=t2,hi=t) = a*b */ 87 "smull %1, %2, %4, %3\n\t" /* (lo=%1,hi=%2) = a*b */
|
/external/pdfium/third_party/freetype/include/freetype/internal/ |
H A D | ftcalc.h | 55 smull t2, t, b, a /* (lo=t2,hi=t) = a*b */ 87 "smull %1, %2, %4, %3\n\t" /* (lo=%1,hi=%2) = a*b */
|
/external/llvm/test/MC/AArch64/ |
H A D | neon-2velem.s | 219 smull v0.4s, v1.4h, v2.h[2] 220 smull v0.2d, v1.2s, v2.s[2] 221 smull v0.2d, v1.2s, v22.s[2] 226 // CHECK: smull v0.4s, v1.4h, v2.h[2] // encoding: [0x20,0xa0,0x62,0x0f] 227 // CHECK: smull v0.2d, v1.2s, v2.s[2] // encoding: [0x20,0xa8,0x82,0x0f] 228 // CHECK: smull v0.2d, v1.2s, v22.s[2] // encoding: [0x20,0xa8,0x96,0x0f]
|
H A D | neon-3vdiff.s | 209 smull v0.8h, v1.8b, v2.8b 210 smull v0.4s, v1.4h, v2.4h 211 smull v0.2d, v1.2s, v2.2s 213 // CHECK: smull v0.8h, v1.8b, v2.8b // encoding: [0x20,0xc0,0x22,0x0e] 214 // CHECK: smull v0.4s, v1.4h, v2.4h // encoding: [0x20,0xc0,0x62,0x0e] 215 // CHECK: smull v0.2d, v1.2s, v2.2s // encoding: [0x20,0xc0,0xa2,0x0e]
|