/external/llvm/test/MC/AArch64/ |
H A D | neon-scalar-shift-imm.s | 180 sqrshrun b17, h10, #6 181 sqrshrun h10, s13, #15 182 sqrshrun s22, d16, #31 184 // CHECK: sqrshrun b17, h10, #6 // encoding: [0x51,0x8d,0x0a,0x7f] 185 // CHECK: sqrshrun h10, s13, #15 // encoding: [0xaa,0x8d,0x11,0x7f] 186 // CHECK: sqrshrun s22, d16, #31 // encoding: [0x16,0x8e,0x21,0x7f]
|
H A D | neon-simd-shift.s | 314 sqrshrun v0.8b, v1.8h, #3 315 sqrshrun v0.4h, v1.4s, #3 316 sqrshrun v0.2s, v1.2d, #3 321 // CHECK: sqrshrun v0.8b, v1.8h, #3 // encoding: [0x20,0x8c,0x0d,0x2f] 322 // CHECK: sqrshrun v0.4h, v1.4s, #3 // encoding: [0x20,0x8c,0x1d,0x2f] 323 // CHECK: sqrshrun v0.2s, v1.2d, #3 // encoding: [0x20,0x8c,0x3d,0x2f]
|
H A D | arm64-advsimd.s | 1357 sqrshrun b0, h0, #1 1358 sqrshrun h0, s0, #2 1359 sqrshrun s0, d0, #3 1406 ; CHECK: sqrshrun b0, h0, #1 ; encoding: [0x00,0x8c,0x0f,0x7f] 1407 ; CHECK: sqrshrun h0, s0, #2 ; encoding: [0x00,0x8c,0x1e,0x7f] 1408 ; CHECK: sqrshrun s0, d0, #3 ; encoding: [0x00,0x8c,0x3d,0x7f] 1492 sqrshrun.8b v0, v0, #1 1494 sqrshrun.4h v0, v0, #3 1496 sqrshrun.2s v0, v0, #5 1664 ; CHECK: sqrshrun [all...] |
/external/libavc/common/armv8/ |
H A D | ih264_inter_pred_filters_luma_vert_av8.s | 156 sqrshrun v30.8b, v14.8h, #5 // dst[0_0] = CLIP_U8((temp +16) >> 5) 161 sqrshrun v31.8b, v20.8h, #5 // dst[0_8] = CLIP_U8((temp4 +16) >> 5) 169 sqrshrun v30.8b, v16.8h, #5 177 sqrshrun v31.8b, v14.8h, #5 184 sqrshrun v30.8b, v18.8h, #5 191 sqrshrun v31.8b, v16.8h, #5 197 sqrshrun v30.8b, v14.8h, #5 200 sqrshrun v31.8b, v18.8h, #5 216 sqrshrun v30.8b, v14.8h, #5 // dst[0_0] = CLIP_U8((temp +16) >> 5) 219 sqrshrun v3 [all...] |
H A D | ih264_intra_pred_luma_16x16_av8.s | 508 sqrshrun v20.8b, v26.8h, #5 509 sqrshrun v21.8b, v28.8h, #5 512 sqrshrun v22.8b, v26.8h, #5 514 sqrshrun v23.8b, v28.8h, #5 517 sqrshrun v20.8b, v26.8h, #5 519 sqrshrun v21.8b, v28.8h, #5 522 sqrshrun v22.8b, v26.8h, #5 524 sqrshrun v23.8b, v28.8h, #5 527 sqrshrun v20.8b, v26.8h, #5 529 sqrshrun v2 [all...] |
H A D | ih264_inter_pred_chroma_av8.s | 166 sqrshrun v26.8b, v20.8h, #6 174 sqrshrun v27.8b, v22.8h, #6 184 sqrshrun v18.8b, v24.8h, #6 188 sqrshrun v19.8b, v16.8h, #6 194 sqrshrun v26.8b, v20.8h, #6 203 sqrshrun v27.8b, v24.8h, #6 211 sqrshrun v26.8b, v20.8h, #6 214 sqrshrun v27.8b, v22.8h, #6 229 sqrshrun v18.8b, v24.8h, #6 234 sqrshrun v1 [all...] |
H A D | ih264_inter_pred_luma_horz_qpel_vert_qpel_av8.s | 167 sqrshrun v26.8b, v24.8h, #5 179 sqrshrun v28.8b, v28.8h, #5 186 sqrshrun v27.8b, v24.8h, #5 201 sqrshrun v29.8b, v24.8h, #5 211 sqrshrun v26.8b, v16.8h, #5 219 sqrshrun v27.8b, v24.8h, #5 236 sqrshrun v28.8b, v28.8h, #5 250 sqrshrun v29.8b, v24.8h, #5 258 sqrshrun v26.8b, v16.8h, #5 275 sqrshrun v2 [all...] |
H A D | ih264_inter_pred_luma_vert_qpel_av8.s | 163 sqrshrun v30.8b, v14.8h, #5 // dst[0_0] = CLIP_U8((temp +16) >> 5) 168 sqrshrun v31.8b, v20.8h, #5 // dst[0_8] = CLIP_U8((temp4 +16) >> 5) 178 sqrshrun v30.8b, v16.8h, #5 185 sqrshrun v31.8b, v14.8h, #5 195 sqrshrun v30.8b, v18.8h, #5 202 sqrshrun v31.8b, v16.8h, #5 211 sqrshrun v30.8b, v14.8h, #5 214 sqrshrun v31.8b, v18.8h, #5 233 sqrshrun v30.8b, v14.8h, #5 // dst[0_0] = CLIP_U8((temp +16) >> 5) 235 sqrshrun v3 [all...] |
H A D | ih264_inter_pred_filters_luma_horz_av8.s | 175 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row0) 177 sqrshrun v21.8b, v10.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row0) 180 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row1) 182 sqrshrun v24.8b, v16.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row1) 228 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row2) 230 sqrshrun v21.8b, v10.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row2) 233 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row3) 235 sqrshrun v24.8b, v16.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row3) 280 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row4) 282 sqrshrun v2 [all...] |
H A D | ih264_intra_pred_luma_8x8_av8.s | 328 sqrshrun v31.8b, v12.8h, #4 473 sqrshrun v4.8b, v24.8h, #2 474 sqrshrun v5.8b, v26.8h, #2 574 sqrshrun v4.8b, v24.8h, #2 575 sqrshrun v5.8b, v26.8h, #2 675 sqrshrun v4.8b, v20.8h, #1 676 sqrshrun v5.8b, v22.8h, #1 678 sqrshrun v6.8b, v24.8h, #2 679 sqrshrun v7.8b, v26.8h, #2 801 sqrshrun v [all...] |
H A D | ih264_inter_pred_luma_horz_qpel_vert_hpel_av8.s | 210 sqrshrun v18.4h, v26.4s, #10 211 sqrshrun v19.4h, v22.4s, #10 235 sqrshrun v19.4h, v26.4s, #10 236 sqrshrun v18.4h, v22.4s, #10 248 sqrshrun v20.8b, v20.8h, #5 249 sqrshrun v21.8b, v22.8h, #5 297 sqrshrun v18.4h, v26.4s, #10 298 sqrshrun v19.4h, v22.4s, #10 323 sqrshrun v19.4h, v26.4s, #10 324 sqrshrun v1 [all...] |
H A D | ih264_inter_pred_luma_horz_hpel_vert_hpel_av8.s | 147 sqrshrun v18.4h, v26.4s, #10 148 sqrshrun v19.4h, v23.4s, #10 172 sqrshrun v19.4h, v26.4s, #10 173 sqrshrun v25.4h, v22.4s, #10 228 sqrshrun v18.4h, v26.4s, #10 229 sqrshrun v19.4h, v23.4s, #10 254 sqrshrun v19.4h, v26.4s, #10 255 sqrshrun v25.4h, v22.4s, #10 307 sqrshrun v18.4h, v26.4s, #10 308 sqrshrun v1 [all...] |
H A D | ih264_inter_pred_luma_horz_hpel_vert_qpel_av8.s | 264 sqrshrun v26.8b, v26.8h, #5 273 sqrshrun v18.4h, v18.4s, #10 275 sqrshrun v19.4h, v6.4s, #10 308 sqrshrun v28.8b, v8.8h, #5 312 sqrshrun v18.4h, v18.4s, #10 314 sqrshrun v19.4h, v6.4s, #10 349 sqrshrun v26.8b, v10.8h, #5 354 sqrshrun v18.4h, v18.4s, #10 356 sqrshrun v19.4h, v6.4s, #10 379 sqrshrun v1 [all...] |
H A D | ih264_intra_pred_chroma_av8.s | 511 sqrshrun v28.8b, v24.8h, #5 513 sqrshrun v29.8b, v0.8h, #5 516 sqrshrun v28.8b, v2.8h, #5 517 sqrshrun v29.8b, v26.8h, #5 524 sqrshrun v28.8b, v24.8h, #5 525 sqrshrun v29.8b, v0.8h, #5 532 sqrshrun v28.8b, v2.8h, #5 533 sqrshrun v29.8b, v26.8h, #5 540 sqrshrun v28.8b, v24.8h, #5 541 sqrshrun v2 [all...] |
H A D | ih264_inter_pred_luma_horz_qpel_av8.s | 184 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row0) 186 sqrshrun v21.8b, v10.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row0) 191 sqrshrun v18.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row1) 194 sqrshrun v19.8b, v16.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row1) 245 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row2) 247 sqrshrun v21.8b, v10.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row2) 252 sqrshrun v18.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row3) 255 sqrshrun v19.8b, v16.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row3) 303 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row4) 305 sqrshrun v2 [all...] |
H A D | ih264_intra_pred_luma_4x4_av8.s | 419 sqrshrun v3.8b, v24.8h, #2 504 sqrshrun v3.8b, v24.8h, #2 588 sqrshrun v4.8b, v20.8h, #1 589 sqrshrun v3.8b, v24.8h, #2 673 sqrshrun v4.8b, v20.8h, #1 674 sqrshrun v5.8b, v24.8h, #2 761 sqrshrun v4.8b, v20.8h, #1 762 sqrshrun v5.8b, v24.8h, #2 845 sqrshrun v4.8b, v20.8h, #1 846 sqrshrun v [all...] |
/external/libhevc/common/arm64/ |
H A D | ihevc_inter_pred_chroma_vert.s | 158 sqrshrun v6.8b, v6.8h,#6 //shifts right 161 sqrshrun v4.8b, v4.8h,#6 //shifts right 202 sqrshrun v4.8b, v4.8h,#6 //vrshrq_n_s16(vreinterpretq_s16_u16(mul_res1),6) 254 sqrshrun v30.8b, v30.8h,#6 264 sqrshrun v28.8b, v28.8h,#6 279 sqrshrun v26.8b, v26.8h,#6 300 sqrshrun v24.8b, v24.8h,#6 313 sqrshrun v30.8b, v30.8h,#6 335 sqrshrun v28.8b, v28.8h,#6 356 sqrshrun v2 [all...] |
H A D | ihevc_inter_pred_chroma_horz.s | 253 sqrshrun v30.8b, v30.8h,#6 259 sqrshrun v31.8b, v28.8h,#6 308 sqrshrun v22.8b, v22.8h,#6 309 sqrshrun v23.8b, v20.8h,#6 344 sqrshrun v30.8b, v30.8h,#6 345 sqrshrun v31.8b, v28.8h,#6 406 sqrshrun v22.8b, v22.8h,#6 407 sqrshrun v23.8b, v20.8h,#6 421 sqrshrun v30.8b, v30.8h,#6 422 sqrshrun v3 [all...] |
H A D | ihevc_inter_pred_chroma_vert_w16inp.s | 160 sqrshrun v0.8b, v0.8h,#6 //rounding shift 161 sqrshrun v30.8b, v30.8h,#6 //rounding shift 220 sqrshrun v30.8b, v30.8h,#6 //rounding shift 236 sqrshrun v28.8b, v28.8h,#6 //rounding shift 249 sqrshrun v26.8b, v26.8h,#6 //rounding shift 263 sqrshrun v24.8b, v24.8h,#6 //rounding shift 276 sqrshrun v30.8b, v30.8h,#6 //rounding shift 290 sqrshrun v28.8b, v28.8h,#6 //rounding shift 306 sqrshrun v26.8b, v26.8h,#6 //rounding shift 316 sqrshrun v2 [all...] |
H A D | ihevc_inter_pred_filters_luma_vert_w16inp.s | 200 sqrshrun v19.8b, v19.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 218 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 243 sqrshrun v21.8b, v21.8h,#6 260 sqrshrun v30.8b, v30.8h,#6 286 sqrshrun v19.8b, v19.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 309 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 327 sqrshrun v21.8b, v21.8h,#6 341 sqrshrun v30.8b, v30.8h,#6 354 sqrshrun v19.8b, v19.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 366 sqrshrun v2 [all...] |
H A D | ihevc_inter_pred_filters_luma_vert.s | 209 sqrshrun v19.8b, v19.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 224 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 244 sqrshrun v21.8b, v21.8h,#6 278 sqrshrun v30.8b, v30.8h,#6 304 sqrshrun v19.8b, v19.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 333 sqrshrun v20.8b, v20.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 356 sqrshrun v21.8b, v21.8h,#6 375 sqrshrun v30.8b, v30.8h,#6 388 sqrshrun v19.8b, v19.8h,#6 //sto_res = vqmovun_s16(sto_res_tmp)// 401 sqrshrun v2 [all...] |
H A D | ihevc_inter_pred_filters_luma_horz.s | 253 sqrshrun v20.8b, v8.8h,#6 //right shift and saturating narrow result 1 262 sqrshrun v8.8b, v10.8h,#6 //right shift and saturating narrow result 2 372 sqrshrun v8.8b, v8.8h,#6 //right shift and saturating narrow result 1 391 sqrshrun v9.8b, v20.8h,#6 412 sqrshrun v10.8b, v10.8h,#6 //right shift and saturating narrow result 2 442 sqrshrun v11.8b, v22.8h,#6 467 sqrshrun v11.8b, v22.8h,#6 581 sqrshrun v8.8b, v8.8h,#6 //narrow right shift and saturating the result
|
/external/capstone/suite/MC/AArch64/ |
H A D | neon-scalar-shift-imm.s.cs | 40 0x51,0x8d,0x0a,0x7f = sqrshrun b17, h10, #6 41 0xaa,0x8d,0x11,0x7f = sqrshrun h10, s13, #15 42 0x16,0x8e,0x21,0x7f = sqrshrun s22, d16, #31
|
H A D | neon-simd-shift.s.cs | 110 0x20,0x8c,0x0d,0x2f = sqrshrun v0.8b, v1.8h, #3 111 0x20,0x8c,0x1d,0x2f = sqrshrun v0.4h, v1.4s, #3 112 0x20,0x8c,0x3d,0x2f = sqrshrun v0.2s, v1.2d, #3
|
/external/libavc/encoder/armv8/ |
H A D | ih264e_half_pel_av8.s | 174 sqrshrun v20.8b, v8.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row0) 175 sqrshrun v21.8b, v10.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row0) 176 sqrshrun v22.8b, v12.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column3,row0) 177 sqrshrun v23.8b, v14.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row1) 178 sqrshrun v24.8b, v16.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row1) 179 sqrshrun v25.8b, v18.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column3,row1) 332 sqrshrun v2.8b, v20.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column1,row0) 334 sqrshrun v3.8b, v22.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column2,row0) 345 sqrshrun v4.8b, v24.8h, #5 //// (a0 + a5 + 20a2 + 20a3 - 5a1 - 5a4 + 16) >> 5 (column3,row0) 401 sqrshrun v2 [all...] |